In computer data storage, partial-response maximum-likelihood (PRML) is a method for recovering the digital data from the weak analog read-back signal picked up by the head of a magnetic disk drive or tape drive. PRML was introduced to recover data more reliably or at a greater areal-density than earlier simpler schemes such as peak-detection. [1] These advances are important because most of the digital data in the world is stored using magnetic storage on hard disk or tape drives.
Ampex introduced PRML in a tape drive in 1984. IBM introduced PRML in a disk drive in 1990 and also coined the acronym PRML. Many advances have taken place since the initial introduction. Recent read/write channels operate at much higher data-rates, are fully adaptive, and, in particular, include the ability to handle nonlinear signal distortion and non-stationary, colored, data-dependent noise (PDNP or NPML).
Partial response refers to the fact that part of the response to an individual bit may occur at one sample instant while other parts fall in other sample instants. Maximum-likelihood refers to the detector finding the bit-pattern most likely to have been responsible for the read-back waveform.
Partial-response was first proposed by Adam Lender in 1963. [2] The method was generalized by Kretzmer in 1966. Kretzmer also classified the several different possible responses, [3] for example, PR1 is duobinary and PR4 is the response used in the classical PRML. In 1970, Kobayashi and Tang recognized the value of PR4 for the magnetic recording channel. [4]
Maximum-likelihood decoding using the eponymous Viterbi algorithm was proposed in 1967 by Andrew Viterbi as a means of decoding convolutional codes. [5]
By 1971, Hisashi Kobayashi at IBM had recognized that the Viterbi algorithm could be applied to analog channels with inter-symbol interference and particularly to the use of PR4 in the context of Magnetic Recording [6] (later called PRML). (The wide range of applications of the Viterbi algorithm is well described in a review paper by Dave Forney. [7] ) A simplified algorithm, based upon a difference metric, was used in the early implementations. This is due to Ferguson at Bell Labs. [8]
The first two implementations were in Tape (Ampex - 1984) and then in hard disk drives (IBM - 1990). Both are significant milestones with the Ampex implementation focused on very high data-rate for a digital instrumentation recorder and IBM focused on a high level of integration and low power consumption for a mass-market HDD. In both cases, the initial equalization to PR4 response was done with analog circuitry but the Viterbi algorithm was performed with digital logic. In the tape application, PRML superseded 'flat equalization'. In the HDD application, PRML superseded RLL codes with 'peak detection'.
The first implementation of PRML was shipped in 1984 in the Ampex Digital Cassette Recording System (DCRS). The chief engineer on DCRS was Charles Coleman. The machine evolved from a 6-head, transverse-scan, digital video tape recorder. DCRS was a cassette-based, digital, instrumentation recorder capable of extended play times at very high data-rate. [9] It became Ampex' most successful digital product. [10]
The heads and the read/write channel ran at the (then) remarkably high data-rate of 117 Mbits/s. [11] The PRML electronics were implemented with four 4-bit, Plessey analog-to-digital converters (A/D) and 100k ECL logic. [12] The PRML channel outperformed a competing implementation based on "Null-Zone Detection". [13] A prototype PRML channel was implemented earlier at 20 Mbit/s on a prototype 8-inch HDD, [14] but Ampex exited the HDD business in 1985. These implementations and their mode of operation are best described in a paper by Wood and Petersen. [15] Petersen was granted a patent on the PRML channel but it was never leveraged by Ampex. [16]
In 1990, IBM shipped the first PRML channel in an HDD in the IBM 0681 It was full-height 5¼-inch form-factor with up to 12 of 130 mm disks and had a maximum capacity of 857 MB.
The PRML channel for the IBM 0681 was developed in IBM Rochester lab. in Minnesota [17] with support from the IBM Zurich Research lab. in Switzerland. [18] A parallel R&D effort at IBM San Jose did not lead directly to a product. [19] A competing technology at the time was 17ML [20] an example of Finite-Depth Tree-Search (FDTS). [21] [22]
The IBM 0681 read/write channel ran at a data-rate of 24 Mbits/s but was more highly integrated with the entire channel contained in a single 68-pin PLCC integrated circuit operating off a 5 volt supply. As well as the fixed analog equalizer, the channel boasted a simple adaptive digital cosine equalizer [23] after the A/D to compensate for changes in radius and/or changes in the magnetic components.
The presence of nonlinear transition-shift (NLTS) distortion on NRZ recording at high density and/or high data-rate was recognized in 1979. [24] The magnitude and sources of NLTS can be identified using the 'extracted dipulse' technique. [25] [26]
Ampex was the first to recognize the impact of NLTS on PR4. [27] and was first to implement Write precompensation for PRML NRZ recording. 'Precomp.' largely cancels the effect of NLTS. [14] Precompensation is viewed as a necessity for a PRML system and is important enough to appear in the BIOS HDD setup [28] although it is now handled automatically by the HDD.
PR4 is characterized by an equalization target (+1, 0, -1) in bit-response sample values or (1-D)(1+D) in polynomial notation (here, D is the delay operator referring to a one sample delay). The target (+1, +1, -1, -1) or (1-D)(1+D)^2 is called Extended PRML (or EPRML). The entire family, (1-D)(1+D)^n, was investigated by Thapar and Patel. [29] The targets with larger n value tend to be more suited to channels with poor high-frequency response. This series of targets all have integer sample values and form an open eye-pattern (e.g. PR4 forms a ternary eye). In general, however, the target can just as readily have non-integer values. The classical approach to maximum-likelihood detection on a channel with intersymbol interference (ISI) is to equalize to a minimum-phase, whitened, matched-filter target. [30] The complexity of the subsequent Viterbi detector increases exponentially with the target length - the number of states doubling for each 1-sample increase in target length.
Given the rapid increase in complexity with longer targets, a post-processor architecture was proposed, firstly for EPRML. [31] With this approach a relatively simple detector (e.g. PRML) is followed by a post-processor which examines the residual waveform error and looks for the occurrence of likely bit pattern errors. This approach was found to be valuable when it was extended to systems employing a simple parity check [32] [33] [34]
As data detectors became more sophisticated, it was found important to deal with any residual signal nonlinearities as well as pattern-dependent noise (noise tends to be largest when there is a magnetic transition between bits) including changes in noise-spectrum with data-pattern. To this end, the Viterbi detector was modified such that it recognized the expected signal-level and expected noise variance associated with each bit-pattern. As a final step, the detectors were modified to include a 'noise predictor filter' thus allowing each pattern to have a different noise-spectrum. Such detectors are referred to as Pattern-Dependent Noise-Prediction (PDNP) detectors [35] or noise-predictive maximum-likelihood detectors (NPML). [36] Such techniques have been more recently applied to digital tape recorders. [37]
Although the PRML acronym is still occasionally used, advanced detectors are more complex PRML operate at higher data-rates. The analog front-end typically includes AGC, correction for the nonlinear read-element response, and a low-pass filter with control over the high-frequency boost or cut. Equalization is done after the ADC with a digital FIR filter. (TDMR uses a 2-input, 1-output equalizer.) The detector uses the PDNP/NPML approach but the hard-decision Viterbi algorithm is replaced with a detector providing soft-outputs (additional information about the reliability of each bit). Such detectors using a soft Viterbi algorithm or BCJR algorithm are essential in iteratively decoding the low-density parity-check code used in modern HDDs. A single integrated circuit contains the entire read and write channels (including the iterative decoder) as well as all the disk control and interface functions. There are currently two suppliers: Broadcom and Marvell. [38]
A hard disk drive (HDD), hard disk, hard drive, or fixed disk, is an electro-mechanical data storage device that stores and retrieves digital data using magnetic storage with one or more rigid rapidly rotating platters coated with magnetic material. The platters are paired with magnetic heads, usually arranged on a moving actuator arm, which read and write data to the platter surfaces. Data is accessed in a random-access manner, meaning that individual blocks of data can be stored and retrieved in any order. HDDs are a type of non-volatile storage, retaining stored data when powered off. Modern HDDs are typically in the form of a small rectangular box.
In telecommunication, a convolutional code is a type of error-correcting code that generates parity symbols via the sliding application of a boolean polynomial function to a data stream. The sliding application represents the 'convolution' of the encoder over the data, which gives rise to the term 'convolutional coding'. The sliding nature of the convolutional codes facilitates trellis decoding using a time-invariant trellis. Time invariant trellis decoding allows convolutional codes to be maximum-likelihood soft-decision decoded with reasonable complexity.
In telecommunication, intersymbol interference (ISI) is a form of distortion of a signal in which one symbol interferes with subsequent symbols. This is an unwanted phenomenon as the previous symbols have a similar effect as noise, thus making the communication less reliable. The spreading of the pulse beyond its allotted time interval causes it to interfere with neighboring pulses. ISI is usually caused by multipath propagation or the inherent linear or non-linear frequency response of a communication channel causing successive symbols to blur together.
The Viterbi algorithm is a dynamic programming algorithm for obtaining the maximum a posteriori probability estimate of the most likely sequence of hidden states—called the Viterbi path—that results in a sequence of observed events. This is done especially in the context of Markov information sources and hidden Markov models (HMM).
In information theory, turbo codes are a class of high-performance forward error correction (FEC) codes developed around 1990–91, but first published in 1993. They were the first practical codes to closely approach the maximum channel capacity or Shannon limit, a theoretical maximum for the code rate at which reliable communication is still possible given a specific noise level. Turbo codes are used in 3G/4G mobile communications and in satellite communications as well as other applications where designers seek to achieve reliable information transfer over bandwidth- or latency-constrained communication links in the presence of data-corrupting noise. Turbo codes compete with low-density parity-check (LDPC) codes, which provide similar performance.
Voice activity detection (VAD), also known as speech activity detection or speech detection, is the detection of the presence or absence of human speech, used in speech processing. The main uses of VAD are in speaker diarization, speech coding and speech recognition. It can facilitate speech processing, and can also be used to deactivate some processes during non-speech section of an audio session: it can avoid unnecessary coding/transmission of silence packets in Voice over Internet Protocol (VoIP) applications, saving on computation and on network bandwidth.
In 1953, IBM recognized the immediate application for what it termed a "Random Access File" having high capacity and rapid random access at a relatively low cost. After considering technologies such as wire matrices, rod arrays, drums, drum arrays, etc., the engineers at IBM's San Jose California laboratory invented the hard disk drive. The disk drive created a new level in the computer data hierarchy, then termed Random Access Storage but today known as secondary storage, less expensive and slower than main memory but faster and more expensive than tape drives.
Multiuser detection deals with demodulation of the mutually interfering digital streams of information that occur in areas such as wireless communications, high-speed data transmission, DSL, satellite communication, digital television, and magnetic recording. It is also being currently investigated for demodulation in low-power inter-chip and intra-chip communication. Multiuser detection encompasses both receiver technologies devoted to joint detection of all the interfering signals or to single-user receivers which are interested in recovering only one user but are robustified against multiuser interference and not just background noise.
Underwater acoustic communication is a technique of sending and receiving messages in water. There are several ways of employing such communication but the most common is by using hydrophones. Underwater communication is difficult due to factors such as multi-path propagation, time variations of the channel, small available bandwidth and strong signal attenuation, especially over long ranges. Compared to terrestrial communication, underwater communication has low data rates because it uses acoustic waves instead of electromagnetic waves.
Peter A. Franaszek is an American information theorist, an IEEE Fellow, a research staff member emeritus at the IBM T.J. Watson Research Center and a former member of the IBM Academy of Technology. He received his Sc.B. from Brown University in 1962, and his Ph.D. from Princeton University in 1966.
Noise-Predictive Maximum-Likelihood (NPML) is a class of digital signal-processing methods suitable for magnetic data storage systems that operate at high linear recording densities. It is used for retrieval of data recorded on magnetic media.
Evangelos Eleftheriou is a Greek electrical engineer. He is an IBM Fellow and was responsible for the Cloud and Computing Infrastructure department at the IBM Research – Zurich laboratory in Rüschlikon, Switzerland.
Yoshihiro Shiroishi was born in 1951 in Tokyo, Japan. He is a Chief Architect and Technical Advisor at the Hitachi Research & Dev. Group, Tokyo, Japan. Shiroishi was named Fellow of the Institute of Electrical and Electronics Engineers (IEEE) in 2015 for leadership in the development of high density magnetic recording technologies and devices.
Charles Hubert Coleman Jr. was an American electronic engineer and a pioneer in the field of color video tape recording and later in high data-rate digital tape recording. He was also an amateur explorer and avid pilot.
Christopher Henry Bajorek is a data storage engineer noted for his leadership in developing and implementing magnetoresistive sensors into magnetic stripe readers, tape drives and hard disk drives.
John C. Mallinson was a British physicist who made significant contributions to the understanding of magnetism and magnetic recording. He is perhaps best remembered for his theoretical work on structures with one-sided magnetic flux. Their use is exemplified in the Halbach array and in the familiar refrigerator magnet.
Two-dimensional magnetic recording (TDMR) is a technology introduced in 2017 in hard disk drives (HDD) used for computer data storage. Most of the world's data is recorded on HDDs, and there is continuous pressure on manufacturers to create greater data storage capacity in a given HDD form-factor and for a given cost. In an HDD, data is stored using magnetic recording on a rotating magnetic disk and is accessed through a write-head and read-head. TDMR allows greater storage capacity by advantageously combining signals simultaneously from multiple read-back heads to enhance the recovery of one or more data-tracks. In this manner, data can be stored with higher areal-density on the disks thus providing higher capacity in each HDD. TDMR is a read-back technology and thus applies equally well to future recording (writing) technologies such as heat-assisted magnetic recording (HAMR) and microwave-assisted magnetic recording (MAMR).
Mason Lamar Williams III was an engineer and physicist, noted for his contributions in the areas of magnetic recording and data storage on hard disk drives (HDD). A large part of his career was with the IBM Almaden Research Center in San Jose, California. After retiring, Williams played a major role in the restoration and demonstration of the IBM RAMAC at the Computer History Museum in Mountain View, California
Robert E Fontana is an engineer, physicist, and author who is noted for his contributions in the areas of magnetic recording and data storage on hard disk drives (HDD) and on digital tape recorders. His work has concentrated on developing thin film processing techniques for nano-fabrication of magnetic devices including Giant Magnetoresistance read heads now used universally in magnetic recording. Much of his career was with IBM in San Jose, California. He is a Fellow of the Institute of Electrical and Electronics Engineers and a member of the National Academy of Engineering.
Neal Bertram is a physicist noted for his contributions to the theory of magnetic recording. From 1968 to 1985, he worked for Ampex Corporation in Redwood City. From 1985 to 2004, he was an Endowed Chair Professor at the Center for Memory and Recording Research (CMRR), University of California at San Diego. He is the author of the book "Theory of Magnetic Recording". He is an elected Fellow of the Institute of Electrical and Electronics Engineers. In 2003, he won the IEEE Reynold B. Johnson Information Storage Systems Award.