Bishnu S. Atal

Last updated

Bishnu S. Atal
Born1933
India
Alma mater
Awards

Bishnu S. Atal (born 1933) is an Indian physicist and engineer. He is a noted researcher in acoustics, and is best known for developments in speech coding. He advanced linear predictive coding (LPC) during the late 1960s to 1970s, and developed code-excited linear prediction (CELP) with Manfred R. Schroeder in 1985.

Contents

In 1987, Atal was elected as a member into the National Academy of Engineering for innovative research in the area of linear predictive coding of speech.

Biography

Atal was born in India, and received his BS degree in physics (1952) from the University of Lucknow, a diploma in electrical communication engineering (1955) from the Indian Institute of Science, Bangalore and a PhD in electrical engineering (1968) from Brooklyn Polytechnic Institute.

From 1957 to 1960, he was a lecturer in acoustics at the Department of Electrical Communication Engineering, Indian Institute of Science, Bangalore.

In 1961 Atal joined Bell Laboratories, where his subsequent research focused on acoustics and speech, making major contributions in the field of speech analysis, synthesis, and coding, including low bit-rate speech coding and automatic speech recognition. He advanced and promoted linear predictive coding (1967), and developed code-excited linear prediction (1985) with Manfred R. Schroeder.

He retired in 2002 to become affiliate professor of Electrical Engineering at the University of Washington.

Atal holds more than 16 U.S. patents, and is a member of the National Academy of Engineering and National Academy of Sciences, and a fellow of the Acoustical Society of America and of the Institute of Electrical and Electronics Engineers. He received the 1986 IEEE Morris N. Liebmann Memorial Award "for pioneering contributions to linear predictive coding for speech processing", and the 1993 IEEE ASSP Society Award for contributions to linear prediction of speech, multipulse, and code-excited source coding. He is the Franklin Institute's 2003 Benjamin Franklin Medal Laureate in Engineering.

Selected works

Related Research Articles

Audio signal processing is a subfield of signal processing that is concerned with the electronic manipulation of audio signals. Audio signals are electronic representations of sound waves—longitudinal waves which travel through air, consisting of compressions and rarefactions. The energy contained in audio signals or sound level is typically measured in decibels. As audio signals may be represented in either digital or analog format, processing may occur in either domain. Analog processors operate directly on the electrical signal, while digital processors operate mathematically on its digital representation.

In information theory, data compression, source coding, or bit-rate reduction is the process of encoding information using fewer bits than the original representation. Any particular compression is either lossy or lossless. Lossless compression reduces bits by identifying and eliminating statistical redundancy. No information is lost in lossless compression. Lossy compression reduces bits by removing unnecessary or less important information. Typically, a device that performs data compression is referred to as an encoder, and one that performs the reversal of the process (decompression) as a decoder.

Speech processing is the study of speech signals and the processing methods of signals. The signals are usually processed in a digital representation, so speech processing can be regarded as a special case of digital signal processing, applied to speech signals. Aspects of speech processing includes the acquisition, manipulation, storage, transfer and output of speech signals. Different speech processing tasks include speech recognition, speech synthesis, speaker diarization, speech enhancement, speaker recognition, etc.

Linear predictive coding (LPC) is a method used mostly in audio signal processing and speech processing for representing the spectral envelope of a digital signal of speech in compressed form, using the information of a linear predictive model.

<span class="mw-page-title-main">Digital audio</span> Technology that records, stores, and reproduces sound

Digital audio is a representation of sound recorded in, or converted into, digital form. In digital audio, the sound wave of the audio signal is typically encoded as numerical samples in a continuous sequence. For example, in CD audio, samples are taken 44,100 times per second, each with 16-bit sample depth. Digital audio is also the name for the entire technology of sound recording and reproduction using audio signals that have been encoded in digital form. Following significant advances in digital audio technology during the 1970s and 1980s, it gradually replaced analog audio technology in many areas of audio engineering, record production and telecommunications in the 1990s and 2000s.

<span class="mw-page-title-main">Andrew Viterbi</span> Italian-American engineer, businessman

Andrew James Viterbi is an Italian Jewish–American electrical engineer and businessman who co-founded Qualcomm Inc. and invented the Viterbi algorithm. He is the Presidential Chair Professor of Electrical Engineering at the University of Southern California's Viterbi School of Engineering, which was named in his honor in 2004 in recognition of his $52 million gift.

Lawrence R. Rabiner is an electrical engineer working in the fields of digital signal processing and speech processing; in particular in digital signal processing for automatic speech recognition. He has worked on systems for AT&T Corporation for speech recognition.

Code-excited linear prediction (CELP) is a linear predictive speech coding algorithm originally proposed by Manfred R. Schroeder and Bishnu S. Atal in 1985. At the time, it provided significantly better quality than existing low bit-rate algorithms, such as residual-excited linear prediction (RELP) and linear predictive coding (LPC) vocoders. Along with its variants, such as algebraic CELP, relaxed CELP, low-delay CELP and vector sum excited linear prediction, it is currently the most widely used speech coding algorithm. It is also used in MPEG-4 Audio speech coding. CELP is commonly used as a generic term for a class of algorithms and not for a particular codec.

Line spectral pairs (LSP) or line spectral frequencies (LSF) are used to represent linear prediction coefficients (LPC) for transmission over a channel. LSPs have several properties that make them superior to direct quantization of LPCs. For this reason, LSPs are very useful in speech coding.

<span class="mw-page-title-main">Thomas Huang</span> Chinese-American engineer and computer scientist (1936–2020)

Thomas Shi-Tao Huang was a Chinese-born American computer scientist, electrical engineer, and writer. He was a researcher and professor emeritus at the University of Illinois at Urbana-Champaign (UIUC). Huang was one of the leading figures in computer vision, pattern recognition and human computer interaction.

Gerhard M. Sessler is a German inventor and scientist. He is Professor emeritus at the Department of Electrical Engineering and Information Technology of the Technische Universität Darmstadt.

James Loton Flanagan was an American electrical engineer. He was Rutgers University's vice president for research until 2004. He was also director of Rutgers' Center for Advanced Information Processing and the Board of Governors Professor of Electrical and Computer Engineering. He is known for co-developing adaptive differential pulse-code modulation (ADPCM) with P. Cummiskey and Nikil Jayant at Bell Labs.

Fumitada Itakura is a Japanese scientist. He did pioneering work in statistical signal processing, and its application to speech analysis, synthesis and coding, including the development of the linear predictive coding (LPC) and line spectral pairs (LSP) methods.

<span class="mw-page-title-main">Manfred R. Schroeder</span>

Manfred Robert Schroeder was a German physicist, most known for his contributions to acoustics and computer graphics. He wrote three books and published over 150 articles in his field.

The Medical Intelligence and Language Engineering Laboratory, also known as MILE lab, is a research laboratory at the Indian Institute of Science, Bangalore under the Department of Electrical Engineering. The lab is known for its work on Image processing, online handwriting recognition, Text-To-Speech and Optical character recognition systems, all of which are focused mainly on documents and speech in Indian languages. The lab is headed by A. G. Ramakrishnan.

<span class="mw-page-title-main">Georgios B. Giannakis</span> American computer scientist (born 1958)

Georgios B. Giannakis is a Greek-American Computer Scientist, engineer and inventor. He has been an Endowed Chair Professor of Wireless Telecommunications, he was Director of the Digital Technology Center, and at present he is a McKnight Presidential Chair with the Department of Electrical and Computer Engineering at the University of Minnesota.

<span class="mw-page-title-main">Audio coding format</span> Digitally coded format for audio signals

An audio coding format is a content representation format for storage or transmission of digital audio. Examples of audio coding formats include MP3, AAC, Vorbis, FLAC, and Opus. A specific software or hardware implementation capable of audio compression and decompression to/from a specific audio coding format is called an audio codec; an example of an audio codec is LAME, which is one of several different codecs which implements encoding and decoding audio in the MP3 audio coding format in software.

Nikil S. Jayant is an Indian-American communications engineer. He was a researcher at Bell Laboratories and subsequently a professor at Georgia Institute of Technology. He received his Ph.D. in Electrical Communication Engineering from the Indian Institute of Science, Bangalore, India in 1970.

Ranjan Kumar Mallik is an Indian electrical and communications engineer and a professor at the Department of Electrical Engineering of the Indian Institute of Technology, Delhi. He held the Jai Gupta Chair at IIT Delhi from 2007 to 2012 and the Brigadier Bhopinder Singh Chair from 2012 to 2017. He is known for his researches on multiple-input multi-output systems and is an elected fellow of all the three major Indian science academies viz. Indian Academy of Sciences, Indian National Science Academy, and The National Academy of Sciences, India. He is also an elected fellow of The World Academy of Sciences, Indian National Academy of Engineering, and The Institute of Electrical and Electronics Engineers, Inc.

<span class="mw-page-title-main">John Makhoul</span> American computer scientist

John Makhoul is a Lebanese-American computer scientist who works in the field of speech and language processing. Dr. Makhoul's work on linear predictive coding was used in the establishment of the Network Voice Protocol, which enabled the transmission of speech signals over the ARPANET. Makhoul is recognized in the field for his vital role in the areas of speech and language processing, including speech analysis, speech coding, speech recognition and speech understanding. He has made a number of significant contributions to the mathematical modeling of speech signals, including his work on linear prediction, and vector quantization. His patented work on the direct application of speech recognition techniques for accurate, language-independent optical character recognition (OCR) has had a dramatic impact on the ability to create OCR systems in multiple languages relatively quickly.

References