Acoustic camera

Last updated

An acoustic camera (or noise camera) is an imaging device used to locate sound sources and to characterize them. It consists of a group of microphones, also called a microphone array, from which signals are simultaneously collected and processed to form a representation of the location of the sound sources.

Contents

Terminology

The term acoustic camera has first appeared at the end of the 19th century: A physiologist, J.R. Ewald, [1] was investigating the function of the inner ear and introduced an analogy with the Chladni plates (a domain nowadays called Cymatics), a device enabling to visually see the modes of vibration of a plate. He called this device an acoustic camera. The term has then been widely used during the 20th century [2] [3] [4] to designate various types of acoustic devices, such as underwater localization systems [5] or active systems used in medicine. [6] It designates nowadays any transducer array used to localize sound sources (the medium is usually the air), especially when coupled with an optical camera.

Technology

General principles

An acoustic camera generally consists of a microphone array and optionally an optical camera. The microphones – analog or digital – are acquired simultaneously or with known relative time delays to be able to use the phase difference between the signals. As the sound propagates in the medium (air, water...) at a finite known speed, a sound source is perceived by the microphones at different time instants and at different sound intensities that depend on both the sound source location and the microphone location. One popular method to obtain an acoustic image from the measurement of the microphone is to use beamforming: By delaying each microphone signal relatively and adding them, the signal coming from a specific direction is amplified while signals coming from other directions are canceled. The power of this resulting signal is then calculated and reported on a power map at a pixel corresponding to the direction . The process is iterated at each direction where the power needs to be computed.

While this method has many advantages – robustness, easy to understand, highly parallelizable because each direction can be computed independently, versatile (there exist many types of beamformers to include various types of hypothesis), relatively fast – it also has some drawbacks: the produced acoustic map has artifacts (also called side lobes or ghost sources) and it does not model correctly correlated sound sources. Various methods have been introduced to reduce the artifacts such as DAMAS [7] or to take in account correlated sources such as CLEAN-SC, [8] both at the price of a higher computational cost.

When the sound sources are near the acoustic camera, the relative intensity perceived by the different microphones as well as the waves not being any more seen as planar but spherical by the acoustic camera add new information compared to the case of sources being far from the camera. It enables to use more effective methods such as acoustic holography.

Reprojection

Results of far-field beamforming can be reprojected onto planar or non-planar surfaces.

Two-dimensional

Some acoustic cameras use two-dimensional acoustic mapping, which uses a unidirectional microphone array (e.g. a rectangle of microphones, all facing the same direction). Two-dimensional acoustic mapping works best when the surface to be examined is planar and the acoustic camera can be set up facing the surface perpendicularly. However, the surfaces of real-world objects are not often flat, and it is not always possible to optimally position the acoustic camera. [9]

Additionally, the two-dimensional method of acoustic mapping introduces error into the calculations of the sound intensity at a point. Two-dimensional mapping approximates three-dimensional surfaces into a plane, allowing the distance between each microphone and the focus point to be calculated relatively easily. However, this approximation ignores the distance differences caused by surfaces having different depths at different points. In most applications of the acoustic camera, this error is small enough to be ignored; however, in confined spaces, the error becomes significant. [9]

Three-dimensional

Three-dimensional acoustic cameras fix the errors of two-dimensional cameras by taking into account surface depths, and therefore correctly measuring the distances between the microphone and each spatial point. These cameras produce a more accurate picture, but require a 3-D model of the object or space being analyzed. Additionally, if the acoustic camera picks up sound from a point in space that is not part of the model, the sound may be mapped to a random space in the model, or the sound may not show up at all. 3-D acoustic cameras can also be used to analyze confined spaces, such as room interiors; however, in order to do this, a microphone array that is omnidirectional (e.g. a sphere of microphones, each facing a different direction) is required. This is in addition to the first requirement of having a 3-D model. [9]

Applications

There are many applications of the acoustic camera, with most focusing on noise reduction. The camera is frequently applied to improve the noise emission of vehicles (such as cars, airplanes [10] ),trains, structures—such as wind turbines [11] and heavy machinery operations such as mining [12] or drilling.

Acoustic cameras are not only used to measure the exterior emission of products but also to improve the comfort inside cabins of cars, [9] train or airplanes. Spherical acoustic cameras are preferred in this type of application because the three-dimensional placement of the microphone allows to localize sound sources in all directions.

Troubleshooting of faults that occur in machines and mechanical parts can be accomplished with an acoustic camera. To find where the problem lies, the sound mapping of a properly functional machine can be compared to one of a dysfunctional machine.

A similar setup of the acoustic camera can be used to study the noise inside passenger carts during train operation. Alternatively, the camera can be set up outside, in an area near the train tracks, to observe the train as it goes by. This can give another perspective of the noise that might be heard inside the train. Additionally, an outside setup can be used to examine the squealing of train wheels caused by a curve in the tracks.

Acoustic camera may be for to aid legal enforcement of noise nuisances caused by people or motor vehicles. Epidemiologist Erica Walker has said this is a "lazy" solution to the problem of noise, and expressed concern acoustic cameras could be used to over-police ethic minority neighbourhoods. [13]

Challenges

Dynamic range

The effective dynamic range in the imaging plane can be interpreted as the maximum contrast achievable within the target area. An inherent challenge related to the dynamic range of acoustic cameras lies in its dependency on the sound's wavelength and the size of the array. These physical constraints pose difficulties for far-field acoustic cameras aiming to resolve multiple low-frequency sources. As the aperture size would need to be significantly large to tackle low-frequency issues, it often results in inconclusive or less definitive results within this frequency range. This underlines the unique challenges faced in enhancing the dynamic range of acoustic cameras, particularly in applications involving low-frequency sounds.

Low frequencies in the far-field

The lowest frequency that can be localized with a far-field acoustic camera is primarily determined by the size of the array's aperture (its largest dimension). Challenges arise when dealing with low-frequency issues, particularly those below 300 Hz, as they require large array sizes for effective sound source localization. Alternatively, there are a number of effective solutions, such as acoustic vector sensors, either standalone or in an array configuration, or near-field acoustic cameras, both can serve as valuable tools for addressing non-stationary issues. On the other hand, methods that employ direct sound mapping using sound intensity probes and/or particle velocity probes offer robust alternatives for identifying and visualizing time-stationary sound sources. [14]

Computational power

The signal processing required by the acoustic camera is very intensive and needs powerful hardware and plenty of memory storage. Because of this, signal processing is frequently done after the recording of data, which can hinder or prevent the use of the camera in analyzing sounds that only occur occasionally or at varying locations. Cameras that do perform signal processing in real time tend to be large and expensive. Hardware and signal processing improvements can help to overcome these difficulties. Signal processing optimizations often focus on reduction of computational complexity, storage requirements, and memory bandwidth (rate of data consumption). [15]

Related Research Articles

<span class="mw-page-title-main">Sonar</span> Acoustic sensing method

Sonar is a technique that uses sound propagation to navigate, measure distances (ranging), communicate with or detect objects on or under the surface of the water, such as other vessels.

<span class="mw-page-title-main">Microphone array</span> Group of microphones operating in tandem

A microphone array is any number of microphones operating in tandem. There are many applications:

<span class="mw-page-title-main">Microphone</span> Device that converts sound into an electrical signal

A microphone, colloquially called a mic, is a transducer that converts sound into an electrical signal. Microphones are used in many applications such as telephones, hearing aids, public address systems for concert halls and public events, motion picture production, live and recorded audio engineering, sound recording, two-way radios, megaphones, and radio and television broadcasting. They are also used in computers and other electronic devices, such as mobile phones, for recording sounds, speech recognition, VoIP, and other purposes, such as ultrasonic sensors or knock sensors.

<span class="mw-page-title-main">Binaural recording</span> Method of recording sound

Binaural recording is a method of recording sound that uses two microphones, arranged with the intent to create a 3-D stereo sound sensation for the listener of actually being in the room with the performers or instruments. This effect is often created using a technique known as dummy head recording, wherein a mannequin head is fitted with a microphone in each ear. Binaural recording is intended for replay using headphones and will not translate properly over stereo speakers. This idea of a three-dimensional or "internal" form of sound has also translated into useful advancement of technology in many things such as stethoscopes creating "in-head" acoustics and IMAX movies being able to create a three-dimensional acoustic experience.

<span class="mw-page-title-main">Ambisonics</span> Full-sphere surround sound format

Ambisonics is a full-sphere surround sound format: in addition to the horizontal plane, it covers sound sources above and below the listener.

<span class="mw-page-title-main">Surround sound</span> System with loudspeakers that surround the listener

Surround sound is a technique for enriching the fidelity and depth of sound reproduction by using multiple audio channels from speakers that surround the listener. Its first application was in movie theaters. Prior to surround sound, theater sound systems commonly had three screen channels of sound that played from three loudspeakers located in front of the audience. Surround sound adds one or more channels from loudspeakers to the side or behind the listener that are able to create the sensation of sound coming from any horizontal direction around the listener.

<span class="mw-page-title-main">Spectrogram</span> Visual representation of the spectrum of frequencies of a signal as it varies with time

A spectrogram is a visual representation of the spectrum of frequencies of a signal as it varies with time. When applied to an audio signal, spectrograms are sometimes called sonographs, voiceprints, or voicegrams. When the data are represented in a 3D plot they may be called waterfall displays.

Sound localization is a listener's ability to identify the location or origin of a detected sound in direction and distance.

Beamforming or spatial filtering is a signal processing technique used in sensor arrays for directional signal transmission or reception. This is achieved by combining elements in an antenna array in such a way that signals at particular angles experience constructive interference while others experience destructive interference. Beamforming can be used at both the transmitting and receiving ends in order to achieve spatial selectivity. The improvement compared with omnidirectional reception/transmission is known as the directivity of the array.

<span class="mw-page-title-main">Acoustic location</span> Use of reflected sound waves to locate objects

Acoustic location is a method of determining the position of an object or sound source by using sound waves. Location can take place in gases, liquids, and in solids.

Analysis of sound and acoustics plays a role in such engineering tasks as product design, production test, machine performance, and process control. For instance, product design can require modification of sound level or noise for compliance with standards from ANSI, IEC, and ISO. The work might also involve design fine-tuning to meet market expectations. Here, examples include tweaking an automobile door latching mechanism to impress a consumer with a satisfying click or modifying an exhaust manifold to change the tone of an engine's rumble. Aircraft designers are also using acoustic instrumentation to reduce the noise generated on takeoff and landing.

Noise, vibration, and harshness (NVH), also known as noise and vibration (N&V), is the study and modification of the noise and vibration characteristics of vehicles, particularly cars and trucks. While noise and vibration can be readily measured, harshness is a subjective quality, and is measured either via jury evaluations, or with analytical tools that can provide results reflecting human subjective impressions. The latter tools belong to the field psychoacoustics.

<span class="mw-page-title-main">Underwater acoustics</span> Study of the propagation of sound in water

Underwater acoustics is the study of the propagation of sound in water and the interaction of the mechanical waves that constitute sound with the water, its contents and its boundaries. The water may be in the ocean, a lake, a river or a tank. Typical frequencies associated with underwater acoustics are between 10 Hz and 1 MHz. The propagation of sound in the ocean at frequencies lower than 10 Hz is usually not possible without penetrating deep into the seabed, whereas frequencies above 1 MHz are rarely used because they are absorbed very quickly.

Ambiophonics is a method in the public domain that employs digital signal processing (DSP) and two loudspeakers directly in front of the listener in order to improve reproduction of stereophonic and 5.1 surround sound for music, movies, and games in home theaters, gaming PCs, workstations, or studio monitoring applications. First implemented using mechanical means in 1986, today a number of hardware and VST plug-in makers offer Ambiophonic DSP. Ambiophonics eliminates crosstalk inherent in the conventional stereo triangle speaker placement, and thereby generates a speaker-binaural soundfield that emulates headphone-binaural sound, and creates for the listener improved perception of reality of recorded auditory scenes. A second speaker pair can be added in back in order to enable 360° surround sound reproduction. Additional surround speakers may be used for hall ambience, including height, if desired.

<span class="mw-page-title-main">Thermoacoustic imaging</span>

Thermoacoustic imaging was originally proposed by Theodore Bowen in 1981 as a strategy for studying the absorption properties of human tissue using virtually any kind of electromagnetic radiation. But Alexander Graham Bell first reported the physical principle upon which thermoacoustic imaging is based a century earlier. He observed that audible sound could be created by illuminating an intermittent beam of sunlight onto a rubber sheet. Shortly after Bowen's work was published, other researchers proposed methodology for thermoacoustic imaging using microwaves. In 1994 researchers used an infrared laser to produce the first thermoacoustic images of near-infrared optical absorption in a tissue-mimicking phantom, albeit in two dimensions (2D). In 1995 other researchers formulated a general reconstruction algorithm by which 2D thermoacoustic images could be computed from their "projections," i.e. thermoacoustic computed tomography (TCT). By 1998 researchers at Indiana University Medical Center extended TCT to 3D and employed pulsed microwaves to produce the first fully three-dimensional (3D) thermoacoustic images of biologic tissue [an excised lamb kidney ]. The following year they created the first fully 3D thermoacoustic images of cancer in the human breast, again using pulsed microwaves. Since that time, thermoacoustic imaging has gained widespread popularity in research institutions worldwide. As of 2008, three companies were developing commercial thermoacoustic imaging systems – Seno Medical, Endra, Inc. and OptoSonics, Inc.

<span class="mw-page-title-main">Brüel & Kjær</span> Danish multinational company

Brüel & Kjær is a Danish multinational engineering and electronics company headquartered in Nærum, near Copenhagen. It was the largest producer in the world of equipment for acoustic and vibrational measurements. Brüel & Kjær is a subsidiary of Spectris.

A sound intensity probe is a probe capable of measuring sound intensity. The sound intensity is the product of sound pressure and particle velocity. Common probes measure the sound intensity using two microphones. Another type measures the sound intensity using a microphone and a particle velocity probe.

3D sound localization refers to an acoustic technology that is used to locate the source of a sound in a three-dimensional space. The source location is usually determined by the direction of the incoming sound waves and the distance between the source and sensors. It involves the structure arrangement design of the sensors and signal processing techniques.

3D sound reconstruction is the application of reconstruction techniques to 3D sound localization technology. These methods of reconstructing three-dimensional sound are used to recreate sounds to match natural environments and provide spatial cues of the sound source. They also see applications in creating 3D visualizations on a sound field to include physical aspects of sound waves including direction, pressure, and intensity. This technology is used in entertainment to reproduce a live performance through computer speakers. The technology is also used in military applications to determine location of sound sources. Reconstructing sound fields is also applicable to medical imaging to measure points in ultrasound.

<span class="mw-page-title-main">United Kingdom patent 394325</span> Seminal work on stereophonic sound by Alan Blumlein

The United Kingdom patent 394325 'Improvements in and relating to Sound-transmission, Sound-recording and Sound-reproducing Systems' is a fundamental work on stereophonic sound, written by Alan Blumlein in 1931 and published in 1933. The work exists only in the form of a patent and two accompanying memos addressed to Isaac Shoenberg. The text is exceptionally long for a patent of the period, having 70 numbered claims. It contains a brief summary of sound localization theory, a roadmap for introduction of surround sound in sound film and recording industry, and a description of Blumlein's inventions related to stereophony, notably the matrix processing of stereo signals, the Blumlein stereo microphone and the 45/45 mechanical recording system.

References

  1. Ewald, J.R. (1898). Wiener klinische Wochenschrift. 11: 721.{{cite journal}}: CS1 maint: untitled periodical (link)
  2. Whitman, R. L.; Ahmed, M.; Korpel, A. (1972). "A progress report on the laser scanned acoustic camera". Acoustical Holography. Springer US. 20: 11–32. doi:10.1007/978-1-4615-8213-7_2. ISBN   978-1-4615-8215-1.
  3. USpatent 3895340,"Acoustic camera apparatus"
  4. Hansen, Rolf Kahrs; Andersen, Poul Arndt (1993). "3D acoustic camera for underwater imaging". Acoustical Imaging. Springer US. 20: 723–727. doi:10.1007/978-1-4615-2958-3_98. ISBN   978-1-4613-6286-9.
  5. Haslett, R. W. G.; Pearce, G.; Welsh, A. W.; Hussey, K. (1966). "The underwater acoustic camera". Acta Acustica United with Acustica. S. Hirzel Verlag. 17, 4: 187–203.
  6. Maginness, M. G.; Plummer, J. D.; Meindl, J. D. (1974). "An acoustic image sensor using a transmit-receive array". Acoustical Holography. Springer US: 619–631. doi:10.1007/978-1-4757-0827-1_36. ISBN   978-1-4757-0829-5.
  7. Brooks, Thomas F.; Humphreys, William M. (2004). "Deconvolution Approach for the Mapping of Acoustic Sources". NASA Invention Disclosure. NASA Langley Research. LAR-16907-1.
  8. Sijtsma, P. (2007). "CLEAN based on spatial source coherence". International Journal of Aeroacoustics. 6 (4): 357–374. doi:10.1260/147547207783359459. S2CID   122396368.
  9. 1 2 3 4 Meyer, Andy, and Döbler, Dirk. "Noise source localization within a car interior using 3D-microphone arrays." Proceedings of the BeBeC (2006).
  10. Leon, Brusniak; Underbrink, James R.; Stoker, Robert W. (2006). "Acoustic imaging of aircraft noise sources using large aperture phased arrays". AIAA/CEAS Aeroacoustics Conference. 12.
  11. Gwang-Se, Lee; Cheong, Cheolung; Shin, Su-Hyun; Jung, Sung-Soo (2012). "A case study of localization and identification of noise sources from a pitch and a stall regulated wind turbine". Applied Acoustics. 73 8: 817–827.
  12. Oberholster, Abrie J. "Solving issues in the mining industry". Siemens Digital Industries Software. University Of Pretoria. Retrieved November 12, 2021.
  13. Demopoulos A (4 October 2023). "Honk honk! Can noise cameras reduce 'potentially fatal' sound pollution?". The Guardian.
  14. Sound Source Localization Solutions
  15. Zimmermann, B.; Studer, C., "FPGA-based real-time acoustic camera prototype," Circuits and Systems (ISCAS), Proceedings of 2010 IEEE International Symposium on , vol., no., pp.1419,1419, May 30 2010-June 2 2010

Further reading