Interaural time difference

Last updated
Interaural time difference (ITD) between left (top) and right (bottom) ears. (sound source: 100 ms white noise from 90deg azimuth, 0deg elevation) Interaural Diferencia de tiempo (ITD) entre izquierdo (inferior) y correcto (superior) orejas.jpg
Interaural time difference (ITD) between left (top) and right (bottom) ears. (sound source: 100 ms white noise from 90° azimuth, 0° elevation)

The interaural time difference (or ITD) when concerning humans or animals, is the difference in arrival time of a sound between two ears. It is important in the localization of sounds, as it provides a cue to the direction or angle of the sound source from the head. If a signal arrives at the head from one side, the signal has further to travel to reach the far ear than the near ear. This pathlength difference results in a time difference between the sound's arrivals at the ears, which is detected and aids the process of identifying the direction of sound source.

Contents

When a signal is produced in the horizontal plane, its angle in relation to the head is referred to as its azimuth, with 0 degrees (0°) azimuth being directly in front of the listener, 90° to the right, and 180° being directly behind.

Different methods for measuring ITDs

Duplex theory

The Duplex theory proposed by Lord Rayleigh (1907) provides an explanation for the ability of humans to localise sounds by time differences between the sounds reaching each ear (ITDs) and differences in sound level entering the ears (interaural level differences, ILDs). But there still lies a question whether ITD or ILD is prominent.

The duplex theory states that ITDs are used to localise low frequency sounds, in particular, while ILDs are used in the localisation of high frequency sound inputs. However, the frequency ranges for which the auditory system can use ITDs and ILDs significantly overlap, and most natural sounds will have both high and low frequency components, so that the auditory system will in most cases have to combine information from both ITDs and ILDs to judge the location of a sound source. [1] A consequence of this duplex system is that it is also possible to generate so-called "cue trading" or "time–intensity trading" stimuli on headphones, where ITDs pointing to the left are offset by ILDs pointing to the right, so the sound is perceived as coming from the midline. A limitation of the duplex theory is that the theory does not completely explain directional hearing, as no explanation is given for the ability to distinguish between a sound source directly in front and behind. Also the theory only relates to localising sounds in the horizontal plane around the head. The theory also does not take into account the use of the pinna in localisation.(Gelfand, 2004)

Experiments conducted by Woodworth (1938) tested the duplex theory by using a solid sphere to model the shape of the head and measuring the ITDs as a function of azimuth for different frequencies. The model used had a distance between the two ears of approximately 22–23 cm. Initial measurements found that there was a maximum time delay of approximately 660 μs when the sound source was placed at directly 90° azimuth to one ear. This time delay correlates to the wavelength of a sound input with a frequency of 1500 Hz. The results concluded that when a sound played had a frequency less than 1500 Hz the wavelength is greater than this maximum time delay between the ears. Therefore, there is a phase difference between the sound waves entering the ears providing acoustic localisation cues. With a sound input with a frequency closer to 1500 Hz the wavelength of the sound wave is similar to the natural time delay. Therefore, due to the size of the head and the distance between the ears there is a reduced phase difference so localisations errors start to be made. When a high frequency sound input is used with a frequency greater than 1500 Hz, the wavelength is shorter than the distance between the 2 ears, a head shadow is produced and ILD provide cues for the localisation of this sound.

Feddersen et al. (1957) also conducted experiments taking measurements on how ITDs alter with changing the azimuth of the loudspeaker around the head at different frequencies. But unlike the Woodworth experiments human subjects were used rather than a model of the head. The experiment results agreed with the conclusion made by Woodworth about ITDs. The experiments also concluded that is there is no difference in ITDs when sounds are provided from directly in front or behind at 0° and 180° azimuth. The explanation for this is that the sound is equidistant from both ears. Interaural time differences alter as the loudspeaker is moved around the head. The maximum ITD of 660 μs occurs when a sound source is positioned at 90° azimuth to one ear.

Current findings

Starting in 1948, the prevailing theory on interaural time differences centered on the idea that inputs from the medial superior olive differentially process inputs from the ipsilateral and contralateral side relative to the sound. This is accomplished through a discrepancy in arrival time of excitatory inputs into the medial superior olive, based on differential conductance of the axons, which allows both sounds to ultimately converge at the same time through neurons with complementary intrinsic properties.

Franken et al., attempted to further elucidate the mechanisms underlying ITD in mammalian brains. [2] One experiment they performed was to isolate discrete inhibitory post-synaptic potentials to try and determine if inhibitory inputs to the superior olive were allowing the faster excitatory input to delay firing until the two signals were synced. However, after blocking EPSPs with a glutamate receptor blocker, they determine that the size of inhibitory inputs was too marginal to appear to play a significant role in phase locking. This was verified when the experimenters blocked inhibitory input and still saw clear phase locking of the excitatory inputs in their absence. This led them to the theory that in-phase excitatory inputs are summated such that the brain can process sound localization by counting the number of action potentials that arise from various magnitudes of summated depolarization.

Franken, et al. also examined anatomical and functional patterns within the superior olive to clarify previous theories about the rostrocaudal axis serving as a source of tonotopy. Their results showed a significant correlation between tuning frequency and relative position along the dorsoventral axis, while they saw no distinguishable pattern of tuning frequency on the rostrocaudal axis.

Lastly, they went on to further explore the driving forces behind the interaural time difference, specifically whether the process is simply the alignment of inputs that is processed by a coincidence detector, or whether the process is more complicated. Evidence from Franken, et al. shows that the processing is affected by inputs that precede the binaural signal, which would alter the functioning of voltage-gated sodium and potassium channels to shift the membrane potential of the neuron. Furthermore, the shift is dependent on the frequency tuning of each neuron, ultimately creating a more complex confluence and analysis of sound. Franken, et al.'s findings provide several pieces of evidence that contradict existing theories about binaural audition.

The anatomy of the ITD pathway

The auditory nerve fibres, known as the afferent nerve fibres, carry information from the organ of Corti to the brainstem and brain. Auditory afferent fibres consist of two types of fibres called type I and type II fibres. Type I fibres innervate the base of one or two inner hair cells and Type II fibres innervate the outer hair cells. Both leave the organ of Corti through an opening called the habenula perforata. The type I fibres are thicker than the type II fibres and may also differ in how they innervate the inner hair cells. Neurons with large calyceal endings ensure preservation of timing information throughout the ITD pathway.

Next in the pathway is the cochlear nucleus, which receives mainly ipsilateral (that is, from the same side) afferent input. The cochlear nucleus has three distinct anatomical divisions, known as the antero-ventral cochlear nucleus (AVCN), postero-ventral cochlear nucleus (PVCN) and dorsal cochlear nucleus (DCN) and each have different neural innervations.

The AVCN contains predominant bushy cells, with one or two profusely branching dendrites; it is thought that bushy cells may process the change in the spectral profile of complex stimuli. The AVCN also contain cells with more complex firing patterns than bushy cells called multipolar cells, these cells have several profusely branching dendrites and irregular shaped cell bodies. Multipolar cells are sensitive to changes in acoustic stimuli and in particular, onset and offset of sounds, as well as changes in intensity and frequency. The axons of both cell types leave the AVCN as large tract called the ventral acoustic stria, which forms part of the trapezoid body and travels to the superior olivary complex.

A group of nuclei in pons make up the superior olivary complex (SOC). This is the first stage in auditory pathway to receive input from both cochleas, which is crucial for our ability to localise the sounds source in the horizontal plane. The SOC receives input from cochlear nuclei, primarily the ipsilateral and contralateral AVCN. Four nuclei make up the SOC but only the medial superior olive (MSO) and the lateral superior olive (LSO) receive input from both cochlear nuclei.

The MSO is made up of neurons which receive input from the low-frequency fibers of the left and right AVCN. The result of having input from both cochleas is an increase in the firing rate of the MSO units. The neurons in the MSO are sensitive to the difference in the arrival time of sound at each ear, also known as the interaural time difference (ITD). Research shows that if stimulation arrives at one ear before the other, many of the MSO units will have increased discharge rates. The axons from the MSO continue to higher parts of the pathway via the ipsilateral lateral lemniscus tract.(Yost, 2000)

The lateral lemniscus (LL) is the main auditory tract in the brainstem connecting SOC to the inferior colliculus. The dorsal nucleus of the lateral lemniscus (DNLL) is a group of neurons separated by lemniscus fibres, these fibres are predominantly destined for the inferior colliculus (IC). In studies using an unanesthetized rabbit the DNLL was shown to alter the sensitivity of the IC neurons and may alter the coding of interaural timing differences (ITDs) in the IC.(Kuwada et al., 2005) The ventral nucleus of the lateral lemniscus (VNLL) is a chief source of input to the inferior colliculus. Research using rabbits shows the discharge patterns, frequency tuning and dynamic ranges of VNLL neurons supply the inferior colliculus with a variety of inputs, each enabling a different function in the analysis of sound.(Batra & Fitzpatrick, 2001) In the inferior colliculus (IC) all the major ascending pathways from the olivary complex and the central nucleus converge. The IC is situated in the midbrain and consists of a group of nuclei the largest of these is the central nucleus of inferior colliculus (CNIC). The greater part of the ascending axons forming the lateral lemniscus will terminate in the ipsilateral CNIC however a few follow the commissure of Probst and terminate on the contralateral CNIC. The axons of most of the CNIC cells form the brachium of IC and leave the brainstem to travel to the ipsilateral thalamus. Cells in different parts of the IC tend to be monaural, responding to input from one ear, or binaural and therefore respond to bilateral stimulation.

The spectral processing that occurs in the AVCN and the ability to process binaural stimuli, as seen in the SOC, are replicated in the IC. Lower centres of the IC extract different features of the acoustic signal such as frequencies, frequency bands, onsets, offsets, changes in intensity and localisation. The integration or synthesis of acoustic information is thought to start in the CNIC.(Yost, 2000)

Effect of a hearing loss

A number of studies have looked into the effect of hearing loss on interaural time differences. In their review of localisation and lateralisation studies, Durlach, Thompson, and Colburn (1981), cited in Moore (1996) found a "clear trend for poor localization and lateralization in people with unilateral or asymmetrical cochlear damage." This is due to the difference in performance between the two ears. In support of this, they did not find significant localisation problems in individuals with symmetrical cochlear losses. In addition to this, studies have been conducted into the effect of hearing loss on the threshold for interaural time differences. The normal human threshold for detection of an ITD is up to a time difference of 10μs (microseconds). Studies by Gabriel, Koehnke, & Colburn (1992), Häusler, Colburn, & Marr (1983) and Kinkel, Kollmeier, & Holube (1991)(cited by Moore, 1996) have shown that there can be great differences between individuals regarding binaural performance. It was found that unilateral or asymmetric hearing losses can increase the threshold of ITD detection in patients. This was also found to apply to individuals with symmetrical hearing losses when detecting ITDs in narrowband signals. However, ITD thresholds seem to be normal for those with symmetrical losses when listening to broadband sounds.

See also

Related Research Articles

<span class="mw-page-title-main">Auditory system</span> Sensory system used for hearing

The auditory system is the sensory system for the sense of hearing. It includes both the sensory organs and the auditory parts of the sensory system.

<span class="mw-page-title-main">Lateral lemniscus</span> Brain structure

The lateral lemniscus is a tract of axons in the brainstem that carries information about sound from the cochlear nucleus to various brainstem nuclei and ultimately the contralateral inferior colliculus of the midbrain. Three distinct, primarily inhibitory, cellular groups are located interspersed within these fibers, and are thus named the nuclei of the lateral lemniscus.

Sound localization is a listener's ability to identify the location or origin of a detected sound in direction and distance.

<span class="mw-page-title-main">Inferior colliculus</span> Midbrain structure involved in the auditory pathway

The inferior colliculus (IC) is the principal midbrain nucleus of the auditory pathway and receives input from several peripheral brainstem nuclei in the auditory pathway, as well as inputs from the auditory cortex. The inferior colliculus has three subdivisions: the central nucleus, a dorsal cortex by which it is surrounded, and an external cortex which is located laterally. Its bimodal neurons are implicated in auditory-somatosensory interaction, receiving projections from somatosensory nuclei. This multisensory integration may underlie a filtering of self-effected sounds from vocalization, chewing, or respiration activities.

<span class="mw-page-title-main">Medial geniculate nucleus</span>

The medial geniculate nucleus (MGN) or medial geniculate body (MGB) is part of the auditory thalamus and represents the thalamic relay between the inferior colliculus (IC) and the auditory cortex (AC). It is made up of a number of sub-nuclei that are distinguished by their neuronal morphology and density, by their afferent and efferent connections, and by the coding properties of their neurons. It is thought that the MGN influences the direction and maintenance of attention.

<span class="mw-page-title-main">Cochlear nerve</span> Nerve carrying auditory information from the inner ear to the brain

The cochlear nerve is one of two parts of the vestibulocochlear nerve, a cranial nerve present in amniotes, the other part being the vestibular nerve. The cochlear nerve carries auditory sensory information from the cochlea of the inner ear directly to the brain. The other portion of the vestibulocochlear nerve is the vestibular nerve, which carries spatial orientation information to the brain from the semicircular canals, also known as semicircular ducts.

Virtual acoustic space (VAS), also known as virtual auditory space, is a technique in which sounds presented over headphones appear to originate from any desired direction in space. The illusion of a virtual sound source outside the listener's head is created.

<span class="mw-page-title-main">Cochlear nucleus</span> Two cranial nerve nuclei of the human brainstem

The cochlear nuclear (CN) complex comprises two cranial nerve nuclei in the human brainstem, the ventral cochlear nucleus (VCN) and the dorsal cochlear nucleus (DCN). The ventral cochlear nucleus is unlayered whereas the dorsal cochlear nucleus is layered. Auditory nerve fibers, fibers that travel through the auditory nerve carry information from the inner ear, the cochlea, on the same side of the head, to the nerve root in the ventral cochlear nucleus. At the nerve root the fibers branch to innervate the ventral cochlear nucleus and the deep layer of the dorsal cochlear nucleus. All acoustic information thus enters the brain through the cochlear nuclei, where the processing of acoustic information begins. The outputs from the cochlear nuclei are received in higher regions of the auditory brainstem.

<span class="mw-page-title-main">Dorsal cochlear nucleus</span>

The dorsal cochlear nucleus is a cortex-like structure on the dorso-lateral surface of the brainstem. Along with the ventral cochlear nucleus (VCN), it forms the cochlear nucleus (CN), where all auditory nerve fibers from the cochlea form their first synapses.

<span class="mw-page-title-main">Superior olivary complex</span> Collection of brainstem nuclei related to hearing

The superior olivary complex (SOC) or superior olive is a collection of brainstem nuclei that is located in pons, functions in multiple aspects of hearing and is an important component of the ascending and descending auditory pathways of the auditory system. The SOC is intimately related to the trapezoid body: most of the cell groups of the SOC are dorsal to this axon bundle while a number of cell groups are embedded in the trapezoid body. Overall, the SOC displays a significant interspecies variation, being largest in bats and rodents and smaller in primates.

Eric Knudsen is a professor of neurobiology at Stanford University. He is best known for his discovery, along with Masakazu Konishi, of a brain map of sound location in two dimensions in the barn owl, tyto alba. His work has contributed to the understanding of information processing in the auditory system of the barn owl, the plasticity of the auditory space map in developing and adult barn owls, the influence of auditory and visual experience on the space map, and more recently, mechanisms of attention and learning. He is a recipient of the Lashley Award, the Gruber Prize in Neuroscience, and the Newcomb Cleveland prize and is a member of the National Academy of Sciences.

Binaural fusion or binaural integration is a cognitive process that involves the combination of different auditory information presented binaurally, or to each ear. In humans, this process is essential in understanding speech as one ear may pick up more information about the speech stimuli than the other.

<span class="mw-page-title-main">Ventral cochlear nucleus</span>

In the ventral cochlear nucleus (VCN), auditory nerve fibers enter the brain via the nerve root in the VCN. The ventral cochlear nucleus is divided into the anterior ventral (anteroventral) cochlear nucleus (AVCN) and the posterior ventral (posteroventral) cochlear nucleus (PVCN). In the VCN, auditory nerve fibers bifurcate, the ascending branch innervates the AVCN and the descending branch innervates the PVCN and then continue to the dorsal cochlear nucleus. The orderly innervation by auditory nerve fibers gives the AVCN a tonotopic organization along the dorsoventral axis. Fibers that carry information from the apex of the cochlea that are tuned to low frequencies contact neurons in the ventral part of the AVCN; those that carry information from the base of the cochlea that are tuned to high frequencies contact neurons in the dorsal part of the AVCN. Several populations of neurons populate the AVCN. Bushy cells receive input from auditory nerve fibers through particularly large endings called end bulbs of Held. They contact stellate cells through more conventional boutons.

<span class="mw-page-title-main">Calyx of Held</span>

The Calyx of Held is a particularly large synapse in the mammalian auditory central nervous system, so named after Hans Held who first described it in his 1893 article Die centrale Gehörleitung because of its resemblance to the calyx of a flower. Globular bushy cells in the anteroventral cochlear nucleus (AVCN) send axons to the contralateral medial nucleus of the trapezoid body (MNTB), where they synapse via these calyces on MNTB principal cells. These principal cells then project to the ipsilateral lateral superior olive (LSO), where they inhibit postsynaptic neurons and provide a basis for interaural level detection (ILD), required for high frequency sound localization. This synapse has been described as the largest in the brain.

Coincidence detection is a neuronal process in which a neural circuit encodes information by detecting the occurrence of temporally close but spatially distributed input signals. Coincidence detectors influence neuronal information processing by reducing temporal jitter and spontaneous activity, allowing the creation of variable associations between separate neural events in memory. The study of coincidence detectors has been crucial in neuroscience with regards to understanding the formation of computational maps in the brain.

The olivocochlear system is a component of the auditory system involved with the descending control of the cochlea. Its nerve fibres, the olivocochlear bundle (OCB), form part of the vestibulocochlear nerve, and project from the superior olivary complex in the brainstem (pons) to the cochlea.

Stimulus filtering occurs when an animal's nervous system fails to respond to stimuli that would otherwise cause a reaction to occur. The nervous system has developed the capability to perceive and distinguish between minute differences in stimuli, which allows the animal to only react to significant impetus. This enables the animal to conserve energy as it is not responding to unimportant signals.

Perceptual-based 3D sound localization is the application of knowledge of the human auditory system to develop 3D sound localization technology.

<span class="mw-page-title-main">Sound localization in owls</span> Ability of owls to locate sounds in 3D space

Most owls are nocturnal or crepuscular birds of prey. Because they hunt at night, they must rely on non-visual senses. Experiments by Roger Payne have shown that owls are sensitive to the sounds made by their prey, not the heat or the smell. In fact, the sound cues are both necessary and sufficient for localization of mice from a distant location where they are perched. For this to work, the owls must be able to accurately localize both the azimuth and the elevation of the sound source.

Bushy cells are two types of second order neuron found in the anterior part of the ventral cochlear nucleus, the AVCN. They can be globular or spherical giving outputs to different parts of the superior olivary complex.

References

  1. Jan Schnupp, Israel Nelken and Andrew King (2011). Auditory Neuroscience. MIT Press. ISBN   978-0-262-11318-2. Archived from the original on 2011-01-29. Retrieved 2011-04-14.
  2. Franken, Tom P; Roberts, Michael T; Wei, Liting; Golding, Nace L; Joris, Philip X (March 2015). "In vivo coincidence detection in mammalian sound localization generates phase delays". Nature Neuroscience. 18 (3): 444–452. doi:10.1038/nn.3948. PMID   25664914.{{cite journal}}: |access-date= requires |url= (help)

Further reading