Notes - Auditory Systems - Hall PDF
Notes - Auditory Systems - Hall PDF
Notes - Auditory Systems - Hall PDF
As shown below, sound plays a variety of important roles in the lives of all animals, including humans.
What Is Sound?
Sound is produced when something vibrates →The vibration disturbs the air around it.
This creates periodic changes in air pressure.
• Rarefaction (low pressure)
• Condensation (high pressure)
These changes in air pressure move through the air as sound waves.
1
Measuring Sound Amplitude (Intensity)
Derived parameter for measuring sound amplitude is sound pressure level (SPL).
The human ear is incredibly sensitive - In terms of power, the sound of a jet engine is
about 1,000,000,000,000 times more powerful than the smallest audible sound.
As a result, The decibel (dB) is the unit used to measure the intensity of a sound.
On the decibel scale, the smallest audible sound (near total silence) is 0 dB. A sound
10 times more powerful is 10 dB. A sound 100 times more powerful than near total
silence is 20 dB. A sound 1,000 times more powerful than near total silence is 30 dB.
Both the amplitude and frequency of a sound can vary with time.
1. Amplitude Modulation (AM)
2. Frequency Modulation (FM)
Sound Features Used for Sound Recognition
Both the time and frequency information contained within a sound are extracted by the
auditory system and used for sound recognition.
2
Oscillographic (top) and sonographic (bottom) analysis of two different spoken phrases showing their unique
spectral and temporal features.
The intensity of a particular frequency component is represented on a gray scale. The darker the shading, the more
energy there is at that frequency.
How is Time and Frequency Information Extracted by the Nervous System?
Pinna funnels sound pressure waves into the meatus where they induce
movement of the tympanic membrane.
The outer ear amplifies frequencies in the range of 2 kHz – 6 kHz; these are critical spectral components of speech.
Middle Ear
The middle ear is comprised of the following structures:
1. Malleus
2. Incus
3. Stapes
The three bones of the middle ear also called the ossicles.
Functions as an impedance matching device for the efficient transfer of acoustic energy from compression waves
in air (low impedance) to fluid–membrane waves (high impedance) within the cochlea.
Inner Ear
3
The Cochlea
Cochlea contains the auditory end organ, the Organ of Corti, responsible for mechanoelectrical transduction.
The Organ of Corti extends along the entire length of the Cochlea.
• Stereocilia of outer row of outer hair cells are embedded in the tectorial membrane.
• Stereocilia of the remaining two rows of outer hair cells and those of the inner hair cells are loosely attached to
the underside of the tectorial membrane.
4
Inner hair cells transform the sound vibrations in the fluids of the cochlea into electrical signals that are then relayed
via the auditory nerve (cranial nerve VIII) to the auditory brainstem.
Outer hair cells do not send neural signals to the brain, but mechanically tune and amplify low-level sounds that
enters the cochlea.
The role of the cochlea in the process of auditory transduction is described in the video below.
The transduction of an auditory signal into a neural signal is referred to as mechanoelectrical transduction.
As you will see, this is due the energy involved in the mechanical displacement of the basilar membrane into a
receptor potential at the level of the inner hair cell.
Transduction takes place within the Organ of Corti, a confocal microscope image of which
is shown at left. Hair cells have been labeled with a green fluorescent stain. The single
row of inner hair cells can be seen inf front, with the three rows of outer hair cells in back.
The transduction channel responds to tension in the system, which is increased by net
displacements between intracellular and extracellular structures.
Does such an arrangement exist at the level of the hair cell? The answer of
course is, yes!
In the EM photomicrograph above you can see the protein-based tip links connecting the stereocilia of an inner hair
cell. These are the "extracellular links" proposed in the model presented above.
Shown below left, is more biologically realistic molecular model for hair cell transduction.
5
When tension on the tip link increases (bending of the stereocilia towards the tallest row),
transduction channels open and the cell depolarizes.
When tension on the tip link decreases (bending of the stereocilia towards the shortest row), transduction channels
close and the cell hyperpolarizes.
The relative motion of the basilar membrane and tectorial membrane resulting from fluid displacement in the
cochlea exerts a shearing force on the stereocilia triggering a transduction event.
To summarize:
• The apical end of the inner hair cells is bathed in endolymph (high
K+).
• The basal end of the hair cells is bathed in perilymph (low K+).
6
Information Processing - Frequency Coding
How is the frequency information embedded in the Wren’s song, and that of other sounds for that matter,
extracted and coded by the vertebrate ear?
As shown below, the width of the basilar membrane changes with length - it is narrow at base; wide at apex.
As a result, the basilar membrane exhibits tonotopic organization (left) where sound
frequency is represented systematically along its length as shown below.
• As shown at left only those inner hair cells sitting directly above the displaced
regions of the basilar membrane would be maximally excited.
• Moreover, the degree of excitation is dependent upon the amplitude of a particular frequency at a particular
time; greater amplitude, greater displacement of the basilar membrane, greater the depolarization of the hair
cell.
• Thus, within the inner ear, there is a PLACE CODE for frequency, as excitation
of inner hair cells is dependent upon their position along the cochlear partition.
• Complex sounds result in the differential activation of numerous inner hair cells
depending upon the distribution of spectral energy.
The pattern of inner hair cell excitation provides information about the spectral content of
a sound representing an across-fiber coding scheme for the frequency content of a complex
sound.
This is illustrated below:
7
Frequency Coding – Active Mechanisms
Electrical Resonance
During sustained depolarization, the membrane potential of inner hair cells resonates at a specific frequency.
• Above, patch clamp records are shown immediately below the hair
cells, while change in receptor potential is shown immediately below
the patch clamp records.
• Note the damped oscillation in membrane potential as the hair cell
depolarizes, repolarizes, then depolarizes again etc. due to the
continued depolarization of the hair cell - this continues until the
depolarization of the hair cell ends.
• This is due to relative kinetics of IKCa and ICa activation and
inactivation.
• When stimulated acoustically, hair cell output is greatest when the
stimulus frequency matches the resonant frequency of the hair cell.
• Thus, hair cells are “electrically tuned”.
• Electrical resonance is particularly effective at frequencies < 500 Hz.
Electromotility
• Outer, but not inner, hair cells express the contractile protein, prestin (immunohistochemically labeled
below), which is localized to their outer membrane.
• When outer hair cells are depolarized, prestin undergoes a conformational change
resulting in hair cell contraction
• Because of coupling to tectorial and basilar membranes, contraction alters
micromechanics of cochlea.
• Result is to amplify and tune basilar membrane displacement, thereby enhancing signal
detection and frequency selectivity.
• The effect on the amplitude of basilar membrane vibration with and
without the contractile mechanism is shown at left.
• Obviously, hair cells over the region of displacement would
depolarize much greater in response to a specific frequency with the
contractile mechanism than without.
8
Each inner hair cell is innervated by 10 – 15 bipolar cell fibers originating in the spiral ganglion as shown below.
Afferents arising from bipolar neurons in the spiral ganglion (above) form the vestibulocochlear nerve.
As shown below, axons of spiral ganglion cells join those coming from the vestibule to form the vestibulocochlear
nerve (8th cranial nerve)
Fibers in the VIIIth nerve transmit encoded auditory information into the CNS at the
level of the medulla oblongata in the brainstem.
Axons of spiral ganglion cells join those coming from the vestibule to form the
vestibulocochlear nerve (8th cranial nerve) which transmits encoded auditory
information into the CNS .
Frequency Coding
• The frequency sensitivity of 8th nerve fibers is defined by the inner hair cells that they innervate.
o Remember, 8th nerve fibers are the axons of spiral ganglion cells that innervate hair cells.
• The frequency selectivity of 8th nerve fibers is described by their excitatory frequency tuning curves.
• Excitatory frequency tuning curves represent the “excitatory receptive fields” of 8th nerve fibers.
• A typical frequency tuning curve is shown above left. A population of tuning curves derived from
recordings from multiple 8th nerve fibers is shown above right.
o Note, the characteristic frequency (CF) is that frequency requiring the lowest amplitude stimulus
(known as the “threshold”) to excite the nerve fiber.
• As shown above right, a population, the excitatory frequency tuning curves of 8th nerve fibers extends
across the frequency range of hearing.
o This is another example of range fractionation.
• Hence, the frequency content of a sound is represented in an across-fiber coding scheme.
Two-tone Suppression
• Two-tone suppression is a nonlinear property in which the response to a particular
frequency is reduced by the simultaneous presence of a second tone at a different frequency.
• Suppression has been observed in measurements of basilar-membrane displacement and of
neural-firing rate.
• Two-tone suppression plays an important role in the coding of complex signals, such as
speech and has been suggested to result in the enhancement of spectral contrast of complex
sounds, such as vowels.
• In the top panel, a 113 Hz, CF tone is presented at 10 dB above the fiber’s threshold
(smaller, dark line).
• The cumulative response of the fiber to 10 stimulus presentations is represented by the
histogram.
• In the remaining panels, a second tone of 600 Hz (indicated by the oval) is embedded in
the CF tone and presented at increasing intensities.
• Note, that when the second tone is presented at 87 dB, it totally suppresses the response of
the fiber to the CF tone.
• Two-tone suppression is non-neural in origin.
• It is Due to destructive interference of traveling waves along the basilar membrane that
serve to decrease the magnitude of movement such as that illustrated in the movie below.
10
• If one varies the frequency and amplitude of the second tone, one can
construct inhibitory frequency tuning curves as shown in blue at left.
• This is a form of lateral inhibition that serves to sharpen frequency
selectivity.
an interspike-interval histogram. →
• It represents the interval between action potentials (the firing rate)
when an 8th nerve fiber responds to a tone presented multiple times
(shown at left).
• Note that there are times that the fiber does not respond to every
cycle, but sometimes skips one, two, even three cycles at times.
• Using the time represented by the maximum height of the first peak of the histogram (the other peaks
are not used), it is seen that t = 2.4 msec which corresponds to a frequency of 416 Hz.
• The auditory system performs this calculation to determine stimulus frequency from the phase-locked
responses of 8th nerve fibers.
• This represents a “rate code” for stimulus frequency at the level of the 8th nerve.
11
Shown above is the temporal envelope, representing amplitude modulation (red line), for a typical sound.
Note, the modulations occur at different rates, have different durations as well as the time it takes for them to reach
their maximum amplitude (rise time) and return to baseline (fall time).
All of these features are temporal cues that the nervous system uses in concert with frequency cues to recognize
sounds as illustrated in the figure below for the phrase, “she wiped her off with a cloth”. Note the differences in
frequency (top) and amplitude modulation (bottom) and how they are unique for each word.
How is amplitude modulation represented at the level of the 8th nerve? This is illustrated in the figure below.
All 8th nerve fibers respond to acoustic stimulation with a tonic discharge pattern that corresponds directly to the
pattern of amplitude modulation.
In conclusion, sound recognition does not take place at the level of the 8th nerve.
However, the activity of 8th nerve fibers provides the CNS with an accurate
neural copy of acoustic signals which are then processed by higher centers in
the ascending auditory pathway.
12
• As you may know, bats use echolocation to identify and capture insects in flight.
• The signals they emit are brief and in the ultrasonic range beyond the frequency sensitivity of human
hearing.
The below video illustrates how bats utilize echolocation in prey capture.
The FM-FM area measures the delay between the emitted pulse and returning echo
thus providing info re: distance to the target – the longer the delay, the more distant
the target.
The CF-CF & DSCF areas measures the frequency of the emitted pulse and
compares it with the frequency of the echo thus providing info re: degree of doppler
shift and therefore, target velocity. If the frequency of the returning echo is higher
than the emitted pulse, the bat “knows” that it is closing in on its prey.
The bat auditory cortex also illustrates a feature common to the auditory cortex of
other animals, including humans. It, like the visual cortex, is modular with respect
to its functional organization.
Sound Localization
It is not only important to determine “what” a sound is, but also “where” it is coming from.
Auditory system uses multiple mechanisms to determine both the vertical and horizontal location of an acoustic
signals.
13
Vertical Plane
• As explained in the movie below, sound waves reflecting off the
uneven surfaces of the pinna and concha, as well as the rest of the
head, to produce interference patterns at different frequencies, called
head-related transfer functions (HTRFs) depending upon the position
of the sound in the vertical plane.
Horizontal Plane
• According to the “Duplex Theory”, sound localization in the horizontal plane is mediated by two cues:
o Interaural Time Differences (ITDs)
o Interaural Intensity (or level) Differences (IIDs or ILDs)
• Both cues require a comparison of the sound signal arriving at the two ears.
• With respect to the auditory system, where and how does this comparison take place?
• At left is a cross section through the mammalian auditory brainstem showing the superior
olivary complex.
• Lateral Superior Olive (2), Medial Nucleus of the Trapezoid Body (3), Medial Superior
Olive (4)
• The analysis of ITDs takes place in the medial superior olive (MSO; 4
above), a subdivision of the superior olivary complex.
Drawings of labeled MSO neurons are shown →. The arrow indicates the
long axis of the MSO.
• Somata of neurons are aligned parallel to the long axis of the MSO.
• Dendritic tufts extend laterally in both directions.
• Tufts on one side receive excitatory input from the contralateral cochlear nucleus.
• Tufts on other side receive excitatory input from the ipsilateral cochlear nucleus.
14
How does this contribute to ITD analysis?
The Jeffress Model explains how ITD information is extracted via the
creation of “delay lines” and coincidence detectors.
Analysis of IIDs takes place in the lateral superior olive (LSO; 2 in the
above cross section of the superior olive), a subdivision of the superior
olivary complex.
15
Maps of Auditory Space
• Spectral notch information is represented by the activity of neurons in the
cochlear nucleus.
• ITD information is represented by the activity of neurons in the MSO.
• IID information is represented by the activity of neurons in the LSO.
• Information carried over these three pathways converges on neurons in the
inferior colliculus creating maps of auditory space.
16