Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Secondary Microsurgical Reconstruction of the Cervical Esophagus: Safer Flaps and Practical Tips in a Challenging Situation
Previous Article in Journal
Clinical Efficacy of Application-Linked Stretching Ball as Digital Therapeutics in Plantar Fasciitis
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Adaptation at Stimulus Onset and Speed of Neural Processing as Critical Contributors to Speech Comprehension Independent of Hearing Threshold or Age

1
Department of Otolaryngology, Head and Neck Surgery, University of Tübingen, Elfriede-Aulhorn-Str. 5, 72076 Tübingen, Germany
2
Department of Biology, Technical University Darmstadt, 64287 Darmstadt, Germany
3
Department of Information Technology, Ghent University, Technologiepark 126, 9052 Zwijnaarde, Belgium
4
Lyon Neuroscience Research Center, Centre National de la Recherche Scientifique UMR5292, Inserm U1028, Université Lyon 1, Centre Hospitalier Le Vinatier-Bâtiment 462–Neurocampus, 95 Boulevard Pinel, 69675 Bron CEDEX, France
5
Department of Otorhinolaryngology, University Medical Center Groningen (UMCG), Hanzeplein 1, BB21, 9700 RB Groningen, The Netherlands
6
Magnetoencephalography-Centre and Hertie Institute for Clinical Brain Research, University of Tübingen, Otfried-Müller-Straße 27, 72076 Tübingen, Germany
7
Center for Mind and Brain Research, University of Trento, Palazzo Fedrigotti-corso Bettini 31, 38068 Rovereto, Italy
8
Department of Psychiatry & Psychotherapy, University of Tübingen, Calwerstraße 14, 72076 Tübingen, Germany
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Clin. Med. 2024, 13(9), 2725; https://doi.org/10.3390/jcm13092725
Submission received: 28 March 2024 / Revised: 24 April 2024 / Accepted: 26 April 2024 / Published: 6 May 2024
Graphical abstract
">
Figure 1
<p>Elevated hearing thresholds correlated with age, in particular at high frequencies: (<b>A</b>) Individual (grey lines) and mean (black line) PTTs for the three age groups young (left), middle-aged (center), and older (right) used for PTAs of four different frequency ranges: low frequencies “PTA-LF” (0.125–1 kHz), main-language region “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” (6–10 kHz), and extended high frequencies “PTA-EHF” (11.2–16 kHz), illustrated in red on the abscissa of the left-most audiogram. The group mean thresholds are plotted in black (young: circles; middle-aged: triangles; older: squares). (<b>B</b>) Scatterplots for individual hearing thresholds as a function of age, split into the four PTA frequency ranges. The shaded area delineates the age range of the middle-aged group. <span class="html-italic">p</span>-Values (Pearson’s correlation): p(PTA-LF) = 0.000016; p(PTA4) &lt; 0.00001; p(PTA-HF) &lt; 0.00001, p(PTA-EHF) &lt; 0.00001.</p> ">
Figure 2
<p>(<b>A</b>) Power spectrum of the OLSA speech material (broadband, black curve), of high-pass filtered speech (OLSA-HP, red shaded area), and low-pass filtered speech (OLSA-LP, blue shaded area), shown peak-normalized to 0 dB and 1/f-corrected. The four different PTA frequency ranges are depicted: low frequencies “PTA-LF” (0.125–1 kHz), “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” [6–10 kHz], and extended high frequencies “PTA-EHF” (11.2–16 kHz). (<b>B</b>,<b>C</b>) The influence of noise on OLSA SRT<sub>50</sub> was examined using differently filtered speech material. Noise conditions in (<b>B</b>) quiet and (<b>C</b>) ipsilateral noise. Columns provide results for broadband and filtered OLSA stimuli. OLSA SRTs are provided as a function of age.</p> ">
Figure 3
<p>ABR as a function of age, pure-tone averages, and speech comprehension: (<b>A</b>) ABR wave amplitude and latencies grouped by age. Circles represent young, triangles middle-aged, and squares older participants. (<b>B</b>–<b>D</b>) ABR wave amplitudes and latencies grouped for participants with low (green) and high (pink) thresholds of PTA4 (<b>B</b>), PTA-HF (<b>C</b>), and PTA-EHF (<b>D</b>).</p> ">
Figure 4
<p>OLSA speech reception threshold SRT<sub>50</sub> (dB SPL; <span class="html-italic">y</span>-axes) for differently filtered OLSA stimuli ((<b>A</b>,<b>E</b>) broadband, (<b>B</b>,<b>F</b>) low pass, and (<b>C</b>,<b>D</b>,<b>G</b>,<b>H</b>) high-pass) as a function of PTA4 (<b>A</b>,<b>E</b>), PTA-LF (<b>B</b>,<b>F</b>), PTA-HF(<b>C</b>,<b>G</b>), and PTA-EHF (<b>D</b>,<b>H</b>) (<span class="html-italic">x</span>-axes). (<b>A</b>–<b>D</b>) provide results obtained in quiet (n = 89), (<b>E</b>–<b>H</b>) under ipsilateral (n = 63) noise condition. Regression lines are plotted in black and include y-intersections and R<sup>2</sup> values. The different colors assign each subject to one of the three speech comprehension groups: good (blue), standard (grey), and poor (orange).</p> ">
Figure 5
<p>Subjective hearing evaluation by age and speech comprehension: (<b>A</b>) shows age groups and (<b>B</b>) groups according to objective speech comprehension performance based on OLSA thresholds corrected by PNOTs. <span class="html-italic">y</span>-axis: subjective evaluation, <span class="html-italic">x</span>-axis: percentage of all responses given by all participants in age groups (<b>A</b>) and in PNOT groups (<b>B</b>). Participants were asked to rate their hearing as excellent, very good, good, moderate, or bad (<span class="html-italic">y</span>-axis labels).</p> ">
Figure 6
<p>(<b>A</b>) ASSR response amplitudes in µV averaged for 4 and 6 kHz carriers as a function of participant age in years. The blue, grey, and orange-colored symbols refer to the good, standard, and poor speech comprehension groups, respectively. (<b>B</b>) Median (horizontal bar) and individual participants (symbols) ASSR amplitude averaged for 4 and 6 kHz carriers (Mean, left), 4 kHz carrier (middle), and 6 kHz carrier (right) for the quiet listening condition (upper row), or in ipsilateral noise (lower row). Numbers in brackets indicate the number of participants included in the analyses. (<b>C</b>) Regression line (black) of the dependence of OLSA SRT<sub>50</sub> in ipsilateral noise on ASSR amplitudes (averaged for 4 and 6 kHz carriers) normalized for PTT. The y-intersection, R<sup>2</sup> value, and <span class="html-italic">p</span>-value of regression are given close to the trend line.</p> ">
Figure 7
<p>(<b>A</b>,<b>E</b>) L<sub>EDPT</sub> acceptance rates, (<b>B</b>,<b>F</b>) PTT, (<b>C</b>,<b>G</b>) L<sub>EDPT</sub>, (<b>D</b>,<b>H</b>) L<sub>EDPT</sub>-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-quiet comprehension performers. Participants with good speech-in-quiet performance (blue) showed higher acceptance rates (<b>A</b>), equal PTT (<b>B</b>), inconclusive L<sub>EDPT</sub> (<b>C</b>), but a consistent 3 dB better threshold for L<sub>EDPT</sub>-to-PTT although on the right ear with only <span class="html-italic">p</span> = 0.084 (<b>D</b>). Estimated distortion-product thresholds (L<sub>EDPT</sub>) in relation to PTT, when participants are grouped with respect to their speech-in-ipsilateral-noise performance (<b>E</b>) L<sub>EDPT</sub> acceptance rates, (<b>F</b>) PTT, (<b>G</b>) L<sub>EDPT</sub>, (<b>H</b>) L<sub>EDPT</sub>-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-noise comprehension performers. Participants with good speech-in-noise performance (blue) show reduced acceptance rates, reduced PTT and L<sub>EDPT</sub>, but no difference for L<sub>EDPT</sub>-to-PTT.</p> ">
Figure 8
<p>ABR wave amplitude as a function of ABR wave latency in participants matched for PTA thresholds and grouped for good (blue), standard (grey), or poor (orange) speech comprehension in quiet. Significant shifts in latency in poor comprehension, in comparison to the group with good speech comprehension were observed (ABR wave I latency: n = 29, 27, 24, <span class="html-italic">p</span> = 0.218242; wave II latency: n = 24, 22, 16, <span class="html-italic">p</span> = 0.007707, wave III latency: n = 30, 28, 26, <span class="html-italic">p</span> = 0.182784; wave V latency: n = 30, 28, 28, <span class="html-italic">p</span> = 0.026617 and wave VI latency: n = 27, 27, 24, <span class="html-italic">p</span> = 0.001055).</p> ">
Figure 9
<p>Syllable-discrimination scores in relation to speech comprehension. The scores for four pairs of phonemes (/o/-/u/, /i/-/y/, /du/-/bu/, /di/-/bi/) are segregated for participants with poor (orange), good (blue), and standard (grey) speech comprehension selected by PNOT in quiet (<b>A</b>), and ipsilateral noise (<b>B</b>). Each plot consists of a boxplot with perceptual performance [% correct] as a function of PNOT (x-axis). Finally, there is a graphical representation of the significance assessed by Mann–Whitney U tests (<a href="#app1-jcm-13-02725" class="html-app">Supplementary Table S4</a>), significant differences are shown as asterisks with a color code reflecting the three groups.</p> ">
Figure 10
<p>Good and poor speech comprehension in quiet differs from good and poor speech comprehension in ipsilateral noise in the discrimination ability of formant contrasts below PLL (requiring TFS coding), and above PLL (requiring TENV coding). In quiet, poor speech comprehension is associated with poor discrimination below the PLL (e.g., for /o/-/u/), while good speech comprehension is associated with good discrimination above the PLL (e.g., for /i/-/y/). In ipsilateral noise, poor speech comprehension is associated with lower performance for discriminating phoneme pairs with formant contrasts above PLL (/i/-/y/, above 1500 Hz), while good speech comprehension is associated with good discrimination of formants below the PLL (/o/-/u/, below 1500 Hz).</p> ">
Versions Notes

Abstract

:
Background: It is assumed that speech comprehension deficits in background noise are caused by age-related or acquired hearing loss. Methods: We examined young, middle-aged, and older individuals with and without hearing threshold loss using pure-tone (PT) audiometry, short-pulsed distortion-product otoacoustic emissions (pDPOAEs), auditory brainstem responses (ABRs), auditory steady-state responses (ASSRs), speech comprehension (OLSA), and syllable discrimination in quiet and noise. Results: A noticeable decline of hearing sensitivity in extended high-frequency regions and its influence on low-frequency-induced ABRs was striking. When testing for differences in OLSA thresholds normalized for PT thresholds (PTTs), marked differences in speech comprehension ability exist not only in noise, but also in quiet, and they exist throughout the whole age range investigated. Listeners with poor speech comprehension in quiet exhibited a relatively lower pDPOAE and, thus, cochlear amplifier performance independent of PTT, smaller and delayed ABRs, and lower performance in vowel-phoneme discrimination below phase-locking limits (/o/-/u/). When OLSA was tested in noise, listeners with poor speech comprehension independent of PTT had larger pDPOAEs and, thus, cochlear amplifier performance, larger ASSR amplitudes, and higher uncomfortable loudness levels, all linked with lower performance of vowel-phoneme discrimination above the phase-locking limit (/i/-/y/). Conslusions: This study indicates that listening in noise in humans has a sizable disadvantage in envelope coding when basilar-membrane compression is compromised. Clearly, and in contrast to previous assumptions, both good and poor speech comprehension can exist independently of differences in PTTs and age, a phenomenon that urgently requires improved techniques to diagnose sound processing at stimulus onset in the clinical routine.

Graphical Abstract">

Graphical Abstract

1. Introduction

Age-related hearing loss is the most prevalent disorder of aging and is associated with future cognitive impairment [1]. Recent studies indicate that an association between hearing and cognition also exists in individuals with subclinical hearing loss—that is, those with normal pure-tone audiograms below 25 dB hearing level up to 4 kHz [2,3]. A deterioration in speech recognition over the lifespan is also observed, despite unchanged pure-tone thresholds [3].
This suggests that the cognitive decline after hearing loss or the worsening of speech comprehension is not necessarily linked with differences in the pure-tone audiogram as currently measured in the clinical routine. Afferent auditory fiber loss (cochlear synaptopathy) may be a candidate contributor that can precede loss of outer hair cells (OHCs) and an overt threshold loss, as shown in animals [4,5,6] and predicted for humans [7,8,9,10,11,12,13,14]. As such, the so-called ‘hidden hearing loss’ refers specifically to the damage to low spontaneous rate (SR) high-threshold auditory nerve fibers (ANFs) that are said to play a role in coding supra-threshold sound features of speech in noise [6,11,12,15,16,17,18,19]. Thereby compromising temporal envelope (TENV) coding, the phase-locked neural response of a population of peripheral and brainstem neurons to a stimulus envelope [20]. In accordance with that, subjects with hidden hearing loss and with speech comprehension deficits in noise display normal pure-tone and speech audiometry thresholds in quiet, and well-synchronized ABRs [21,22]. This form of cochlear synaptopathy has to be separated from auditory neuropathy spectrum disorder of syndromic or non-syndromic origin, which develops in early childhood and leads to the structural anomalies of cochlear nerve hypoplasia or aplasia [22,23] or from postoperative ANF damage during acoustic neuroma surgery [24,25]. In both cases, ABR peak amplitudes and speech audiometry in quiet and, thus, speech recognition below the phase-locking limit (PLL), which is encoded as temporal fine-structure (TFS) by the ANF, are severely affected in adulthood [22,23]. Besides speech comprehension deficits with a major focus on speech-intelligibility deficits in noise [26,27,28,29], to date, variations in speech comprehension deficits in quiet that cannot be explained by hereditary or postoperative causes are less understood [28]. They are in all cases explained by an increase in the hearing thresholds, and therefore predicted to be compensated for by an increase in loudness. To gain a deeper insight into the causes of speech comprehension deficits, which may possibly exist or worsen independently of clinically normal hearing thresholds [3,30], we have used a combination of measurement methods that are not necessarily used in clinical routine. Thus, PTTs, as typically implemented in clinical audiometry, effectively integrate signals over ~500 ms, and thus reflect the adapted state of nerve firing that follows an overshoot of the discharge rate at the onset of the stimuli [31]. Here, we measured PTTs in a total of 89 young, middle-aged, and older individuals for four different frequency ranges: pure-tone averages (PTA) of low frequencies (i) “PTA-LF” (0.125–1 kHz); (ii) “PTA4” (0.5–4 kHz), which are mostly measured to evaluate the hearing threshold in clinical otolaryngology [32,33]; (iii) high frequencies “PTA-HF” (6–10 kHz) to go beyond the frequencies of 6 kHz that are typically deemed sufficient to intelligibly convey speech in communication systems [34] and that still cover conventional frequency ranges measured in many clinical studies [32,35]; and (iv) extended high frequencies “PTA-EHF” (11.2–16 kHz), which have been assumed to play a decisive role in improved spatial hearing [36].
We combined these measures with the analysis of speech reception thresholds (SRTs) in quiet to reflect the state of audibility, and, in noise, to test for the capacity for speech discrimination. For the measurement of SRT50 in quiet, or at a fixed ipsilateral or contralateral noise level, we used the standard German Matrix test Oldenburger Satztest (OLSA) for either unfiltered “broadband” speech (OLSA-BB); low-pass filtered speech (OLSA-LP), frequency components above 1.5 kHz deleted from the OLSA power spectrum, leaving available TFS cues [37]; and high-pass filtered speech (OLSA-HP), below 1.5 kHz deleted from the OLSA power spectrum, leaving available TENV cues [38,39], as previously described [40]. To evaluate the residual speech comprehension performance independent of PTTs, we next subtracted the three OLSA threshold predictions from the three measured OLSA thresholds and averaged them, using a multivariate regression model based on principal component analysis (PCA). We thereby classified subjects with matched PTTs into groups with good, standard, and poor speech comprehension.
Considering factors that possibly contribute to differences independent of PTT, analysis procedures were chosen that enable the diagnosis of signal transmission at stimulus onset prior to auditory nerve firing-rate adaptations that are reached >500 ms. This early stage of sound transmission around the stimulus onset is not reflected in PTTs. Here, as a metric for the temporal precision of auditory coding, the so-called ASSRs were used, which reflect phase-locked neural activity to periodic stimuli coded by TENV [41]. Carrier tones at 4 and 6 kHz were used to induce ASSRs that were modulated at 116 Hz, the same frequency that was used as the fundamental frequency for the speech stimuli. This was at the same time a modulation frequency that is expected to generate ASSR responses by subcortical rather than cortical components [42,43].
In addition, pDPOAE growth functions were measured in all young, middle-aged, and older individuals, a measure that also identifies pre-neural input signals to the inner hair cells (IHCs) prior to firing-rate adaptation [44]. Indeed, in contrast to PTTs, pDPOAEs reflect the state of the cochlear amplifier with high accuracy by using stimulus pulse widths that are more than an order of magnitude lower than PTTs [39,40,44].
Finally, we analyzed peak amplitudes of ABRs, the short-latency evoked potentials that emanate from the auditory pathways and nuclei of the brain stem, and develop within the first 10 ms of stimulation. This also enables the precise detection of signal transmission at stimulus onset. Here, click-induced, supra-threshold auditory brainstem responses (amplitudes and latencies) of ABR wave I/II (generated by the auditory nerve and dorsal cochlear nucleus) [45,46], ABR wave III (generated by the superior olivary complex (SOC) and lateral lemniscus) [47], and ABR wave V and VI (generated by the inferior colliculus (IC) [48] and the medial geniculate body (MGB) [49] were specifically detected as described [50,51]. Finally, considering that language comprehension is dependent upon the correct discrimination of vowels [52] and consonants [53], which, in turn, requires precise TFS coding (below the human PLL, i.e., below 1.5 k Hz) and TENV coding (above the PLL, above 1.5 kHz) [54,55], we hoped to be able to link differences in speech comprehension in subjects of different ages or PTTs with differential contributions of ANFs to TNF or TENV coding, as previously suggested [56].
Toward this aim, we used decomposed narrowband signals that varied depending on whether nerve discharges were above the PLL as /i/-/y/ and /di/-/bi/ phoneme pairs or below the PLL as /o/-/u/ and /du/-/bu/ phoneme pairs [40,57].
Strikingly, we found that differences in signal transmission at the onset of the stimulus contributed to differences in speech comprehension in quiet and noise, independent of age or threshold. This underlines the high relevance of fast transmission speed of auditory information at the beginning of the stimulus for speech comprehension in human clinical studies, as previously speculated on, based on a numerical model [58]. We may thus have identified a new cause of speech-discrimination disorders that has to date evaded typical diagnostic procedures.

2. Materials and Methods

The study was conducted at the Department of Otolaryngology of the University of Tübingen and approved by the ethics committee of Tübingen University (Faculty of Medicine; ethical approval-number 392/2021BO2). Written informed consent was given by all participants. All methods followed the Declaration of Helsinki by the World Medical Association (WMA) for human research ethics.

2.1. Participants

We recruited 112 participants aged between 18 and 76 years. A checklist to inquire about any comorbidities was used for study exclusion. Among these were as follows: other hearing-related conditions such as tinnitus, or previous ear surgery, as well as systemic diseases known to affect hearing. Ultimately, 89 participants were included in the analysis, and the remaining ones were excluded due to comorbidities (threshold elevation beyond 40 dB hearing loss in one or more frequencies and tinnitus) or because of lacking compliance. These 89 participants were evenly distributed across three age groups, young (18–29 years, n = 29), middle-aged (30–55 years, n = 32), and older (56–76 years, n = 28) (Supplementary Table S1). Participants’ age, gender, handedness, and confirmation of normal middle ear function by tympanometry are provided (Supplementary Table S1). Of the 89 participants, only 63 could be measured in quiet, ipsilateral, and contralateral noise conditions of the German word matrix test OLSA, leaving 26 participants who were only tested in the quiet condition.

2.2. Neuropsychiatric Scores

As an exclusion criteria, we applied two validated questionnaires: “Becks Depression Inventory II” (BDI), shown to screen for depression in a clinical setting [59] and the Geriatric Depression Scale (GDS), with a focus on affective and cognitive domains [60] to exclude depression and a German version of the Mini-Mental State Examination (MMSE) [61,62] to exclude dementia. In this test, the participants answer questions related to orientation in space and time, word short-term memory, subtracting, attentive listening, spelling, reading, writing, executive tests, and visuo-construction. The self-assessment of hearing ability was analyzed using an adapted questionnaire that assessed hearing ability in various conversational situations, also concerning education level [63].

2.3. Otoscopy and Impedance Audiometry

The ear examination was carried out by ENT physicians from the Department of Otolaryngology, Head and Neck Surgery at the University of Tübingen. Tympanometry and stapedial-reflex measurements were performed using an AT235 (Interacoustics, Middelfart, Denmark) tympanometry system using a 226 Hz stimulus to ensure intact middle-ear transmission [64] and generally intact neural pathways [65].

2.4. Pure-Tone Audiometry

Using an AT 1000 Audiometer (Auritec, medizindiagnostische Geräte Gmbh, Hamburg, Germany), PTTs were measured for air and bone conduction, as well as the uncomfortable loudness level (UCL). Bone conduction at 0.25, 0.5, 1, 1.5, 2, 4, and 6 kHz was measured using a B71 bone transducer (Radioear, Middelfart, Denmark). The default pure-tone audiometric thresholds from 0.125 to 10 kHz, and the UCL (0.25, 0.5, 1, 2, 4, and 6 kHz) were measured using Beyerdynamic AT1350A on-ear headphones (Beyerdynamic, Heilbronn, Germany). In addition, EHF thresholds were measured using Sennheiser HDA300Pro (Sennheiser, Wedemark-Wennebostel, Germany) on-ear headphones at the frequencies 11.2, 12.5, 14, and 16 kHz. The Sennheiser HDA300Pro achieves a nominal level of 123 dB SPL (6 Hz to 23 kHz). All measurements were conducted in a sound-attenuating chamber (Industrial Acoustics Company GmbH, Niederkrüchten, Germany).
PTA of low frequencies (PTA-LF; 0.125, 0.25, 0.5, and 1 kHz), high frequencies (PTA-HF; 6, 8, and 10 kHz), extended high frequencies (PTA-EHF; 11.3, 12.5, 14, and 16 kHz), and PTA4 (0.5, 1, 2, and 4 kHz) were derived from the right-ear thresholds.

2.5. Auditory Brainstem Responses (ABRs)

The ABR measurements were performed monaurally using three electrodes (Neuroline 720, Ambu, Bad Nauheim, Germany), with electrode impedance consistently below 2 kΩ (ground: Fpz—above the nasion; reference—inverting input (−): Fz—hairline; non-inverting input (+): mastoid). As an amplifier, the actiCHamp Plus64 (Brain Products GmbH, Gilching, Germany) was set up according to the manufacturer’s specifications and at a sampling rate of 50 kHz. Acoustic click stimuli (83 µs) were presented at two different stimulus levels (70 dB SPL and 80 dB SPL) with 3000 repetitions of alternating polarity. Stimuli were generated using a Scarlet Focusrite 8i8 gen 3 (Focusrite, UK) soundcard and presented through ER2 transducers and disposable ER1-14A earpieces (Etymotic Research, Elk Grove Village, IL, USA). To minimize muscle effects, the participants lay on their backs during the measurements. ER2 in-ear loudspeakers exhibit frequency bandwidth limits to approximately 8 kHz, which allowed conclusions to be drawn about changes caused by the frequency content within this bandwidth.
After band-pass filtering (30–2000 Hz; first order FIR filter, Hamming windowed), ABR waveform components were averaged at each stimulus level. Wave V was determined to be the most prominent peak, typically appearing 5–6 ms after stimulus onset. Waves I, II, III, and VI were then assigned to peaks at 1 to 2 ms, 2 to 3 ms, 3 to 4 ms, and 6 to 7 ms after stimulus onset, respectively. Wave amplitudes were calculated in µV as the difference between leading positive and trailing negative deflections/peaks, as previously described [50,51]. Their latency was measured from the leading positive peak.

2.6. Auditory Steady-State Response (ASSR)

ASSR was measured using the same recording setup and without changing the position of the participants. The modulation frequency was set to 116 Hz (rectangular 100% amplitude modulation as described in [66]) and two blocks of 800 epochs each were recorded at carrier frequencies of 4 and 6 kHz at 70 dB SPL rms. The stimulus duration was set to 400 ms, with an epoch duration of 500 ± 10 ms. Responses from all epochs were averaged and the spectral power was calculated by FFT (MATLAB 2021b). ASSR peak amplitudes (µV) were averaged for the first three harmonics [66]. Measurements with inadequate signal-to-noise ratios (SNR below 2) or ASSR peak amplitudes higher than 0.15 µV were excluded from the statistical evaluation.

2.7. Distortion-Product Otoacoustic Emissions (DPOAEs)

Input–Output (I/O) functions of pDPOAEs were measured to characterize the pre-neural state of the cochlea. Using a pulsed waveform for the second primary (f2), along with onset decomposition [44], a technique to capture the short-latency nonlinear-distortion (ND) component of the DPOAE [67], artefactual interference effects from the longer-latency component can be safely avoided [68]. From these pDPOAE I/O functions, two measures were analyzed in the results section: The extrapolated pDPOAE threshold and the acceptance rate. When stimulus levels of both primaries are chosen according to a so-called scissors paradigm [69], extrapolated DPOAE thresholds (level of the estimated distortion product threshold; LEDPT) based on semi-logarithmically scaled I/O functions have been shown to correlate nearly 1:1 with the pure-tone threshold for hearing losses up to about 50 dB [44,70,71,72]. Most previous studies, such as Kummer 1998 [69], Boege 2002 [70], Georga 2003 [71] and Johnson 2007 [72]) used several criteria for I/O function acceptance; these effectively avoid hard-to-interpret DPOAE I/O functions that lead to large extrapolation errors. Thus, the acceptance rate is the number of I/O functions passing these criteria divided by the number of measurements. The acceptance rate informs about DPOAE levels (because any pDPOAE value to be included in the extrapolation procedure must have an SNR of 10 dB or above), as well as the integrity of the measured I/O function. In the present study, pDPOAE I/O functions were measured using an in-ear-probe at 8 frequencies (f2 = 0.8, 1.2, 1.5, 2, 3, 4, 6, and 8 kHz) using an adaptive algorithm comprising at least four pDPOAE values. For details of recording and method, see Supplementary Material, section “pDPOAE measurements”.

2.8. Speech Reception Thresholds (OLSA)

Speech intelligibility was tested using the “Oldenburger Satz Test” (OLSA), the German version of the International Matrix test [73,74], applying three different configurations of the masker noise presentation in each of the three different speech material filtering conditions. The three noise conditions were no noise (quiet), ipsilateral noise, and contralateral noise.
The three speech material filtering conditions were as follows: unfiltered broadband speech (OLSA-BB), low-pass filtered speech (OLSA-LP, components above 1.5 kHz were deleted from the OLSA power spectrum), and high-pass filtered speech (OLSA-HP, components below 1.5 kHz were deleted from the OLSA power spectrum) (see for details [40]). The nine conditions were presented in random order. Sentences consisted of five words, with a name, verb, number, adjective, and object, and each keyword having ten response possibilities, producing a large number of (105) combinations from an inventory of a total of 50 words. The speech material of the OLSA is spoken by a male speaker [73], the average F0 of which we determined to be 116 Hz. For each condition, participants were presented with 20 sentences. As initial training, an OLSA-BB and OLSA-HP test of 20 sentences each was completed before starting the target sentence presentations. This served to reduce the impact of the training effect, which has been shown to be the largest after the first presentation (approx. 1 dB, [75]). In order to reduce fatigue due to the extent of speech testing, the session was paused after the first three OLSA-BB, OLSA-HP, and OLSA-LP in quiet conditions with an intermittent request, during which psychoacoustic tasks were performed, before it continued for contralateral and ipsilateral noise conditions. The target sentences and the masker noise were presented monaurally (speaker always right, masker noise presented to the same or the other ear) over ER2 transducers (see Section 2.5, ABR). The level of the target sentence varied and was decreased after a correct response (i.e., increasing difficulty) or increased after an incorrect response (i.e., decreasing difficulty). The masker noise was derived from the speech material by randomly shifted overlapping and, thus, exhibits the same long-term spectrum [73], and closely resembles spectra of several speech materials in other languages [76,77]. The level of the masker noise was fixed at 70 dB SPL. Speech reception thresholds for 50% correctly identified words (SRT50) were determined for the three noise masker conditions [40]. For each of the nine different conditions, blocks of 20 sentences were presented.

2.9. Pure-Tone-Normalized OLSA Threshold

This study was particularly focused on factors beyond pure-tone thresholds and how they may relate to OLSA thresholds. To evaluate the role of these other factors, OLSA thresholds were quantitatively normalized for PTTs of all available frequencies (0.125–16 kHz). This correction was calculated independently for the quiet, ipsilateral noise, and contralateral noise conditions by performing a multivariate regression between all three OLSA (BB, LP, and HP) thresholds and the first five principal components (MatLab Version 2021b) of all audiometric thresholds; the latter to avoid overfitting. The PCA was performed by employing a singular-value decomposition algorithm. Together, the first five PCA components captured 93% of the variations in audiometric thresholds. OLSA predictors for each individual subject were derived by evaluating the linear regression model using these first five components of the pure-tone thresholds of each participant. Thus, 77.8% of the participants with good, standard, and poor speech comprehension in contralateral noise were grouped into the same categories in the speech-in-quiet tasks. The same can only be said for 39.9% of participants from the speech comprehension in ipsilateral noise tasks, meaning that subjects with good or poor speech comprehension, independent of PTT in quiet and age, were not necessarily those that exhibited good or poor speech comprehension independent of PTT in ipsilateral noise conditions and age.
To evaluate the residual speech comprehension performance, we subtracted the three OLSA threshold predictions from the three measured OLSA thresholds and averaged them. This average value will be referred to as the PT-normalized OLSA threshold (PNOT). Based on PNOT, the cohort was divided into three equally sized groups with “good”, “normal”, and “poor” speech performance. We verified that this data-driven approach resulted in groups with matched average pure-tone audiometry thresholds within ± 4.5 dB for the four PTA frequency ranges (PTA-LF, PTA4, PTA-HF, PTA-EHF).

2.10. Stimuli for Phoneme Discrimination

The stimuli used for the phoneme discrimination task were computer generated from recordings from a male speaker, and using analysis/re-synthesis as implemented in the WORLD vocoder [78]. To match the OLSA speech material, their average fundamental frequency (F0) was set to 116 Hz during synthesis. A total of eight phonemes, two pairs of steady-state vowels, and two pairs of consonant–vowel syllables were used as stimuli.
The vowel pairs /o/ (like in oder, “or” in German) and /u/ (like in Du, “you” in German) that differed in their first formant (F1, see Supplementary Table S2) and are located well below the supposed PLL in humans (~1.5 kHz) were synthesized with a 30 ms raised cosine ramp at the onset and offset, and had a total duration of approximately 414 ms (corresponding to 48 F0 cycles). Similarly, the /du/-/bu/ syllable pair only differed at frequencies below the PLL, and only within the first 100 ms. The following 371 ms were exactly identical between the two syllables. In addition, the vowel segment of this syllable pair, /u/, was identical to the isolated steady-state /u/ used in the /o-u/ vowel pair, except that it was trimmed such that the overall duration of the syllables was 471 ms.
The vowel pairs /i/ (like in sie, “she” in German) and /y/ (like in üben, “practice” in German) only differed in their second and third formants (F2 and F3), which were above the PLL (see Supplementary Table S2). As a result, it is expected that encoding of this /i/-/y/ contrast could not rely on TFS, but rather on envelope coding. This /i/-/y/ vowel contrast had the same durations and ramps as the other vowel pair described above. The /di/-/bi/ syllable pair was also built to only differ in frequencies above the PLL, and within the spectral power of the first 100 ms. Again, the /i/ from these syllables was identical in spectral shape to the /i/ used in the vowel pair /i/-/y/. All stimuli were then spectrally tilted to ensure similar signal-to-noise ratios above and below the PLL when presented in the speech-shaped noise used in the OLSA task.
From each of the four stimulus pairs, a nine-step continuum was generated by gradually modifying the formant’s frequencies on a log-frequency scale. Following piloting, a large and a small contrast were selected for each pair. Given the large inter-individual variability observed during piloting, this selection aimed to ensure that floor or ceiling effects would be avoided for at least one of these contrast magnitudes.
The stimuli, presented via ER2 earphones, were equalized such that the average level of the stimuli belonging to a given continuum was the same for all pairs, and was adjusted to a 60 dB SPL energy-equivalent continuous sound level (Leq). However, minor level fluctuations within a continuum were preserved to ensure that the level of the formants that remained identical throughout the continuum were not affected. Calibration was performed using a B&K Type 4157 Microphone (Hottinger Brüel & Kjær, Virum, Denmark) in combination with an artificial ear with a volume of 1 cm3 and a 20 s integration time.

2.11. Behavioral Phoneme Discrimination Task

The phoneme discrimination between the pairs (/o/-/u/, /i/-/y/, /du/-/bu/, /di/-/bi/) was measured using a three-alternative forced choice (3AFC) paradigm. For each phoneme pair, we measured two difficulty levels, easy and difficult. We quantified the difference using nine levels that were tested in pilot experiments from which two pairs, one easy and one difficult, were selected for the psychoacoustic session. The differences in difficulty level, e.g., /du/-/bu/ in the easy condition was eight out of nine, while in the difficult condition, the difference was only four out of nine. The respective other level differences for the difficult and easy conditions were for /di/-/bi/ eight and four, for /o/-/u/ four and two, and for /i/-/y/ three and one. Together with the three different noise conditions (quiet, ipsilateral, and contralateral noise), we acquired data under a total of six conditions. Each condition was repeated nine times, producing a total of 54 trials. The noise was the same speech-shaped noise used during the OLSA measurement and presented at 0 dB SNR.
To minimize learning effects, conditions were randomly reordered at the beginning of the measurement. As a short initial training run, three trials of phoneme discrimination of the four syllable pairs were performed.
The right ear was used to test each syllable pair, using the same transducers as used for the OLSA test. Before each condition of the test, the participants were given four training trials with visual feedback. However, responses from this training were not included in the data analysis. During the main test, participants did not receive any feedback on the correctness of their responses.

2.12. Statistical Analysis

To test for the significance of group differences, statistical tests for non-normally distributed data were applied. ABR wave amplitudes and latencies were compared using one-way analysis of variance (ANOVA) for group differences. The resulting p-values smaller than the criterion of α = 0.05 were considered statistically significant. The correlation of two measurement parameters was verified by the Pearson correlation coefficient (r). ASSR amplitudes (µV) were compared by Mann–Whitney U tests between good and poor performers, between poor and standard performers, and between standard and good performers (1-sided hypothesis).

2.13. Pure-Tone-Normalized OLSA (PNOT)

For PNOT, the score of percent-correct answers was compared by Mann–Whitney U tests between good and poor performers, between poor and standard performers, and between standard and good performers (1-sided hypothesis: The group of poor/standard performers contains more participants with low-percentage correct scores relative to the standard/good performers. A resulting p-value equal to or smaller than α = 0.05 was considered statistically significant and noted as an asterisk in the respective figure panel (Supplementary Figure S1; Supplementary Table S4). A p-value smaller than 0.1 was noted with an asterisk in brackets to indicate a trend in the distribution, though not reaching statistical significance. Statistical comparisons were made for the percent-correct scores obtained for the “difficult” discrimination task (small spectral and temporal syllable contrast), and for the scores obtained for the “easy” discrimination task (larger spectral and temporal syllable contrast).

2.14. Variance Analysis

Analysis of variance for speech perception thresholds beyond pure-tone thresholds was performed by least-square multivariate linear fitting of the five added principal components (PCs) derived from pure-tone thresholds, and one additional observable that was tested for its contribution to total speech comprehension variance. To ensure the uniqueness of the multivariate linear model, we first removed all linear correlations between the five PCs and the tested observable. This can be understood as removing the influence of PTTs on ABR wave amplitude or latency or other parameters such as LEDPT thresholds, ASSR amplitudes, or phoneme discrimination. An inherent risk of the increase in dimensions of the regression model is the possibility of overfitting. To eliminate this effect, we compared the observed increase in explainable variance in the observed six-dimensional model to the variance of 10,000 pseudo models in which we randomly shuffled the additional observable before fitting the model. This gave us a reliable estimate of what gain in the explained variance was achieved based on chance. The results are presented as stacked bar diagrams showing the percentage of variance that could be attributed to each of the observables (Supplementary Figure S1, Supplementary Table S3). To better illustrate the magnitude of the effect, we additionally computed the standard deviation in the unit of the SRT (dB), which can be attributed to the new observable, by taking the ratio of the variance of the observable to the overall variance, and multiplying by the standard deviation (SD) of the sample in (dB) (Supplementary Figure S1; Supplementary Table S3). However, during this computation, we assumed a normal distribution of the five PCs and the additional tested observable, an assumption that is not required for the statistical evaluation based on permutation analysis.

2.15. Data Distributions

If not indicated otherwise, data are presented as group mean and SD for the number of participants or ears (n), as specified in the figure legends. For visualization of the different performances of poor, good, and standard speech reception, syllable discrimination performance (% correct) was classified in histograms with logarithmic class sizes.

3. Results

From all 89 young, middle-aged, and older participants (Supplementary Table S1), a five-grade custom questionnaire for subjective self-evaluation of hearing performance (excellent, very good, good, moderate, bad) in different conversational situations was conducted, and three psychometric tests (BDI, GDS, MMSE) were performed to exclude any confounding, severe psychiatric factors such as depression or dementia onset (Supplementary Table S1). There were no confounding abnormalities identified in the included group.

3.1. Pure-Tone Thresholds Are Elevated with Age

As described in Section 2.4, PTTs were collected for frequencies between 0.125 and 16 kHz. The distinct frequency ranges PTA-LF [0.125–1 kHz], PTA4 [0.5–4 kHz), and PTA-HF [6–10 kHz], and extended high frequencies PTA-EHF [11.2–16 kHz] are illustrated on the abscissa of the leftmost audiogram (Figure 1A). Comparison of the three age groups revealed group differences that shifted toward significantly elevated thresholds above 8 kHz, which, as reported in previous studies, was particularly prominent for EHF thresholds between 11.2 and 16 kHz (Figure 1A) [79,80,81,82].
In addition, all four PTAs correlated significantly with age (Figure 1B, PTA-LF: p = 0.000016, R2 = 0.1929; PTA4: p < 0.00001, R2 = 0.3182; PTA-HF: p < 0.00001, R2 = 0.8426; PTA-EHF: p < 0.00001, R2 = 0.8185). The slope of the regression lines (see Figure 1B, R2 values) is much steeper if the correlation is computed with respect to the HF and EHF averages, with considerably higher R2 values than for the lower-frequency averages (0.84 and 0.81 as compared to 0.19 and 0.32 for PTA-LF and PTA4, Figure 1B). This supports the notion that age-dependent hearing loss is predominantly an increasing loss of high-frequency hearing.

3.2. Speech Reception Thresholds Elevate with PTA-Threshold and Age

SRT50 were next analyzed using either an unfiltered speech-in-noise signal (OLSA-BB), a low-pass filtered condition (OLSA-LP), or a high-pass filtered condition (OLSA-HP, see Section 2.8 and Figure 2A). SRT50 were determined for OLSA-BB, OLSA-LP, and OLSA-HP (Figure 2B,C) and plotted as a function of age. We observed that SRT50 in quiet were significantly positively correlated with age, with the strongest dependence for the high-pass filtered condition, as shown for quiet (Figure 2B, OLSA-BB: R2 = 0.256, p < 0.00001; OLSA-LP: R2 = 0.138, p = 0.000172; OLSA-HP: R2 = 0.363, p < 0.00001).
SRT50 under speech-shaped ipsilateral noise were also significantly correlated with age, but the dependence on age was markedly reduced (Figure 2C, OLSA-BB: R2 = 0.184, p = 0.000223; OLSA-LP: R2 = 0.065, p = 0.0221; OLSA-HP: R2 = 0.129, p = 0.001921), and the respective slopes were shallower than the slopes for the speech-in-quiet condition (Figure 2C).
To summarize: SRT50, as analyzed using defined OLSA filter conditions, revealed that speech comprehension significantly depended on age under all conditions. The variance and regression curve increased most with OLSA-HP, when frequencies below 1.5 kHz were deleted from the speech power spectra. This is surprising, considering the low age-dependent threshold increase in the range of these low-frequency spectra.

3.3. The Supra-Threshold ABR Wave Decreases with Elevated, Age-Dependent PTA-EHF

Supra-threshold ABR waves (amplitudes and latencies) of ABR wave I–VI were measured by acoustic stimulation at 80 dB SPL rms in young, (Figure 3A, solid line, circle), middle-aged (Figure 3A, dotted line, triangle), and older participants (Figure 3A, dotted line, square) using loudspeakers with bandwidth, limited to ~8 kHz (see Section 2.5). ABR wave I had significantly lower amplitude in the middle-aged and older participants than in the young (Figure 3A, young: 0.203 µV, n = 27; middle-aged: 0.111 µV; n = 27; older: 0.119 µV; n = 19; p = 0.00029). However, this lower input was somewhat compensated for by ABR wave VI amplitudes in the middle-aged group (Figure 3A, young: 0.341 µV, n = 27; middle-aged: 0.335 µV, n = 31; older: 0.291, n = 20, p = 0.35). Nonetheless, it was still significantly delayed in the older group for wave III and wave V (Figure 3A, wave III, young: 3.67 ms, n = 29; middle-aged: 3.77 ms; n = 31; older: 3.86 ms, n = 24; p = 0.0095; wave V, young: 5.55 ms, n = 29; middle-aged: 5.66 ms, n = 32; older: 5.76 ms, n = 25; p = 0.018). Interestingly, the lower-amplitude and delayed ABR wave VI in the older group was neither linked to PTT differences in the PTA-4 frequency range (Figure 3B) nor to PTT differences in the PTA-HF frequency range (Figure 3C). When ABR amplitudes were plotted as a function of latencies for PTA-EHFs frequency ranges, altered late ABR wave amplitudes in the older (Figure 3A) were linked to lower and delayed ABR wave V/VI (Figure 3D).
To summarize: Poorer PTA-EHF thresholds as a function of age are associated with a persistently reduced and delayed click-evoked ABR wave V/VI. This strongly indicates that the elevation of PTA-EHF with age may negatively influence both late ABR wave V/VI peak amplitudes and latencies in lower frequency ranges, as also suggested in previous studies [79].

3.4. Speech Comprehension Exhibits Components That Are Dependent on and Independent of Pure-Tone Threshold and Age

To investigate the dependence of frequency-specific filtered speech bands on the pure-tone threshold in more detail, we next tested the dependence of all OLSA-SRTs (BB, LP, HP) on their corresponding pure-tone averages, such as PTA-LF, PTA-4, PTA-HF, and PTA-EHF in quiet (Figure 4A–D) and ipsilateral noise conditions (Figure 4E–H, Table 1). In all 12 comparisons, OLSA SRT50 significantly depended on the corresponding PTA measure, with stronger scatter for the dependence of OLSA-HP SRT50 on PTA-HF, and OLSA-HP on PTA-EHF, under all three noise conditions (Figure 4, Table 1). For OLSA in ipsilateral noise, the correlation shows a clearly reduced slope of the regressions under all frequency conditions, with a fairly homogenous distribution around the regression line for hearing losses up to 40 dB, as shown in previous studies [83].
Under all conditions, the OLSA SRT50 correlated better (higher R2) with the PTA (Figure 4, Table 1) than with age (Figure 2). This effect was very clear for OLSA-BB in quiet (Figure 4A), with a mean R2 = 0.5368 for PTA (Figure 4A) vs. R2 = 0.26 for age (Figure 2B). This effect was less pronounced for OLSA-BB in ipsilateral noise with R2 = 0.31 for PTA4 (Figure 4E, Table 1) vs. R2 = 0.18 for age (Figure 2C), where it almost vanished only for OLSA-LP vs. PTA-LF (Figure 4 and Figure 2).
To examine the variance of OLSA-SRT with the distinct PTA in more detail, we performed a post hoc classification based on PNOT, as described in Section 2.9. This approach was used to separate the cohort into three groups, with a maximum spread in their speech comprehension (Figure 4, see blue = “good” and orange = “poor” dots), as well as the “standard” group with comprehension performance between good and poor groups (Figure 4, grey dots). As expected, participants with poor, standard, and good speech comprehension based on PNOT showed a well-matched mean PTT (Table 1). Interestingly, they were almost the same mean age (Table 1), although the standard PNOT group was on average slightly younger, and exhibited slightly, but not significantly, better PTA-EHF thresholds: PNOT-quiet: p(age) = 0.100; p(PTA-EHF) = 0.150; PNOT-ipsi: p(age) = 0.713; p(PTA-EHF) = 0.547 (Table 1).
Speech-in-quiet comprehension, used here essentially as a control, depended strongly on the PTT, with a slope of more than 0.6 dB/dB. When removing the effect of the PTT upon the PNOT (see Section 2, Supplementary Figure S1, Supplementary Table S3), the multivariate regression of the PTT on OLSA thresholds returned to R2 = 0.49.
The remaining variance of the speech intelligibility that was not explained by the PTT comprised in OLSA-BB was 38.7%, corresponding to an SD of 3.7 dB in SRT50. In contrast, broadband speech-in-noise comprehension depended on the PTT, with a slope of approximately 0.1 dB/dB, leaving 51.4% of the variance unexplained when removing the effect of the PTT, corresponding to an SD of 1.0 dB in SRT50.
In summary: Analyzing the PNOT, we identified good and poor speech comprehension independent of PTT and age. To explain the remaining variance of OLSA-SRT independent of PTT, we (i) compared subjective speech understanding dependent on age with that obtained independent of PTT, and compared good and poor PNOT groups using diagnostic tools that enabled us to identify the efficiency of signal transmission at stimulus onset not reflected in PTTs (see Introduction) as (ii) ASSRs, (iii) precise measurements of the cochlear amplifier using pDPOAEs, (iv) click-evoked central supra-threshold ABRs, and (v) phoneme discrimination ability. The respective shares of the contributions to variance in speech comprehension of each of these measurements are depicted in Supplementary Figure S1 and calculated as the percentage of significant contributions to the variance of OLSA SRT50 in Supplementary Table S3.

3.5. Differences between Good and Poor Pure-Tone-Normalized OLSA Thresholds (PNOTs) Is a Better Indicator of Self-Assessed Hearing Ability than Age

When comparing the self-evaluation of speech comprehension of young, middle-aged, and older participants (Figure 5A), with that obtained when participants are grouped by good, standard, and poor speech comprehension according to PNOTs (Figure 5B), we found PNOT classification of participants to be a better predictor of the subjective assessment of their hearing than their age. In particular, a higher percentage of the middle-aged and older participants rated themselves as hearing very well—comparable to the young population—when the sub-division was assessed according to age (Figure 5A, young, n = 29, middle-aged, n = 32, older n = 28, p = 0.50, one-sided Fisher Exact Probability Test for “very good” and “good” assessments). However, when patients were instead classified as having good, standard, or poor speech comprehension by the PNOTs in quiet, we found that the expected self-rated decrease in hearing ability with age was much more congruent with real hearing performance, though not reaching statistical significance (Figure 5B, good, n = 30, standard. n = 29, poor, n = 30, p = 0.10, one-sided Fisher Exact Probability Test for “very good” and “good” assessments). There were no statistically significant differences when grouped for PNOTs in ipsilateral noise (good, standard, poor, n = 21, p = 0.50).
In summary: The differences between good and poor speech comprehension that remained when OLSA thresholds were normalized for PTTs was a better indicator of self-assessed hearing ability than age, highlighting the relevance of the factors that may contribute to good and poor speech comprehension independent of PTT.

3.6. The Difference between Good and Poor PNOTs Shows Low Dependence on Temporal Envelope (TENV) Coding (ASSR)

The averaged ASSR amplitudes were analyzed next as a function of age at both 4 and 6 kHz carrier frequencies (Figure 6A), and inspected for significance in groups with poor, standard, and good speech comprehension in quiet and in ipsilateral noise (Figure 6B) conditions. A tendency of lower amplitude with increasing age was noted for ASSR amplitudes at both 4 and 6 kHz carrier frequencies (shown for the average of 4 kHz and 6 kHz ASSR responses in Figure 6A), with nearly equal contribution of poor or good PNOTs along the different age.
The grouped ASSR amplitudes were found to be significantly different (larger) in the group with poor speech comprehension for the mean ASSR response in the quiet condition (Figure 6B, top panel, p = 0.037) and in ipsilateral noise masking for 4 kHz (Figure 6B, lower panel, p = 0.045). Also, in the post hoc linear-mixed model analysis after permutation, ASSR was the only electrophysiological measure that significantly explained a considerable amount (7%) of speech-in-noise comprehension (p = 0.012); this corresponded to 0.4 dB under the broadband condition. Therefore, under quiet and ipsilateral noise conditions, variations in speech comprehension of OLSA-BB normalized for the PTT showed a significant association between larger ASSR amplitudes and poor speech reception thresholds of OLSA-BB (Figure 6B,C, p = 0.0037).
In summary: Poor speech comprehension in quiet and noise, which remained when the OLSA threshold was normalized for the PTT, was associated with slightly increased amplitudes of TENV coding, although over age, ASSR amplitudes decreased.

3.7. The Difference between Good and Poor PNOT Is Reflected in Differences in Cochlear Amplifier Efficiency at Stimulus Onset

When pDPOAE growth functions were analyzed in participants that were classified by good (Figure 7A–D, blue) or poor (Figure 7A–D, orange) speech-in-quiet recognition, four out of ten factors became significant, and one became a tendency. (i) The percentage of accepted estimates of pDPOAE thresholds (LEDPT) was higher for participants with good speech reception in comparison to participants with poor speech reception for both the left (n = 60 ears; p = 0.039) and right (n = 60 ears; p = 0.039) ear (Figure 7A). The difference in acceptance rate was assessed with a chi-squared test, using the two speech performance groups as the first dimension, and above or below-average LEDPT acceptance rate as a second dimension. (ii) The PTT was not different between groups (Figure 7B). (iii) For the left ear, LEDPT was significantly lower for the good performers (p = 0.012) (Figure 7C). (iv) When LEDPT was normalized for the PTT, a significantly lower cochlear threshold persisted in participants with good speech comprehension in comparison to those with poor speech comprehension in the left ear (p = 0.017) and remained different with a tendency (p = 0.084) for the right ear (Figure 7D). The difference in normalized LEDPT–PTT between the groups was 2.8 and 3.1 dB for the right and left ear, respectively (Figure 7D). Even when excluding the results for 8 kHz, higher LEDPT–PTT differences were observed for subjects with poor speech performance, showing again significance in the left and a tendency in the right ear. Thus, semi-logarithmic DPOAE I/O functions, a measure that relates to cochlear amplification at near-threshold sound-pressure levels (LEDPT), as well as a measure that is influenced by cochlear amplification at stimulus levels up to 55 dB SPL (acceptance rate), represent a stronger cochlear amplifier (lower LEDPT values, higher acceptance rates) if a subject has good speech-in-quiet recognition, or weaker amplifier if a subject has poor speech-in quiet recognition. If one disregards the lack of complete consistency, the conclusion would be that a stronger pre-neural input signal to the IHC is an advantage for speech-in-quiet recognition for subjects with an equal behavioral PTT, and thus could point to a previously unrecognized influence of cochlear amplification in speech reception in quiet.
When pDPOAEs were analyzed in participants that were classified by good (Figure 7E–H, blue) and poor (Figure 7E–H, orange) speech-in-ipsilateral-noise recognition, five out of ten factors became significant, and one a tendency (only four measures for the two ears shown in Figure 7; we omitted the slope of the I/O functions that, with one exception, never became significant). In contrast to quiet conditions, the acceptance rate was significantly higher for poor performers in the right ear (p = 0.031) (Figure 7E), and the PTT was lower for poor performers, with a tendency in the left ear (p = 0.066), and significant in the right ear (p = 0.023; Figure 7F). Moreover, LEDPT was significantly lower for poor performers in the left (p = 0.0022) and the right ear (p = 0.00015, Figure 7G). In addition, the slope in the right ear was significantly steeper for poor performers (p = 0.041, not included in Figure 7). Thus, measures of hearing sensitivity close to the threshold, the PTT, the distortion-product threshold (LEDPT), and a measure that is at least influenced by cochlear amplification at levels up to 55 dB SPL (acceptance rate) represent stronger cochlear amplification (lower LEDPT values, lower behavioral thresholds, higher acceptance rates) if a subject has poor speech-in-noise recognition.
This supports the finding that, in contrast to subjects with poor speech comprehension in quiet, those with poor speech comprehension in noise do not have low, but rather higher, cochlear-amplification performance.
Finally, we tested whether there were contributions to the total variance of speech comprehension performance based on each of the three differently filtered versions of OLSA. We found that the DPOAE I/O function acceptance rate (Figure 7A), as well as the difference between LEDPT and PTT (Figure 7D), survived the most restrictive post hoc linear mixed-model analysis after permutation (p = 0.001–0.033), explaining 2.0–8.3% of the variance, or 0.8 dB, and 3.2–4.8 dB of the SRT50 variation in the broadband and high-pass condition, respectively (Supplementary Figure S1, Supplementary Table S4). Under ipsilateral noise, the acceptance rate of LEDPT measurements (Figure 7A) was significant to explain the variance of SRT50 in broadband and high-pass condition and almost significant (p = 0.051) in the low-pass condition, accounting for 3.1 to 5.5% of the variance, but only 0.3 to 1.0 dB in SRT50. Here, the PTT-corrected LEDPT thresholds explained up to 7% (0.3 dB in SRT50) of the variance of the OLSA, but only in the low-pass condition (Supplementary Figure S1, Supplementary Table S3).
In summary: Poor speech comprehension in quiet, independent of PTT and age, is linked to elevated pDPOAE thresholds, putatively reflecting a poorer pre-neural input signal at stimulus onset. Poorer speech comprehension in noise, in contrast, is linked with lower pDPOAE thresholds, thus reflecting rather stronger pre-neural input signals at stimulus onset.

3.8. The Difference between Good and Poor PNOTs Is Reflected in Variations in Supra-Threshold Amplitude and Response Latencies of ANFs

Lower efficiency of cochlear amplification, resulting in a poorer pre-neural input signal at stimulus onset, as observed here in individuals with poor speech comprehension in quiet, may influence the fast onset peak of the ANF spike rate that contributes to ABR peak amplitude. This reflects the peak spike-rate increase before the ANF firing rate declines to a steady-state value [84]. To measure this, we analyzed the amplitudes of supra-threshold ABR waves for PNOT groups (Figure 8). We observed that the amplitude of wave I differed by 0.0148 +/− 0.0129 µV in participants with poor speech comprehension, in comparison to those with good speech comprehension (Figure 8). This limits the detection threshold for input amplitude differences to 5 dB when assuming the ABR wave I growth with respect to the sound presentation level, as described by [85]. Moreover, we found a significantly smaller ABR wave II amplitude in participants with poor speech comprehension in comparison to those with good speech comprehension (Figure 8, good: 0.0767 µV, n = 24; standard: 0.0905 µV, n = 22; poor: 0.0458 µV, n = 16; p = 0.0458), while no difference was observed in the ABR wave III, V, or VI amplitude between participants with good and poor speech comprehension (Figure 8).
Correspondingly, when testing the predictive ability of ABR waves for variance of speech reception thresholds in quiet (Supplementary Figure S1, Table 1), we found that the ABR wave I amplitude was negatively correlated with the PNOT in OLSA-BB and OLSA-LP condition, explaining 2.5 and 2.8% of the remaining variance, respectively. This means that smaller early supra-threshold ABR wave peak amplitudes reflect poor PNOTs (Figure 8, wave I, II).
In addition, significant latency shifts were observed in participants with poor speech comprehension in comparison to those with good or standard speech comprehension, as shown for wave II (Figure 8, II, good: 2.66 ms, n = 24; standard: 2.63 ms, n = 22; poor: 2.85 ms, n = 16; p = 0.00771), wave V (Figure 8, V, good: 5.63 ms, n = 30; standard: 5.56 ms, n = 28; poor: 5.76 ms, n = 28; p = 0.027), and wave VI (Figure 8, VI, good: 7.17 ms, n = 27; standard: 7.04 ms, n = 27; poor: 7.37 ms, n = 24; p = 0.0011). It is important to note that these latency differences in the good and poor PNOT groups in quiet did not differ by age (Table 1), or by PTA-EHF (Table 1), and thus exist in addition to the observed age-dependent supra-threshold amplitude and latency differences (Figure 3). Delayed ABR latencies of wave I and V survived the most restrictive post hoc linear mixed-model analysis after permutation (p = 0.004–0.030), and significantly explained 2.2% and 2.2% of OLSA-BB variance, and 4.6% and 3.2% of OLSA-LP variation in quiet; this, however, only corresponded to 0.8 to 1.3 dB of the SRT50 variation. ABR wave VI latency explained 4.2% of the OLSA-HP in quiet (Supplementary Figure S1, Supplementary Table S3).
In summary: Poor speech comprehension in quiet that remained when OLSA thresholds were normalized for PTT was linked to significantly delayed supra-threshold ABR I–VI peak amplitudes, and slightly smaller early ABR wave I–II amplitudes. This indicates that reduced neural response and/or synchronization at stimulus onset contribute to poor speech comprehension in quiet, independent of PTT.

3.9. Delta to Poor and Good PNOTs Show Differences in Phoneme Discrimination below and above the PLL

Aiming next to provide insight if poor and good speech comprehension independent of PTT and age may be linked to a difference in TFS or TENV coding, we presented phoneme pairs with formant contrasts below the PLL requiring TFS coding (/o/-/u/ and /du/-/bu/) or above the PLL requiring TENV coding (/i/-/y/ and /di/-/bi/; Table 2).
All phoneme pairs were presented in randomized blocks to the right ear in quiet or ipsilateral noise (Figure 9). For all tested phoneme pairs, two grades of difficulties were chosen, depending on the size of the physical contrast (here, labeled as “difficult” and “easy”). When the discrimination ability in percent was plotted as a function of age for phoneme-pair discrimination in quiet and ipsilateral noise conditions, a weak correlation was found in the quiet condition for /di/-/bi/ and in the ipsilateral noise condition for /du/-/bu/.
The ability to discriminate between the phoneme pairs in quiet and ipsilateral noise was plotted against the PNOTs obtained from the corresponding groups (Figure 9). Both easy and difficult discrimination conditions were averaged. Supplementary Table S4 provides statistics on behavioral accuracy.
In general, the performance of all participants was better for the discrimination of /di/-/bi/ than of /du/-/bu/. Thus, the phoneme /du/-/bu/ (Figure 9A,B, /du-bu/) showed the smallest variation in behavioral results across the cohort, with performance exceeding the 66th percentile only for a single participant (CS083). This was regardless of age; on average 29.4% to 58.9% of the participants responded below or at the 33rd percentile mark (=chance level), depending on the noise condition. The highest percentage of correct behavioral responses was achieved for discrimination of /o/-/u/ and /di/-/bi/, less for /i/-/y/ (Figure 9A,B).
The most prominent difference between participants with good or poor PNOT was in the differentiation of /o/-/u/, with formant contrasts below the PLL, as shown for quiet (Figure 9A), and ipsilateral noise (Figure 9B). When PNOT categorization in quiet (Figure 9A) was analyzed for phoneme discrimination, it became evident that the /o/-/u/ discrimination performance in participants with poor speech comprehension in quiet was poorer than in the group with standard speech comprehension (Figure 9A). Even under easy conditions, in which the two stimuli had large spectral differences, groups with poor speech comprehension in quiet performed worse in comparison to those with good or standard speech comprehension in quiet (Supplementary Table S4). On the other hand, participants with good speech comprehension (categorized from PNOT) in quiet (Figure 9A, /i/-/y/) were significantly better in their discrimination of /i/-/y/ in comparison to participants with standard speech discrimination ability.
Under ipsilateral noise conditions, participants with good speech comprehension (categorized from PNOTs) were better able to discriminate between /o/-/u/ than participants with standard speech comprehension. However, groups with poor speech comprehension were poorer at discriminating between /i/-/y/ in comparison to those with standard speech comprehension (Figure 9B, /i/-/y/) for both easy and difficult discrimination conditions in ipsilateral noise (Supplementary Table S4).
The discrimination of /du/-/bu/ was not different between participants with good and poor speech comprehension (categorized from PNOTs) in quiet and ipsilateral noise (Figure 9A,B, /du/-bu/), likely because the performance rate among all participants almost never exceeded 60%.
Further, groups with good, standard, and poor speech comprehension (categorized from PNOTs) did not differ in their discrimination ability between /di/-/bi/ under any of the listening conditions (Figure 9A,B, /di/-/bi/), likely because the performance rate among all participants almost always exceeded 90%.
Overall, our findings show that good and poor speech comprehension in quiet differs from good and poor speech comprehension in ipsilateral noise in its discrimination ability of formant contrasts below the PLL (requiring TFS coding), and above the PLL (requiring TENV coding) (Figure 10).
In quiet, poor speech comprehension was associated with poor discrimination of phoneme pairs with formant contrasts below the PLL (/o/-/u/), while good speech comprehension was associated with better discrimination of phoneme pairs with formant contrasts above the PLL (/i/-/y/) (Figure 10, quiet).
In ipsilateral noise, poor speech comprehension was associated with poorer discrimination of phoneme pairs with formant contrasts above the PLL (/i/-/y/), while good speech comprehension was associated with good discrimination of phoneme pairs with formant contrasts below the PLL (/o/-/u/) (Figure 10, ipsilateral noise).
Finally, the differentiation of consonant-based phoneme contrasts that required TFS coding (i.e., /du/-/bu/) was too difficult for both good and poor PNOT groups, while the phoneme contrasts that required TENV coding (i.e., /di/-/bi/) were too easy for both good and poor PNOT groups. Neither of these stimulus pairs therefore resulted in any group differences. Thus, the dynamic range of the phoneme task as implemented here was insufficient for differentiating the influence of PNOTs for /du/-/bu/ and /di/-/bi/ for the different speech-coding mechanisms (TFS vs. TENV).

4. Discussion

The present study investigated contributing factors of SRT50 for young, middle-aged, and older participants with mostly normal hearing, or mild hearing loss up to 8 kHz. We found that the PTT, although it was the most dominant factor for SRT50, only explained approximately half of the variance in quiet and in noise. The variance around this dominant relationship between the PTT and SRT50 was then operationally split into three groups of relatively good, standard, and poor speech comprehension, after removing PTT influence by the PNOT method. By analyzing the contributions to the remaining variance (pDPOAEs, ASSRs, supra-threshold ABR wave analysis), a previously undescribed influence of cochlear amplifier efficiency and effectivity and/or synchronicity changes at stimulus onset can be described as contributing to good and poor speech understanding in quiet and noise, regardless of hearing threshold and age. Hair-cell transmission weakness at the beginning of the stimulus should be urgently considered in future for inclusion in clinical diagnostics as a possible cause of speech intelligibility deficits in the young and old.

4.1. PTTs and SRT50 Show Age-Dependent Differences

In line with previous studies [13,36,86], we observed minor hearing loss in lower frequency ranges (PTA4 and PTA-LF) and prominent hearing loss at HF and EHF (PTA-HF and PTA-EHF) with increasing age (Figure 1). Also, using a differently filtered OLSA spectrum for testing speech comprehension in quiet and ipsilateral noise conditions, in analogy to [40,66] (Figure 2), we here showed speech intelligibility to depend significantly on age under all conditions. Particularly high-pass filtering of the speech material at 1.5 kHz (OLSA-HP) led to the steepest dependence on age (Figure 2). Although frequency spectra beyond 8 kHz have traditionally been noted to exhibit a limited perceptual role in speech sound quality [87,88], evidence is accumulating that high-frequency energy provides at least non-qualitative perceptual information, including cues for speech-source localization and intelligibility (reviewed in [34,36]). From the present study, we cannot exclude the possibility that the considerable loss of PTA-EHF over age (Figure 1) and the strong influence of age on the HP-filtered OLSA spectrum (Figure 2) are related events, and elevated PTA-EHF negatively impacts speech comprehension through lowering the late supra-threshold ABR wave, as discussed below.

4.2. Supra-Threshold ABR Wave Decrease with Elevated Age-Dependent PTA-EHF

The peak amplitudes of supra-threshold ABR waves are defined through the precise discharge rate of IHCs onto individual ANFs [89] and the precision with which ANFs fire synchronously at the sound onset [90]. The synchronous firing rate at the onset of the stimulus is a feature that critically depends upon the sensitivity of high-spontaneous firing rate (high-SR) low-threshold ANFs, which define latencies and perception thresholds [91,92,93,94]. In contrast, low-spontaneous firing rate (low-SR) high-threshold ANFs contribute little to the synchronization of ANFs [95]. In the present study, ABR wave I amplitudes were found to be reduced in middle-aged and older subjects (Figure 3A), as also observed in previous work [39,66,96], suggesting that an age-dependent synaptopathy exists in humans. In the present study, the central ABR wave amplitude differed between middle-aged and older individuals, being linked to sustained reduced and delayed ABR wave III–VI in older, but not young or middle-aged individuals (Figure 3A). This lack of compensation in the older group could be associated with a considerable loss of PTA-EHF (Figure 3D), but not PTA-4 or PTA-HF (Figure 3B,C). This relates to a specific effect of EHF on the coding of acoustic signals below 8 kHz, particularly if we take filtering of the in-ear-loudspeakers < 8 kHz used in the present study into account (see Section 2.5). Previous findings analyzing frequency-following responses (FFR) in humans observed a moderate influence of PTA-EHF on lower-frequency changes. Thus, FFR response amplitudes that are the periodic responses to the TFS of frequencies < 1.5 kHz of pure tones, thus below the PLL [97], were negatively influenced by elevated PTA-EHF [79]. While deficits in age-dependent temporal resolution are expected to typically diminish phase-locking at higher stimulation rates [98,99], the present (Figure 2 and Figure 3) and previous [79] findings suggest that in humans, EHF hearing loss may impact synchronous activity at lower frequencies. As low-SR high threshold fibers hardly contribute to the synchronicity of ANFs [95], and OHC dysfunction had no negative effect on temporal coding when phase-locked ANF responses were measured using FFR protocols [79], we may conclude that high-SR low-threshold ANFs driven by low-frequency tones may contribute to the observed differences in central response amplitudes between middle-aged and older individuals (Figure 3A). It remains to be clarified in future studies whether elevated PTA-EHF can influence the transformation of high-SR low-threshold ANFs to a nominal low-SR high-threshold ANF phenotype, as predicted when acoustic overexposure damage of stereocilia contributes to raised thresholds [100]. This aspect should also be taken into account in other studies that describe the influence of PTA-EHF on speech comprehension [101].

4.3. Difference between Good and Poor PNOTs Is a Better Indicator of Self-Assessed Hearing Ability than Age

Although SRTs in quiet strongly depend on PTTs, variance analysis in the present study indicates that 38.7% of the variance of OLSA-BB remained unexplained, which corresponds to an SD of 3.7 dB in SRT50 (Supplementary Figure S1). SRT50 in ipsilateral noise also depends on PTTs, but leaves more of the variance unexplained (Supplementary Figure S1, i.e., 51.4%, corresponding to an SD of 1.0 dB). When speech comprehension differences were normalized for PTT, subjects could be grouped into three categories based on their PNOT (poor, standard, and good). Interestingly, this PNOT categorization revealed a better association between the self-reported evaluations of speech comprehension ability when listening to speech in quiet than when subjects were categorized by age (Figure 5). Although, this correlation was only a statistical tendency, it strengthens previous findings of no correlation between self-reported speech comprehension ability and age [102]. This, moreover, challenges the hypothesis that those factors that influence speech understanding regardless of age and PTT, here identified in the broadest sense as strength or weakness of sound transfer at stimulus onset, may have relevance for the self-assessment of how (well) we hear.

4.4. Difference between Good and Poor PNOTs Show Low Dependence on Temporal Coding (ASSR)

As found in previous studies [103,104], the present finding confirmed that the ASSR amplitude declined with age, although only showing a statistical tendency (Figure 6A). Individuals with poor or good speech comprehension independent of PTT appear equally distributed over age (Figure 6A, orange and blue dots). On the other hand, higher ASSR amplitudes were found in poor PNOTs in quiet and ipsilateral noise conditions (Figure 6B) and higher ASSR amplitude in poor PNOTs strongly correlated with OLSA SRT in ipsilateral noise (Figure 6C). Considering an explanation for this surprising finding, we note that the ASSR growth function is known to correlate well with loudness [43,105,106]. This may suggest that our data are driven by a subgroup of poor performers who show extraordinarily high ASSR amplitudes (Figure 6B) and a significant increase in UCL (Supplementary Figure S1), thus exhibiting maladaptive loudness sensation. Here, the underlying mechanism may be linked to a change in the compressive nonlinearity that was suggested to contribute to categorical differences in loudness scaling with steeper growth in loudness for older adults [107], and here possibly to differences in cochlear amplifier efficiency observed in poor PNOTs in ipsilateral noise conditions, as discussed next.

4.5. Differences between Good and Poor PNOTs Are Reflected in Variations in Cochlear Amplifier Efficacy at Stimulus Onset

Poor and good speech comprehension, independent of age and PTT, differed between LEDPT and PTT (Figure 7), explaining 2% and 8.3% of the variance in OLSA-BB and OLSA-HP, respectively. A plausible reason for this effect is that LEDPT, which is in general closely related to the PTT, is not subject to adaptation of the ANF firing rate [108] or to adaptation caused by the medial olivocochlear reflex, since its time constant [109,110] is well above the DPOAE stimulus pulse widths used here for pDPOAE growth functions. As mentioned before, PTTs as implemented in clinical audiometry are effectively integrated over ~500 ms, reflecting the adapted state of nerve firing [111], while pDPOAEs would rather provide information on the non-adapted, pre-neural input signal to the IHCs [44]. In mammalian IHCs, weakness of nerve adaptation at the onset is associated with vesicle depletion, and is characteristically linked to synaptic fatigue or the desensitization kinetics of postsynaptic receptors [111,112,113,114]. Thus, a larger or smaller LEDPT–PTT difference that correlates with better or worse speech-in-quiet comprehension (Figure 7A,D) would reflect stronger or lower firing-rate adaptation, linked to less or more synaptic fatigue at IHCs. These are factors that now need to be considered as a plausible mechanism for differences in the detection of signal-onset features during speech presentation.
In the current data, LEDPT–PTT and the acceptance rate of DPOAE I/O functions under ipsilateral noise conditions (Figure 7, Supplementary Table S3) explained 7.0% and 5.5% of the remaining variance of SRT50 in the OLSA-LP condition, respectively corresponding to 0.3 dB in SRT50 for both measures, and explaining 0.7 and 3.1% in the OLSA-HP condition in quiet, respectively, corresponding to 0.5 and 1.1 dB in SRT50. The sign of the variance in LEDPT–PTT and the acceptance rate switched in ipsilateral noise in comparison to the quiet condition (Supplementary Figure S1, Supplementary Table S3), indicating that in ipsilateral noise, the larger DPOAE I/O function acceptance rate and larger LEDPT–PTT are linked to poorer speech comprehension (Figure 7E–H). This finding might be explained by compression of the cochlear input signal to the neural system. For the speech-in-noise test, temporal information is only used within a narrow dynamic level range. If, then, the DPOAE I/O function acceptance rate is comparatively poorer, it would be at levels above which basilar-membrane compression basically ends, and the growth behavior approaches linear dependency, meaning growth behavior would start at lower levels. This may be, on one side, an advantage, because a larger part of the dynamic level range used in the test would be almost linear, yielding uncompromised modulation contrast of the speech signal, as also previously discussed [29]. On the other side, this phenomenon, known as recruitment [115,116], shapes not only DPOAE I/O functions [117] or loudness scaling [107], thereby providing a rationale for the larger ASSR and UCL responses observed here, but also limits high TENV coding as discussed later.

4.6. Differences in Good and Poor PNOTs Are Reflected in Variations in Supra-Threshold Amplitude and Response Latencies of Auditory Nerve Fibers

Strikingly, a functional synaptopathy was evident as an ABR wave I amplitude reduction and an ABR wave II–VI latency shift in participants with poor PNOTs in quiet (Figure 8), which were of similar age to those with good PNOTs in quiet groups (Table 1). This factor explained 2.8%–1.0 dB—of OLSA-LP (ABR wave I) and 4.6%–1.3 dB—of OLSA-LP (ABR wave I latency), when SRT50 was corrected for PTT (Supplementary Table S3). Independent of OHC loss, which is expected to define the PTT, cochlear synaptopathy is not expected to explain more than 1 dB of OLSA. Indeed, a model of the effects of synapse loss on basic perceptual tasks, which calculated the effects of more than 50% loss of synapses, came up with a just-noticeable difference of up to 1.4 dB [118,119]. In contrast to previous findings reviewed by [120] that argued against a direct role for cochlear synaptopathy in the coding of moderate-to-high-level speech sounds, or assumptions that speech comprehension deficits in quiet are mainly linked to an increase in hearing thresholds [28], the present findings provide strong evidence that cochlear synaptopathy in humans exists even independently of age and PTT, as shown through reduced early and delayed early and late ABR waves in poor PNOTs (Figure 8). The peak amplitudes of supra-threshold ABR waves are defined through the precise discharge rate of individual auditory fibers [89] and the precision with which auditory fibers fire synchronously at the onset [90]. Synchronous firing at the onset of the stimulus is a feature that is critically dependent on the sensitivity of high-SR low-threshold ANFs, which not only define latencies and perception thresholds [91,92,93,94], but through its specific contribution to the rise in spike rate at the onset of sound stimulation [121,122], also define synchronized ANF responses at stimulus onset [95,123]. Therefore, differences in early ABR wave I/II amplitudes and late ABR wave latencies, as observed between subjects with poor and good PNOTs in quiet (Figure 8), may be best explained by differences in high-SR ANF functions at stimulus onset that influence speech comprehension due to an altered impact on synchronized ANFs at stimulus onset. In conclusion, as discussed for the lower pDPOAE acceptance rate in poor PNOTs, a weaker firing-rate onset peak as a result of pre- or postsynaptic changes linked to synaptic fatigue or a desensitization kinetic at IHCs [111,112,113,114] should be discussed as a novel contributor to speech comprehension in quiet, independent of age and PTT.

4.7. Differences in Good and Poor PNOTs Are Reflected in Variations in Phoneme Discrimination below and above the PLL

As a most striking feature of the phoneme discrimination test in PNOT groups, we observed poorer speech coding below the PLL in poor PNOT in quiet. Previous findings suggested a crucial role of high-SR ANFs for a perceptional threshold in the phase-locking range [56,124], which must be seen in connection with poorer speech coding below the PLL (Figure 9A, /o/-/u/, Figure 10, Supplementary Table S4). The poor discrimination of phonemes with formant contrasts above the PLL (/i/-/y/) in subjects with poor vs. standard PNOT (Figure 9) is best explained through the higher acceptance rate and LEDPT–PTT difference in this group (Figure 7E,H). This is likely linked to the lower basilar-membrane compression, and subsequent diminished width of the dynamic range of low- and medium-SR fibers, required for TENV coding [6], which provides a disadvantage for the differentiation of formant contrasts above the PLL (Figure 9, Supplementary Table S3, Figure 10).

5. Conclusions

In conclusion, apart from the dominating threshold dependence evidenced by PTTs, and typically used in clinical routine, we discovered several effects that affect speech discrimination independently of PTTs, which differ depending on whether the speech signal is close to threshold (speech-in-quiet), or clearly supra-threshold (speech-in-noise). Thus, using diagnostic procedures that enable the detection of changes in auditory processing at the beginning of the stimulus, such as pDPOAEs, ABR peak amplitudes, the phoneme discrimination test, and ASSRs as a metric of temporal coding, we identified elements that contribute to speech comprehension independent of age and PTT. As new factors contributing to speech comprehension in quiet, by comparison between DPOAEs and PTTs, we here identified the state of the cochlear amplifier and high-SR cochlear synaptopathy, influencing synchronized ANF responses at stimulus onset.
In noise, it appears that the recruitment phenomenon can partially counteract the discrimination deficits brought about by hearing loss due to reduced cochlear amplification. Differences in the nerve adaptation rate at stimulus onset in quiet and the recruitment phenomenon in noise must therefore be re-included in the 50% differences in human speech understanding that were previously not explained by hearing thresholds, and that likely contribute to the predicted 80% of auditory information that is transmitted at stimulus onset during speech [58]. These elements should be considered as a new mechanism behind the different coding principles predicted to be dependent on the PLL [40,57]. The findings also emphasize the need for improved routine clinical techniques to diagnose sound processing at stimulus onset.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/jcm13092725/s1, Supplementary Figure S1: Post hoc analysis of OLSA variance based on a linear mixed model; Supplementary Table S1: Proband information; Supplementary Table S2: Frequencies of the first four formants; Supplementary Table S3: Post hoc analysis of OLSA variance based on a linear mixed model; Supplementary Table S4: p-values for syllable discrimination scores for different speech performers. Supplementary Material pDPOAE measurements. References [44,67,68,69,70,71,72,125,126,127,128,129] has been cited in the Supplementary Materials file.

Author Contributions

Conceptualization: S.W., E.G., S.V., L.R., M.H.J.M., E.D. and M.K.; Methodology: J.S., S.W., K.D., W.S. and D.B.; Formal Analysis: J.S., K.D., M.R., J.W., M.W., T.E., K.B. and W.S.; Investigation: J.S.; Writing—original draft: L.R. and M.K.; Writing—review and editing: S.W., E.G., D.B., S.V., C.B., L.R., M.H.J.M. and E.D.; Supervision: M.H.J.M., E.D. and M.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Deutsche Forschungsgemeinschaft DFG KN 316/13-1, DFG RU 713/6-1, KL 1093/12-1; ERA-NET NEURON JTC 2020: BMBF 01EW2102 CoSySpeech and FWO G0H6420N; IZKF Promotionskolleg of the Faculty of Medicine, University Hospital of Tübingen. VICI Grant (Grant No. 918-17-603), Netherlands Organization for Scientific Research (NWO) and the Netherlands Organization for Health Research and Development (ZonMw).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of the University of Tübingen (ethical approval number 392/2021BO2, date of approval: 30 June 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We thank the audiometrists of the HNO-Klinik Tübingen for their expertise in recording audiograms. We thank Geoffrey Manley for excellent English-language service.

Conflicts of Interest

The authors have no conflicts of interest to declare. All co-authors have seen and agreed with the manuscript’s contents, and there is no financial interest to report.

References

  1. Livingston, G.; Sommerlad, A.; Orgeta, V.; Costafreda, S.G.; Huntley, J.; Ames, D.; Ballard, C.; Banerjee, S.; Burns, A.; Cohen-Mansfield, J.; et al. Dementia prevention, intervention, and care. Lancet 2017, 390, 2673–2734. [Google Scholar] [CrossRef] [PubMed]
  2. Golub, J.S.; Brickman, A.M.; Ciarleglio, A.J.; Schupf, N.; Luchsinger, J.A. Association of Subclinical Hearing Loss With Cognitive Performance. JAMA Otolaryngol. Head Neck Surg. 2020, 146, 57–67. [Google Scholar] [CrossRef] [PubMed]
  3. Hoppe, U.; Hocke, T.; Iro, H. Age-Related Decline of Speech Perception. Front. Aging Neurosci. 2022, 14, 891202. [Google Scholar] [CrossRef] [PubMed]
  4. Sergeyenko, Y.; Lall, K.; Liberman, M.C.; Kujawa, S.G. Age-related cochlear synaptopathy: An early-onset contributor to auditory functional decline. J. Neurosci. 2013, 33, 13686–13694. [Google Scholar] [CrossRef] [PubMed]
  5. Mohrle, D.; Ni, K.; Varakina, K.; Bing, D.; Lee, S.C.; Zimmermann, U.; Knipper, M.; Ruttiger, L. Loss of auditory sensitivity from inner hair cell synaptopathy can be centrally compensated in the young but not old brain. Neurobiol. Aging 2016, 44, 173–184. [Google Scholar] [CrossRef] [PubMed]
  6. Monaghan, J.J.M.; Garcia-Lazaro, J.A.; McAlpine, D.; Schaette, R. Hidden Hearing Loss Impacts the Neural Representation of Speech in Background Noise. Curr. Biol. 2020, 30, 4710–4721.e4. [Google Scholar] [CrossRef] [PubMed]
  7. Bharadwaj, H.M.; Mai, A.R.; Simpson, J.M.; Choi, I.; Heinz, M.G.; Shinn-Cunningham, B.G. Non-Invasive Assays of Cochlear Synaptopathy—Candidates and Considerations. Neuroscience 2019, 407, 53–66. [Google Scholar] [CrossRef] [PubMed]
  8. Mepani, A.M.; Verhulst, S.; Hancock, K.E.; Garrett, M.; Vasilkov, V.; Bennett, K.; de Gruttola, V.; Liberman, M.C.; Maison, S.F. Envelope following responses predict speech-in-noise performance in normal-hearing listeners. J. Neurophysiol. 2021, 125, 1213–1222. [Google Scholar] [CrossRef]
  9. Viana, L.M.; O’Malley, J.T.; Burgess, B.J.; Jones, D.D.; Oliveira, C.A.; Santos, F.; Merchant, S.N.; Liberman, L.D.; Liberman, M.C. Cochlear neuropathy in human presbycusis: Confocal analysis of hidden hearing loss in post-mortem tissue. Hear. Res. 2015, 327, 78–88. [Google Scholar] [CrossRef]
  10. Kobel, M.; Le Prell, C.G.; Liu, J.; Hawks, J.W.; Bao, J. Noise-induced cochlear synaptopathy: Past findings and future studies. Hear. Res. 2017, 349, 148–154. [Google Scholar] [CrossRef]
  11. Liberman, M.C.; Kujawa, S.G. Cochlear synaptopathy in acquired sensorineural hearing loss: Manifestations and mechanisms. Hear. Res. 2017, 349, 138–147. [Google Scholar] [CrossRef] [PubMed]
  12. Plack, C.J.; Barker, D.; Prendergast, G. Perceptual consequences of “hidden” hearing loss. Trends Hear. 2014, 18, 2331216514550621. [Google Scholar] [CrossRef] [PubMed]
  13. Fullgrabe, C.; Moore, B.C.; Stone, M.A. Age-group differences in speech identification despite matched audiometrically normal hearing: Contributions from auditory temporal processing and cognition. Front. Aging Neurosci. 2014, 6, 347. [Google Scholar] [CrossRef] [PubMed]
  14. Frisina, R.D. Age-related hearing loss: Ear and brain mechanisms. Ann. N. Y. Acad. Sci. 2009, 1170, 708–717. [Google Scholar] [CrossRef] [PubMed]
  15. Wu, P.Z.; Liberman, L.D.; Bennett, K.; de Gruttola, V.; O’Malley, J.T.; Liberman, M.C. Primary Neural Degeneration in the Human Cochlea: Evidence for Hidden Hearing Loss in the Aging Ear. Neuroscience 2019, 407, 8–20. [Google Scholar] [CrossRef]
  16. Chambers, A.R.; Resnik, J.; Yuan, Y.; Whitton, J.P.; Edge, A.S.; Liberman, M.C.; Polley, D.B. Central Gain Restores Auditory Processing following Near-Complete Cochlear Denervation. Neuron 2016, 89, 867–879. [Google Scholar] [CrossRef] [PubMed]
  17. Bakay, W.M.H.; Anderson, L.A.; Garcia-Lazaro, J.A.; McAlpine, D.; Schaette, R. Hidden hearing loss selectively impairs neural adaptation to loud sound environments. Nat. Commun. 2018, 9, 4298. [Google Scholar] [CrossRef]
  18. Asokan, M.M.; Williamson, R.S.; Hancock, K.E.; Polley, D.B. Publisher Correction: Sensory overamplification in layer 5 auditory corticofugal projection neurons following cochlear nerve synaptic damage. Nat. Commun. 2018, 9, 3158. [Google Scholar] [CrossRef]
  19. Hesse, L.L.; Bakay, W.; Ong, H.C.; Anderson, L.; Ashmore, J.; McAlpine, D.; Linden, J.; Schaette, R. Non-Monotonic Relation between Noise Exposure Severity and Neuronal Hyperactivity in the Auditory Midbrain. Front. Neurol. 2016, 7, 133. [Google Scholar] [CrossRef]
  20. Kraus, N.; Anderson, S.; White-Schwoch, T. The Frequency-Following Response: A Window into Human Communication. In The Frequency-Following Response: A Window into Human Communication; Kraus, N., Anderson, S., White-Schwoch, T., Fay, R.R., Popper, A.N., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 1–15. [Google Scholar]
  21. Chen, G.D. Hidden cochlear impairments. J. Otol. 2018, 13, 37–43. [Google Scholar] [CrossRef]
  22. De Siati, R.D.; Rosenzweig, F.; Gersdorff, G.; Gregoire, A.; Rombaux, P.; Deggouj, N. Auditory Neuropathy Spectrum Disorders: From Diagnosis to Treatment: Literature Review and Case Reports. J. Clin. Med. 2020, 9, 1074. [Google Scholar] [CrossRef] [PubMed]
  23. Iliadou, V.V.; Ptok, M.; Grech, H.; Pedersen, E.R.; Brechmann, A.; Deggouj, N.; Kiese-Himmel, C.; Sliwinska-Kowalska, M.; Nickisch, A.; Demanez, L.; et al. A European Perspective on Auditory Processing Disorder-Current Knowledge and Future Research Focus. Front. Neurol. 2017, 8, 622. [Google Scholar] [CrossRef] [PubMed]
  24. Yingling, C.D.; Gardi, J.N. Intraoperative monitoring of facial and cochlear nerves during acoustic neuroma surgery. Otolaryngol. Clin. North Am. 1992, 25, 413–448. [Google Scholar] [CrossRef] [PubMed]
  25. Eggink, M.C.; Frijns, J.H.M.; Sagers, J.E.; O’Malley, J.T.; Liberman, M.C.; Stankovic, K.M. Human vestibular schwannoma reduces density of auditory nerve fibers in the osseous spiral lamina. Hear. Res. 2022, 418, 108458. [Google Scholar] [CrossRef] [PubMed]
  26. Le Prell, C.G. Effects of noise exposure on auditory brainstem response and speech-in-noise tasks: A review of the literature. Int J. Audiol. 2019, 58, S3–S32. [Google Scholar] [CrossRef] [PubMed]
  27. Snell, K.B.; Frisina, D.R. Relationships among age-related differences in gap detection and word recognition. J. Acoust. Soc. Am. 2000, 107, 1615–1626. [Google Scholar] [CrossRef] [PubMed]
  28. Johannesen, P.T.; Perez-Gonzalez, P.; Kalluri, S.; Blanco, J.L.; Lopez-Poveda, E.A. The Influence of Cochlear Mechanical Dysfunction, Temporal Processing Deficits, and Age on the Intelligibility of Audible Speech in Noise for Hearing-Impaired Listeners. Trends Hear. 2016, 20, 2331216516641055. [Google Scholar] [CrossRef]
  29. Marrufo-Perez, M.I.; Lopez-Poveda, E.A. Adaptation to noise in normal and impaired hearing. J. Acoust. Soc. Am. 2022, 151, 1741. [Google Scholar] [CrossRef]
  30. Ayasse, N.D.; Penn, L.R.; Wingfield, A. Variations Within Normal Hearing Acuity and Speech Comprehension: An Exploratory Study. Am. J. Audiol. 2019, 28, 369–375. [Google Scholar] [CrossRef]
  31. Smith, R.L.; Zwislocki, J.J. Short-term adaptation and incremental responses of single auditory-nerve fibers. Biol. Cybern. 1975, 17, 169–182. [Google Scholar] [CrossRef]
  32. von Gablenz, P.; Holube, I. Hearing Loss and Speech Recognition in the Elderly. Laryngorhinootologie 2017, 96, 759–764. [Google Scholar] [CrossRef] [PubMed]
  33. den Besten, C.A.; Monksfield, P.; Bosman, A.; Skarzynski, P.H.; Green, K.; Runge, C.; Wigren, S.; Blechert, J.I.; Flynn, M.C.; Mylanus, E.A.M.; et al. Audiological and clinical outcomes of a transcutaneous bone conduction hearing implant: Six-month results from a multicentre study. Clin. Otolaryngol. 2019, 44, 144–157. [Google Scholar] [CrossRef] [PubMed]
  34. Monson, B.B.; Hunter, E.J.; Lotto, A.J.; Story, B.H. The perceptual significance of high-frequency energy in the human voice. Front. Psychol. 2014, 5, 587. [Google Scholar] [CrossRef] [PubMed]
  35. Sliwinska-Kowalska, M. Hearing. Handb. Clin. Neurol. 2015, 131, 341–363. [Google Scholar] [CrossRef] [PubMed]
  36. Hunter, L.L.; Monson, B.B.; Moore, D.R.; Dhar, S.; Wright, B.A.; Munro, K.J.; Zadeh, L.M.; Blankenship, C.M.; Stiepan, S.M.; Siegel, J.H. Extended high frequency hearing and speech perception implications in adults and children. Hear. Res. 2020, 397, 107922. [Google Scholar] [CrossRef]
  37. Lorenzi, C.; Gilbert, G.; Carn, H.; Garnier, S.; Moore, B.C. Speech perception problems of the hearing impaired reflect inability to use temporal fine structure. Proc. Natl. Acad. Sci. USA 2006, 103, 18866–18869. [Google Scholar] [CrossRef] [PubMed]
  38. Shaheen, L.A.; Valero, M.D.; Liberman, M.C. Towards a Diagnosis of Cochlear Neuropathy with Envelope Following Responses. J. Assoc. Res. Otolaryngol. 2015, 16, 727–745. [Google Scholar] [CrossRef] [PubMed]
  39. Parthasarathy, A.; Bartlett, E.L.; Kujawa, S.G. Age-related Changes in Neural Coding of Envelope Cues: Peripheral Declines and Central Compensation. Neuroscience 2019, 407, 21–31. [Google Scholar] [CrossRef] [PubMed]
  40. Garrett, M.; Vasilkov, V.; Mauermann, M.; Wilson, J.L.; Henry, K.S.; Verhulst, S. Speech-in-noise intelligibility difficulties with age: The role of cochlear synaptopathy. bioRxiv 2024, 2020.06.09.142950. [Google Scholar] [CrossRef]
  41. Rance, G. The Auditory Steady-State Response: Generation, Recording, and Clinical Application; Plural Publishing: San Diego, CA, USA, 2008. [Google Scholar]
  42. Kuwada, S.; Anderson, J.S.; Batra, R.; Fitzpatrick, D.C.; Teissier, N.; D’Angelo, W.R. Sources of the scalp-recorded amplitude-modulation following response. J. Am. Acad. Audiol. 2002, 13, 188–204. [Google Scholar] [CrossRef]
  43. Lu, H.; Mehta, A.H.; Oxenham, A.J. Methodological considerations when measuring and analyzing auditory steady-state responses with multi-channel EEG. Curr. Res. Neurobiol. 2022, 3, 100061. [Google Scholar] [CrossRef] [PubMed]
  44. Zelle, D.; Lorenz, L.; Thiericke, J.P.; Gummer, A.W.; Dalhoff, E. Input-output functions of the nonlinear-distortion component of distortion-product otoacoustic emissions in normal and hearing-impaired human ears. J. Acoust. Soc. Am. 2017, 141, 3203. [Google Scholar] [CrossRef]
  45. Portmann, M.; Cazals, Y.; Negrevergne, M.; Aran, J.M. Transtympanic and surface recordings in the diagnosis of retrocochlear disorders. Acta Oto-Laryngol. 1980, 89, 362–369. [Google Scholar] [CrossRef]
  46. Wei, L.; Ding, D.; Sun, W.; Xu-Friedman, M.A.; Salvi, R. Effects of sodium salicylate on spontaneous and evoked spike rate in the dorsal cochlear nucleus. Hear. Res. 2010, 267, 54–60. [Google Scholar] [CrossRef] [PubMed]
  47. Melcher, J.R.; Guinan, J.J., Jr.; Knudson, I.M.; Kiang, N.Y. Generators of the brainstem auditory evoked potential in cat. II. Correlating lesion sites with waveform changes. Hear. Res. 1996, 93, 28–51. [Google Scholar] [CrossRef] [PubMed]
  48. Møller, A.R.; Jannetta, P.J.; Jho, H.D. Click-evoked responses from the cochlear nucleus: A study in human. Electroencephalogr. Clin. Neurophysiol. 1994, 92, 215–224. [Google Scholar] [CrossRef]
  49. Hashimoto, I.; Ishiyama, Y.; Yoshimoto, T.; Nemoto, S. Brain-stem auditory-evoked potentials recorded directly from human brain-stem and thalamus. Brain 1981, 104, 841–859. [Google Scholar] [CrossRef]
  50. Hofmeier, B.; Wolpert, S.; Aldamer, E.S.; Walter, M.; Thiericke, J.; Braun, C.; Zelle, D.; Ruttiger, L.; Klose, U.; Knipper, M. Reduced sound-evoked and resting-state BOLD fMRI connectivity in tinnitus. Neuroimage Clin. 2018, 20, 637–649. [Google Scholar] [CrossRef]
  51. Hofmeier, B.; Wertz, J.; Refat, F.; Hinrichs, P.; Saemisch, J.; Singer, W.; Ruttiger, L.; Klose, U.; Knipper, M.; Wolpert, S. Functional biomarkers that distinguish between tinnitus with and without hyperacusis. Clin. Transl. Med. 2021, 11, e378. [Google Scholar] [CrossRef]
  52. Won, J.H.; Tremblay, K.; Clinard, C.G.; Wright, R.A.; Sagi, E.; Svirsky, M. The neural encoding of formant frequencies contributing to vowel identification in normal-hearing listeners. J. Acoust. Soc. Am. 2016, 139, 1–11. [Google Scholar] [CrossRef]
  53. Hornickel, J.; Skoe, E.; Nicol, T.; Zecker, S.; Kraus, N. Subcortical differentiation of stop consonants relates to reading and speech-in-noise perception. Proc. Natl. Acad. Sci. USA 2009, 106, 13022–13027. [Google Scholar] [CrossRef]
  54. Verschooten, E.; Shamma, S.; Oxenham, A.J.; Moore, B.C.J.; Joris, P.X.; Heinz, M.G.; Plack, C.J. The upper frequency limit for the use of phase locking to code temporal fine structure in humans: A compilation of viewpoints. Hear. Res. 2019, 377, 109–121. [Google Scholar] [CrossRef] [PubMed]
  55. Weiss, T.F.; Rose, C. A comparison of synchronization filters in different auditory receptor organs. Hear. Res. 1988, 33, 175–179. [Google Scholar] [CrossRef] [PubMed]
  56. Huet, A.; Desmadryl, G.; Justal, T.; Nouvian, R.; Puel, J.L.; Bourien, J. The Interplay Between Spike-Time and Spike-Rate Modes in the Auditory Nerve Encodes Tone-In-Noise Threshold. J. Neurosci. 2018, 38, 5727–5738. [Google Scholar] [CrossRef] [PubMed]
  57. Moore, B.C.J. Effects of hearing loss and age on the binaural processing of temporal envelope and temporal fine structure information. Hear. Res. 2021, 402, 107991. [Google Scholar] [CrossRef] [PubMed]
  58. Wang, H.; Isik, M.; Borst, A.; Hemmert, W. Auditory information coding by modeled cochlear nucleus neurons. J. Comput. Neurosci. 2011, 30, 529–542. [Google Scholar] [CrossRef]
  59. Beck, A.T.; Steer, R.A.; Ball, R.; Ranieri, W. Comparison of Beck Depression Inventories -IA and -II in psychiatric outpatients. J. Pers. Assess 1996, 67, 588–597. [Google Scholar] [CrossRef] [PubMed]
  60. Yesavage, J.A.; Brink, T.L.; Rose, T.L.; Lum, O.; Huang, V.; Adey, M.; Leirer, V.O. Development and validation of a geriatric depression screening scale: A preliminary report. J. Psychiatr. Res. 1982, 17, 37–49. [Google Scholar] [CrossRef] [PubMed]
  61. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  62. Tombaugh, T.N.; McIntyre, N.J. The mini-mental state examination: A comprehensive review. J. Am. Geriatr. Soc. 1992, 40, 922–935. [Google Scholar] [CrossRef]
  63. Maele, T.V.; Keshishzadeh, S.; Poortere, N.; Dhooge, I.; Keppler, H.; Verhulst, S. The Variability in Potential Biomarkers for Cochlear Synaptopathy After Recreational Noise Exposure. J. Speech Lang Hear. Res. 2021, 64, 4964–4981. [Google Scholar] [CrossRef] [PubMed]
  64. Shanks, J.E. Tympanometry. Ear Hear. 1984, 5, 268–280. [Google Scholar] [CrossRef] [PubMed]
  65. Margolis, R.H.; Goycoolea, H.G. Multifrequency tympanometry in normal adults. Ear Hear. 1993, 14, 408–413. [Google Scholar] [CrossRef] [PubMed]
  66. Vasilkov, V.; Garrett, M.; Mauermann, M.; Verhulst, S. Enhancing the sensitivity of the envelope-following response for cochlear synaptopathy screening in humans: The role of stimulus envelope. Hear. Res. 2021, 400, 108132. [Google Scholar] [CrossRef] [PubMed]
  67. Shera, C.A.; Guinan, J.J., Jr. Evoked otoacoustic emissions arise by two fundamentally different mechanisms: A taxonomy for mammalian OAEs. J. Acoust. Soc. Am. 1999, 105, 782–798. [Google Scholar] [CrossRef] [PubMed]
  68. Zelle, D.; Gummer, A.W.; Dalhoff, E. Extraction of distortion-product otoacoustic emission source components and its relevance for objective audiometry. Procedia IUTAM 2016, 100, 38–47. [Google Scholar] [CrossRef]
  69. Kummer, P.; Janssen, T.; Arnold, W. The level and growth behavior of the 2 f1-f2 distortion product otoacoustic emission and its relationship to auditory sensitivity in normal hearing and cochlear hearing loss. J. Acoust. Soc. Am. 1998, 103, 3431–3444. [Google Scholar] [CrossRef] [PubMed]
  70. Boege, P.; Janssen, T. Pure-tone threshold estimation from extrapolated distortion product otoacoustic emission I/O-functions in normal and cochlear hearing loss ears. J. Acoust. Soc. Am. 2002, 111, 1810–1818. [Google Scholar] [CrossRef]
  71. Gorga, M.P.; Neely, S.T.; Dorn, P.A.; Hoover, B.M. Further efforts to predict pure-tone thresholds from distortion product otoacoustic emission input/output functions. J. Acoust. Soc. Am. 2003, 113, 3275–3284. [Google Scholar] [CrossRef]
  72. Johnson, T.A.; Neely, S.T.; Kopun, J.G.; Dierking, D.M.; Tan, H.; Converse, C.; Kennedy, E.; Gorga, M.P. Distortion product otoacoustic emissions: Cochlear-source contributions and clinical test performance. J. Acoust. Soc. Am. 2007, 122, 3539–3553. [Google Scholar] [CrossRef]
  73. Wagener, K.; Kühnel, V.; Kollmeier, B. Development and evaluation of a German sentence test I: Design of the Oldenburg sentence test. Z. Audiol. 1999, 38, 4–15. [Google Scholar]
  74. Brand, T.; Kollmeier, B. Efficient adaptive procedures for threshold and concurrent slope estimates for psychophysics and speech intelligibility tests. J. Acoust. Soc. Am. 2002, 111, 2801–2810. [Google Scholar] [CrossRef] [PubMed]
  75. Wagener, K.C.; Brand, T.; Kollmeier, B. Development and evaluation of a German sentence test Part III: Evaluation of the Oldenburg sentence test. Z. Audiol. 1999, 38, 86–95. [Google Scholar]
  76. Wagener, K.C.; Brand, T. Sentence intelligibility in noise for listeners with normal hearing and hearing impairment: Influence of measurement procedure and masking parameters. Int. J. Audiol. 2005, 44, 144–156. [Google Scholar] [CrossRef]
  77. Byrne, D.; Dillon, H.; Tran, K.; Arlinger, S.; Wilbraham, K.; Cox, R.; Hagerman, B.; Hetu, R.; Kei, J.; Lui, C.; et al. An international comparison of long-term average speech spectra. J. Acoust. Soc. Am. 1994, 96, 2108–2120. [Google Scholar] [CrossRef]
  78. Morise, M.; Yokomori, F.; Ozawa, K. WORLD: A Vocoder-Based High-Quality Speech Synthesis System for Real-Time Applications. IEICE Trans. Inf. Syst. 2016, E99, 1877–1884. [Google Scholar] [CrossRef]
  79. Marcher-Rorsted, J.; Encina-Llamas, G.; Dau, T.; Liberman, M.C.; Wu, P.Z.; Hjortkjaer, J. Age-related reduction in frequency-following responses as a potential marker of cochlear neural degeneration. Hear. Res. 2022, 414, 108411. [Google Scholar] [CrossRef]
  80. Carcagno, S.; Plack, C.J. Relations between speech-reception, psychophysical temporal processing, and subcortical electrophysiological measures of auditory function in humans. Hear. Res. 2022, 417, 108456. [Google Scholar] [CrossRef] [PubMed]
  81. Osterhammel, D.; Osterhammel, P. High-frequency audiometry. Age and sex variations. Scand. Audiol. 1979, 8, 73–81. [Google Scholar] [CrossRef]
  82. Prendergast, G.; Hymers, M.; Lee, A. A quick and reliable estimate of extended high-frequency hearing. Int. J. Audiol. 2020, 59, 823–827. [Google Scholar] [CrossRef]
  83. Wardenga, N.; Batsoulis, C.; Wagener, K.C.; Brand, T.; Lenarz, T.; Maier, H. Do you hear the noise? The German matrix sentence test with a fixed noise level in subjects with normal hearing and hearing impairment. Int. J. Audiol. 2015, 54 (Suppl. S2), 71–79. [Google Scholar] [CrossRef] [PubMed]
  84. Joris, P.X.; Schreiner, C.E.; Rees, A. Neural processing of amplitude-modulated sounds. Physiol. Rev. 2004, 84, 541–577. [Google Scholar] [CrossRef] [PubMed]
  85. Eggermont, J.J.; Odenthal, D.W. Action potentials and summating potentials in the normal human cochlea. Acta Otolaryngol. Suppl. 1974, 316, 39–61. [Google Scholar] [CrossRef] [PubMed]
  86. Wu, P.Z.; Liberman, M.C. Age-related stereocilia pathology in the human cochlea. Hear. Res. 2022, 422, 108551. [Google Scholar] [CrossRef]
  87. Olson, H.F. Frequency Range Preference for Speech and Music. J. Acoust. Soc. Am. 1947, 19, 549–555. [Google Scholar] [CrossRef]
  88. Moore, B.C. Basic auditory processes involved in the analysis of speech sounds. Philos. Trans. R. Soc. Lond. B Biol. Sci. 2008, 363, 947–963. [Google Scholar] [CrossRef] [PubMed]
  89. Buran, B.N.; Strenzke, N.; Neef, A.; Gundelfinger, E.D.; Moser, T.; Liberman, M.C. Onset coding is degraded in auditory nerve fibers from mutant mice lacking synaptic ribbons. J. Neurosci. 2010, 30, 7587–7597. [Google Scholar] [CrossRef] [PubMed]
  90. Johnson, D.H.; Kiang, N.Y. Analysis of discharges recorded simultaneously from pairs of auditory nerve fibers. Biophys. J. 1976, 16, 719–734. [Google Scholar] [CrossRef] [PubMed]
  91. Rhode, W.S.; Smith, P.H. Encoding timing and intensity in the ventral cochlear nucleus of the cat. J. Neurophysiol. 1986, 56, 261–286. [Google Scholar] [CrossRef]
  92. Zohar, O.; Shackleton, T.M.; Nelken, I.; Palmer, A.R.; Shamir, M. First spike latency code for interaural phase difference discrimination in the guinea pig inferior colliculus. J. Neurosci. Off. J. Soc. Neurosci. 2011, 31, 9192–9204. [Google Scholar] [CrossRef]
  93. Meddis, R. Auditory-nerve first-spike latency and auditory absolute threshold: A computer model. J. Acoust. Soc. Am. 2006, 119, 406–417. [Google Scholar] [CrossRef] [PubMed]
  94. Heil, P.; Neubauer, H.; Brown, M.; Irvine, D.R. Towards a unifying basis of auditory thresholds: Distributions of the first-spike latencies of auditory-nerve fibers. Hear. Res. 2008, 238, 25–38. [Google Scholar] [CrossRef] [PubMed]
  95. Huet, A.; Batrel, C.; Wang, J.; Desmadryl, G.; Nouvian, R.; Puel, J.L.; Bourien, J. Sound Coding in the Auditory Nerve: From Single Fiber Activity to Cochlear Mass Potentials in Gerbils. Neuroscience 2019, 407, 83–92. [Google Scholar] [CrossRef] [PubMed]
  96. Bramhall, N.F. Use of the auditory brainstem response for assessment of cochlear synaptopathy in humans. J. Acoust. Soc. Am. 2021, 150, 4440. [Google Scholar] [CrossRef] [PubMed]
  97. Batra, R.; Kuwada, S.; Maher, V.L. The frequency-following response to continuous tones in humans. Hear. Res. 1986, 21, 167–177. [Google Scholar] [CrossRef] [PubMed]
  98. Anderson, S.; Parbery-Clark, A.; White-Schwoch, T.; Kraus, N. Aging affects neural precision of speech encoding. J. Neurosci. 2012, 32, 14156–14164. [Google Scholar] [CrossRef] [PubMed]
  99. Frisina, D.R.; Frisina, R.D. Speech recognition in noise and presbycusis: Relations to possible neural mechanisms. Hear. Res. 1997, 106, 95–104. [Google Scholar] [CrossRef] [PubMed]
  100. Liberman, M.C.; Dodds, L.W. Single-neuron labeling and chronic cochlear pathology. III. Stereocilia damage and alterations of threshold tuning curves. Hear. Res. 1984, 16, 55–74. [Google Scholar] [CrossRef] [PubMed]
  101. Motlagh Zadeh, L.; Silbert, N.H.; Sternasty, K.; Swanepoel, W.; Hunter, L.L.; Moore, D.R. Extended high-frequency hearing enhances speech perception in noise. Proc. Natl. Acad. Sci. USA 2019, 116, 23753–23759. [Google Scholar] [CrossRef]
  102. Shehabi, A.M.; Prendergast, G.; Plack, C.J. The Relative and Combined Effects of Noise Exposure and Aging on Auditory Peripheral Neural Deafferentation: A Narrative Review. Front. Aging Neurosci. 2022, 14, 877588. [Google Scholar] [CrossRef]
  103. Boettcher, F.A.; Poth, E.A.; Mills, J.H.; Dubno, J.R. The amplitude-modulation following response in young and aged human subjects. Hear. Res. 2001, 153, 32–42. [Google Scholar] [CrossRef] [PubMed]
  104. Rumschlag, J.A.; Razak, K.A. Age-related changes in event related potentials, steady state responses and temporal processing in the auditory cortex of mice with severe or mild hearing loss. Hear. Res. 2021, 412, 108380. [Google Scholar] [CrossRef] [PubMed]
  105. Menard, M.; Gallego, S.; Berger-Vachon, C.; Collet, L.; Thai-Van, H. Relationship between loudness growth function and auditory steady-state response in normal-hearing subjects. Hear. Res. 2008, 235, 105–113. [Google Scholar] [CrossRef] [PubMed]
  106. Oxenham, A.J. How We Hear: The Perception and Neural Coding of Sound. Annu. Rev. Psychol. 2018, 69, 27–50. [Google Scholar] [CrossRef] [PubMed]
  107. Abdala, C.; Ortmann, A.J.; Guardia, Y.C. Weakened Cochlear Nonlinearity During Human Aging and Perceptual Correlates. Ear Hear. 2021, 42, 832–845. [Google Scholar] [CrossRef] [PubMed]
  108. Kiang, N.Y.; Watanabe, T.; Thomas, E.C.; Clark, L.F. Discharge Patterns of Single Fibers in the Cat’s Auditory Nerve; M.I.T. Press: Oxford, UK, 1966; p. xvii, 154. [Google Scholar]
  109. Kim, D.O.; Dorn, P.A.; Neely, S.T.; Gorga, M.P. Adaptation of distortion product otoacoustic emission in humans. J. Assoc. Res. Otolaryngol. 2001, 2, 31–40. [Google Scholar] [CrossRef] [PubMed]
  110. Bassim, M.K.; Miller, R.L.; Buss, E.; Smith, D.W. Rapid adaptation of the 2f1-f2 DPOAE in humans: Binaural and contralateral stimulation effects. Hear. Res. 2003, 182, 140–152. [Google Scholar] [CrossRef] [PubMed]
  111. Goutman, J.D. Mechanisms of synaptic depression at the hair cell ribbon synapse that support auditory nerve function. Proc. Natl. Acad. Sci. USA 2017, 114, 9719–9724. [Google Scholar] [CrossRef]
  112. Moser, T.; Beutner, D. Kinetics of exocytosis and endocytosis at the cochlear inner hair cell afferent synapse of the mouse. Proc. Natl. Acad. Sci. USA 2000, 97, 883–888. [Google Scholar] [CrossRef]
  113. Peterson, A.J.; Huet, A.; Bourien, J.; Puel, J.L.; Heil, P. Recovery of auditory-nerve-fiber spike amplitude under natural excitation conditions. Hear. Res. 2018, 370, 248–263. [Google Scholar] [CrossRef]
  114. Willmore, B.D.B.; King, A.J. Adaptation in auditory processing. Physiol. Rev. 2023, 103, 1025–1058. [Google Scholar] [CrossRef] [PubMed]
  115. Kubota, T.; Ito, T.; Abe, Y.; Chiba, H.; Suzuki, Y.; Kakehata, S.; Aoyagi, M. Detecting the recruitment phenomenon in adults using 80-Hz auditory steady-state response. Auris Nasus Larynx 2019, 46, 696–702. [Google Scholar] [CrossRef] [PubMed]
  116. Denes, P.; Naunton, R.F. The clinical detection of auditory recruitment. J. Laryngol. Otol. 1950, 64, 375–398. [Google Scholar] [CrossRef] [PubMed]
  117. Rasetshwane, D.M.; Argenyi, M.; Neely, S.T.; Kopun, J.G.; Gorga, M.P. Latency of tone-burst-evoked auditory brain stem responses and otoacoustic emissions: Level, frequency, and rise-time effects. J. Acoust. Soc. Am. 2013, 133, 2803–2817. [Google Scholar] [CrossRef] [PubMed]
  118. Suthakar, K.; Liberman, M.C. Noise Masking in Cochlear Synaptopathy: Auditory Brainstem Response vs. Auditory Nerve Response in Mouse. J. Neurophysiol. 2022, 127, 1574–1585. [Google Scholar] [CrossRef] [PubMed]
  119. Oxenham, A.J. Predicting the Perceptual Consequences of Hidden Hearing Loss. Trends Hear. 2016, 20, 2331216516686768. [Google Scholar] [CrossRef] [PubMed]
  120. Carney, L.H. Supra-Threshold Hearing and Fluctuation Profiles: Implications for Sensorineural and Hidden Hearing Loss. J. Assoc. Res. Otolaryngol. 2018, 19, 331–352. [Google Scholar] [CrossRef] [PubMed]
  121. Chimento, T.C.; Schreiner, C.E. Adaptation and recovery from adaptation in single fiber responses of the cat auditory nerve. J. Acoust. Soc. Am. 1991, 90, 263–273. [Google Scholar] [CrossRef] [PubMed]
  122. Young, E.; Sachs, M.B. Recovery from sound exposure in auditory-nerve fibers. J. Acoust. Soc. Am. 1973, 54, 1535–1543. [Google Scholar] [CrossRef]
  123. Huet, A.; Batrel, C.; Dubernard, X.; Kleiber, J.C.; Desmadryl, G.; Venail, F.; Liberman, M.C.; Nouvian, R.; Puel, J.L.; Bourien, J. Peristimulus Time Responses Predict Adaptation and Spontaneous Firing of Auditory-Nerve Fibers: From Rodents Data to Humans. J. Neurosci. 2022, 42, 2253–2267. [Google Scholar] [CrossRef]
  124. Bourien, J.; Tang, Y.; Batrel, C.; Huet, A.; Lenoir, M.; Ladrech, S.; Desmadryl, G.; Nouvian, R.; Puel, J.L.; Wang, J. Contribution of auditory nerve fibers to compound action potential of the auditory nerve. J. Neurophysiol. 2014, 112, 1025–1039. [Google Scholar] [CrossRef] [PubMed]
  125. Zelle, D.; Bader, K.; Dierkes, L.; Gummer, A.W.; Dalhoff, E. Derivation of input-output functions from distortion-product otoacoustic emission level maps. J. Acoust. Soc. Am. 2020, 147, 3169–3187. [Google Scholar] [CrossRef] [PubMed]
  126. Bader, K.; Dierkes, L.; Braun, L.H.; Gummer, A.W.; Dalhoff, E.; Zelle, D. Test-retest reliability of distortion-product thresholds compared to behavioral auditory thresholds. Hear. Res. 2021, 406, 108232. [Google Scholar] [CrossRef] [PubMed]
  127. Whitehead, M.L.; Stagner, B.B.; Martin, G.K.; Lonsbury-Martin, B.L. Visualization of the onset of distortion-product otoacoustic emissions, and measurement of their latency. J. Acoust. Soc. Am. 1996, 100, 1663–1679. [Google Scholar] [CrossRef]
  128. Krokenberger, M. Adaptive DPOAE-Wachstumsfunktionen zur Objektiven Hörschwellenschätzung bei Normalhörenden und Hörgeschädigten Ohren; Universität Tübingen: Tübingen, Germany, 2019. [Google Scholar]
  129. Zelle, D.; Gummer, A.W.; Dalhoff, E. Extraction of otoacoustic distortion product sources using pulse basis functions. J. Acoust. Soc. Am. 2013, 134, EL64–EL69. [Google Scholar] [CrossRef]
Figure 1. Elevated hearing thresholds correlated with age, in particular at high frequencies: (A) Individual (grey lines) and mean (black line) PTTs for the three age groups young (left), middle-aged (center), and older (right) used for PTAs of four different frequency ranges: low frequencies “PTA-LF” (0.125–1 kHz), main-language region “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” (6–10 kHz), and extended high frequencies “PTA-EHF” (11.2–16 kHz), illustrated in red on the abscissa of the left-most audiogram. The group mean thresholds are plotted in black (young: circles; middle-aged: triangles; older: squares). (B) Scatterplots for individual hearing thresholds as a function of age, split into the four PTA frequency ranges. The shaded area delineates the age range of the middle-aged group. p-Values (Pearson’s correlation): p(PTA-LF) = 0.000016; p(PTA4) < 0.00001; p(PTA-HF) < 0.00001, p(PTA-EHF) < 0.00001.
Figure 1. Elevated hearing thresholds correlated with age, in particular at high frequencies: (A) Individual (grey lines) and mean (black line) PTTs for the three age groups young (left), middle-aged (center), and older (right) used for PTAs of four different frequency ranges: low frequencies “PTA-LF” (0.125–1 kHz), main-language region “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” (6–10 kHz), and extended high frequencies “PTA-EHF” (11.2–16 kHz), illustrated in red on the abscissa of the left-most audiogram. The group mean thresholds are plotted in black (young: circles; middle-aged: triangles; older: squares). (B) Scatterplots for individual hearing thresholds as a function of age, split into the four PTA frequency ranges. The shaded area delineates the age range of the middle-aged group. p-Values (Pearson’s correlation): p(PTA-LF) = 0.000016; p(PTA4) < 0.00001; p(PTA-HF) < 0.00001, p(PTA-EHF) < 0.00001.
Jcm 13 02725 g001
Figure 2. (A) Power spectrum of the OLSA speech material (broadband, black curve), of high-pass filtered speech (OLSA-HP, red shaded area), and low-pass filtered speech (OLSA-LP, blue shaded area), shown peak-normalized to 0 dB and 1/f-corrected. The four different PTA frequency ranges are depicted: low frequencies “PTA-LF” (0.125–1 kHz), “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” [6–10 kHz], and extended high frequencies “PTA-EHF” (11.2–16 kHz). (B,C) The influence of noise on OLSA SRT50 was examined using differently filtered speech material. Noise conditions in (B) quiet and (C) ipsilateral noise. Columns provide results for broadband and filtered OLSA stimuli. OLSA SRTs are provided as a function of age.
Figure 2. (A) Power spectrum of the OLSA speech material (broadband, black curve), of high-pass filtered speech (OLSA-HP, red shaded area), and low-pass filtered speech (OLSA-LP, blue shaded area), shown peak-normalized to 0 dB and 1/f-corrected. The four different PTA frequency ranges are depicted: low frequencies “PTA-LF” (0.125–1 kHz), “PTA4” (0.5–4 kHz), high frequencies “PTA-HF” [6–10 kHz], and extended high frequencies “PTA-EHF” (11.2–16 kHz). (B,C) The influence of noise on OLSA SRT50 was examined using differently filtered speech material. Noise conditions in (B) quiet and (C) ipsilateral noise. Columns provide results for broadband and filtered OLSA stimuli. OLSA SRTs are provided as a function of age.
Jcm 13 02725 g002
Figure 3. ABR as a function of age, pure-tone averages, and speech comprehension: (A) ABR wave amplitude and latencies grouped by age. Circles represent young, triangles middle-aged, and squares older participants. (BD) ABR wave amplitudes and latencies grouped for participants with low (green) and high (pink) thresholds of PTA4 (B), PTA-HF (C), and PTA-EHF (D).
Figure 3. ABR as a function of age, pure-tone averages, and speech comprehension: (A) ABR wave amplitude and latencies grouped by age. Circles represent young, triangles middle-aged, and squares older participants. (BD) ABR wave amplitudes and latencies grouped for participants with low (green) and high (pink) thresholds of PTA4 (B), PTA-HF (C), and PTA-EHF (D).
Jcm 13 02725 g003
Figure 4. OLSA speech reception threshold SRT50 (dB SPL; y-axes) for differently filtered OLSA stimuli ((A,E) broadband, (B,F) low pass, and (C,D,G,H) high-pass) as a function of PTA4 (A,E), PTA-LF (B,F), PTA-HF(C,G), and PTA-EHF (D,H) (x-axes). (AD) provide results obtained in quiet (n = 89), (EH) under ipsilateral (n = 63) noise condition. Regression lines are plotted in black and include y-intersections and R2 values. The different colors assign each subject to one of the three speech comprehension groups: good (blue), standard (grey), and poor (orange).
Figure 4. OLSA speech reception threshold SRT50 (dB SPL; y-axes) for differently filtered OLSA stimuli ((A,E) broadband, (B,F) low pass, and (C,D,G,H) high-pass) as a function of PTA4 (A,E), PTA-LF (B,F), PTA-HF(C,G), and PTA-EHF (D,H) (x-axes). (AD) provide results obtained in quiet (n = 89), (EH) under ipsilateral (n = 63) noise condition. Regression lines are plotted in black and include y-intersections and R2 values. The different colors assign each subject to one of the three speech comprehension groups: good (blue), standard (grey), and poor (orange).
Jcm 13 02725 g004
Figure 5. Subjective hearing evaluation by age and speech comprehension: (A) shows age groups and (B) groups according to objective speech comprehension performance based on OLSA thresholds corrected by PNOTs. y-axis: subjective evaluation, x-axis: percentage of all responses given by all participants in age groups (A) and in PNOT groups (B). Participants were asked to rate their hearing as excellent, very good, good, moderate, or bad (y-axis labels).
Figure 5. Subjective hearing evaluation by age and speech comprehension: (A) shows age groups and (B) groups according to objective speech comprehension performance based on OLSA thresholds corrected by PNOTs. y-axis: subjective evaluation, x-axis: percentage of all responses given by all participants in age groups (A) and in PNOT groups (B). Participants were asked to rate their hearing as excellent, very good, good, moderate, or bad (y-axis labels).
Jcm 13 02725 g005
Figure 6. (A) ASSR response amplitudes in µV averaged for 4 and 6 kHz carriers as a function of participant age in years. The blue, grey, and orange-colored symbols refer to the good, standard, and poor speech comprehension groups, respectively. (B) Median (horizontal bar) and individual participants (symbols) ASSR amplitude averaged for 4 and 6 kHz carriers (Mean, left), 4 kHz carrier (middle), and 6 kHz carrier (right) for the quiet listening condition (upper row), or in ipsilateral noise (lower row). Numbers in brackets indicate the number of participants included in the analyses. (C) Regression line (black) of the dependence of OLSA SRT50 in ipsilateral noise on ASSR amplitudes (averaged for 4 and 6 kHz carriers) normalized for PTT. The y-intersection, R2 value, and p-value of regression are given close to the trend line.
Figure 6. (A) ASSR response amplitudes in µV averaged for 4 and 6 kHz carriers as a function of participant age in years. The blue, grey, and orange-colored symbols refer to the good, standard, and poor speech comprehension groups, respectively. (B) Median (horizontal bar) and individual participants (symbols) ASSR amplitude averaged for 4 and 6 kHz carriers (Mean, left), 4 kHz carrier (middle), and 6 kHz carrier (right) for the quiet listening condition (upper row), or in ipsilateral noise (lower row). Numbers in brackets indicate the number of participants included in the analyses. (C) Regression line (black) of the dependence of OLSA SRT50 in ipsilateral noise on ASSR amplitudes (averaged for 4 and 6 kHz carriers) normalized for PTT. The y-intersection, R2 value, and p-value of regression are given close to the trend line.
Jcm 13 02725 g006
Figure 7. (A,E) LEDPT acceptance rates, (B,F) PTT, (C,G) LEDPT, (D,H) LEDPT-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-quiet comprehension performers. Participants with good speech-in-quiet performance (blue) showed higher acceptance rates (A), equal PTT (B), inconclusive LEDPT (C), but a consistent 3 dB better threshold for LEDPT-to-PTT although on the right ear with only p = 0.084 (D). Estimated distortion-product thresholds (LEDPT) in relation to PTT, when participants are grouped with respect to their speech-in-ipsilateral-noise performance (E) LEDPT acceptance rates, (F) PTT, (G) LEDPT, (H) LEDPT-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-noise comprehension performers. Participants with good speech-in-noise performance (blue) show reduced acceptance rates, reduced PTT and LEDPT, but no difference for LEDPT-to-PTT.
Figure 7. (A,E) LEDPT acceptance rates, (B,F) PTT, (C,G) LEDPT, (D,H) LEDPT-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-quiet comprehension performers. Participants with good speech-in-quiet performance (blue) showed higher acceptance rates (A), equal PTT (B), inconclusive LEDPT (C), but a consistent 3 dB better threshold for LEDPT-to-PTT although on the right ear with only p = 0.084 (D). Estimated distortion-product thresholds (LEDPT) in relation to PTT, when participants are grouped with respect to their speech-in-ipsilateral-noise performance (E) LEDPT acceptance rates, (F) PTT, (G) LEDPT, (H) LEDPT-to-PTT difference for left and right ears are compared between good (blue) and poor (orange) speech-in-noise comprehension performers. Participants with good speech-in-noise performance (blue) show reduced acceptance rates, reduced PTT and LEDPT, but no difference for LEDPT-to-PTT.
Jcm 13 02725 g007
Figure 8. ABR wave amplitude as a function of ABR wave latency in participants matched for PTA thresholds and grouped for good (blue), standard (grey), or poor (orange) speech comprehension in quiet. Significant shifts in latency in poor comprehension, in comparison to the group with good speech comprehension were observed (ABR wave I latency: n = 29, 27, 24, p = 0.218242; wave II latency: n = 24, 22, 16, p = 0.007707, wave III latency: n = 30, 28, 26, p = 0.182784; wave V latency: n = 30, 28, 28, p = 0.026617 and wave VI latency: n = 27, 27, 24, p = 0.001055).
Figure 8. ABR wave amplitude as a function of ABR wave latency in participants matched for PTA thresholds and grouped for good (blue), standard (grey), or poor (orange) speech comprehension in quiet. Significant shifts in latency in poor comprehension, in comparison to the group with good speech comprehension were observed (ABR wave I latency: n = 29, 27, 24, p = 0.218242; wave II latency: n = 24, 22, 16, p = 0.007707, wave III latency: n = 30, 28, 26, p = 0.182784; wave V latency: n = 30, 28, 28, p = 0.026617 and wave VI latency: n = 27, 27, 24, p = 0.001055).
Jcm 13 02725 g008
Figure 9. Syllable-discrimination scores in relation to speech comprehension. The scores for four pairs of phonemes (/o/-/u/, /i/-/y/, /du/-/bu/, /di/-/bi/) are segregated for participants with poor (orange), good (blue), and standard (grey) speech comprehension selected by PNOT in quiet (A), and ipsilateral noise (B). Each plot consists of a boxplot with perceptual performance [% correct] as a function of PNOT (x-axis). Finally, there is a graphical representation of the significance assessed by Mann–Whitney U tests (Supplementary Table S4), significant differences are shown as asterisks with a color code reflecting the three groups.
Figure 9. Syllable-discrimination scores in relation to speech comprehension. The scores for four pairs of phonemes (/o/-/u/, /i/-/y/, /du/-/bu/, /di/-/bi/) are segregated for participants with poor (orange), good (blue), and standard (grey) speech comprehension selected by PNOT in quiet (A), and ipsilateral noise (B). Each plot consists of a boxplot with perceptual performance [% correct] as a function of PNOT (x-axis). Finally, there is a graphical representation of the significance assessed by Mann–Whitney U tests (Supplementary Table S4), significant differences are shown as asterisks with a color code reflecting the three groups.
Jcm 13 02725 g009
Figure 10. Good and poor speech comprehension in quiet differs from good and poor speech comprehension in ipsilateral noise in the discrimination ability of formant contrasts below PLL (requiring TFS coding), and above PLL (requiring TENV coding). In quiet, poor speech comprehension is associated with poor discrimination below the PLL (e.g., for /o/-/u/), while good speech comprehension is associated with good discrimination above the PLL (e.g., for /i/-/y/). In ipsilateral noise, poor speech comprehension is associated with lower performance for discriminating phoneme pairs with formant contrasts above PLL (/i/-/y/, above 1500 Hz), while good speech comprehension is associated with good discrimination of formants below the PLL (/o/-/u/, below 1500 Hz).
Figure 10. Good and poor speech comprehension in quiet differs from good and poor speech comprehension in ipsilateral noise in the discrimination ability of formant contrasts below PLL (requiring TFS coding), and above PLL (requiring TENV coding). In quiet, poor speech comprehension is associated with poor discrimination below the PLL (e.g., for /o/-/u/), while good speech comprehension is associated with good discrimination above the PLL (e.g., for /i/-/y/). In ipsilateral noise, poor speech comprehension is associated with lower performance for discriminating phoneme pairs with formant contrasts above PLL (/i/-/y/, above 1500 Hz), while good speech comprehension is associated with good discrimination of formants below the PLL (/o/-/u/, below 1500 Hz).
Jcm 13 02725 g010
Table 1. Pure-tone threshold-normalized SRT50 (PNOT) differentiated for the noise condition and the three speech comprehension groups.
Table 1. Pure-tone threshold-normalized SRT50 (PNOT) differentiated for the noise condition and the three speech comprehension groups.
Noise Conditions PNOTnAge Mean ± SEMPTA4 Mean ± SEMPTA-EHF Mean ± SEM
In quietgood3045.40 ± 2.8712.33 ± 1.2136.22 ± 3.24
standard2938.48 ± 3.1013.41 ± 1.4428.67 ± 3.73
poor3047.60 ± 3.2513.40 ± 1.1838.28 ± 3.73
In ipsilateral noisegood2145.24 ± 3.6414.02 ± 1.4936.00 ± 4.19
standard2141.05 ± 3.8713.21 ± 1.5329.42 ± 4.67
poor2142.10 ± 3.7012.38 ± 1.7834.30 ± 4.25
Table 2. Dependence of all OLSA SRT50 on their corresponding pure-tone averages. OLSA-BB on PTA4 (Figure 4A,D,G).
Table 2. Dependence of all OLSA SRT50 on their corresponding pure-tone averages. OLSA-BB on PTA4 (Figure 4A,D,G).
R2pn
OLSA quietBB over PTA40.5363<0.0000189
LP over PTA-LF0.4574<0.0000189
HP over PTA-LF0.03880.06439489
HP over PTA-EHF0.378<0.0000189
OLSA ipsilateral noiseBB over PTA40.3111<0.0000163
LP over PTA-LF0.0640.04552563
HP over PTA-HF0.18260.00047863
HP over PTA-EHF0.17480.0006563
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Schirmer, J.; Wolpert, S.; Dapper, K.; Rühle, M.; Wertz, J.; Wouters, M.; Eldh, T.; Bader, K.; Singer, W.; Gaudrain, E.; et al. Neural Adaptation at Stimulus Onset and Speed of Neural Processing as Critical Contributors to Speech Comprehension Independent of Hearing Threshold or Age. J. Clin. Med. 2024, 13, 2725. https://doi.org/10.3390/jcm13092725

AMA Style

Schirmer J, Wolpert S, Dapper K, Rühle M, Wertz J, Wouters M, Eldh T, Bader K, Singer W, Gaudrain E, et al. Neural Adaptation at Stimulus Onset and Speed of Neural Processing as Critical Contributors to Speech Comprehension Independent of Hearing Threshold or Age. Journal of Clinical Medicine. 2024; 13(9):2725. https://doi.org/10.3390/jcm13092725

Chicago/Turabian Style

Schirmer, Jakob, Stephan Wolpert, Konrad Dapper, Moritz Rühle, Jakob Wertz, Marjoleen Wouters, Therese Eldh, Katharina Bader, Wibke Singer, Etienne Gaudrain, and et al. 2024. "Neural Adaptation at Stimulus Onset and Speed of Neural Processing as Critical Contributors to Speech Comprehension Independent of Hearing Threshold or Age" Journal of Clinical Medicine 13, no. 9: 2725. https://doi.org/10.3390/jcm13092725

APA Style

Schirmer, J., Wolpert, S., Dapper, K., Rühle, M., Wertz, J., Wouters, M., Eldh, T., Bader, K., Singer, W., Gaudrain, E., Başkent, D., Verhulst, S., Braun, C., Rüttiger, L., Munk, M. H. J., Dalhoff, E., & Knipper, M. (2024). Neural Adaptation at Stimulus Onset and Speed of Neural Processing as Critical Contributors to Speech Comprehension Independent of Hearing Threshold or Age. Journal of Clinical Medicine, 13(9), 2725. https://doi.org/10.3390/jcm13092725

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop