Speech-Evoked Auditory Brainstem Response in Children with Sensorineural Hearing Loss
Bahram Jalaeia1* and Mohd Normani Zakariab2
1Department of Audiology, Arak School of Rehabilitation, Arak University of Medical Sciences, Markazi, Iran
2Audiology and Speech Pathology Programme, School of Health Sciences, University Sains Malaysia, 16150 Kubang Kerian, Kelantan, Malaysia
Submission: July 08, 2019; Published: July 22, 2019
*Corresponding author: Bahram Jalaei, Chairman of Audiology Department., School of Rehabilitation, Arak University of Medical Sciences, Markazi, Iran
How to cite this article: Bahram J, Mohd Normani Z. Speech-Evoked Auditory Brainstem Response in Children with Sensorineural Hearing Loss. Glob J Oto, 2019; 20(4): 556042. DOI: 10.19080/GJO.2019.20.556042
Abstract
Speech-evoked auditory brainstem response (speech-ABR) provides an objective documentation of temporal and spectral encoding of speech cues in the brainstem region. As such, the influence of peripheral hearing loss on speech-ABR has been studied but the current literature is limited. The aim of the present study was to compare speech-ABR findings between children with moderate-to-severe SNHL and their age-matched normal peers. In this case control study, 17 hearing-impaired (HI) children (mean age = 7.1 ± 1.8 years) and 15 typically developing (TD) children (mean age = 6.1 ± 1.6 years) were enrolled. All HI children had bilateral moderately severe sensorineural hearing loss. The speech-ABR was recorded from HI and TD children to 40-ms speech syllable /da/ presented repetitively at 30 dB sensation level (SL). Statistically longer latencies were found in HI children than in TD children for onset peaks (V and A) and transition portion (peak C) (p < 0.05) with very large effect sizes (d = 1.16 - 6.10). In contrast, the sustained components (peaks D, E and F) and offset portion (peak O) of speech-ABR were unaffected by hearing loss (p > 0.05). For speech-ABR composite onset measures, TD children revealed larger VA inter-peak amplitudes (d = 0.87) and steeper VA slopes (d = 1.36) compared to HI children (p < 0.05). Children with moderately severe sensorineural hearing loss had abnormal encoding of speech sounds within the brainstem region. Based on the study outcomes, the usefulness of speech-ABR can be expanded by studying the speech encoding ability in multi-handicapped children).
Keywords: Speech-ABR; Abnormal Encoding; Onset; Sustained; Sensorineural Hearing Loss
Abbrevations: HI: Hearing-Impaired; TD: Typically Developing; ABR: Auditory Brainstem Response; AMLR: Auditory Middle Latency Response; CABR: Cortical Auditory Evoked Potential; OAEs: Otoacoustic Emissions; SD: Standard Deviation; EPSPs: Excitatory Postsynaptic Potentials; EEG: Electroencephalography
Introduction
In everyday life, people use various speech cues in order to understand ongoing conversations. The acoustical features of speech such as onset time, phoneme transition, periodicity, formant and envelope are encoded and analyzed accordingly by the complex hearing system Skoe & Kraus [1]. On the other side, previous researches have suggested the notable effects of sensorineural hearing loss (SNHL) on the affected individuals include decreased audibility, poor frequency discrimination, reduced temporal resolution and integration Moore [2]. These negative consequences of SNHL are in fact more prominent in children who are in the process of acquiring speech and language skills Northern & Downs [3]. Electrophysiological tests offer an objective indication as to how auditory stimuli are processed and encoded at the subcortical and cortical levels Hall [4], Koravand [5], Sharma [6], Skoe & Kraus [1], Studer-Eichenberger [7]. Auditory brainstem response (ABR), auditory middle latency response (AMLR) and cortical auditory evoked potential (CAEP) are among electrophysiological tests that have been studied in children and adults with auditory dysfunctions Hall [4], Nash Kille & Sharma [8], Skoe & Kraus [1], Yang [9]
For many decades, ABR has been used as a standard electrophysiological test for various clinical applications in the fields of audiology and neuro-otology Anwar et al. [10], Hall [1], Koohi [11]. Depending on its specific purposes, ABR is typically recorded with stimuli such as clicks, tone bursts or chirps Birkent et al. [12], Cobb & Stuart [13], Hall [4], Prendergast et al. [14]. More recently, ABR evoked by complex stimuli has been a subject of increased interest among researchers Koravand [15], Skoe & Kraus [1]. As such, ABR elicited by complex sounds such as speech syllables can provide supplementary data to the conventional click-evoked ABR and other clinical tests in assessing the neural function Dhar [16], Skoe & Kraus [1]. When stimulated by a repetitive 40-ms speech syllable /da/, the speech-evoked auditory brainstem response (speech-ABR) waveform typically consists of seven peaks (Figure 1). As shown, peaks V and A reflect the onset (transient) property of speech- ABR, that is analogous to wave V of click-evoked ABR Song [17]. These peaks are highly dependent on the neural synchronization upon auditory stimulations Skoe & Kraus [1]. The consonant-tovowel transition segment of the syllable /da/ is represented by peak C. Peaks D, E, and F represent the phase-locked sustained region (i.e., FFR), while peak O reflects the offset portion of steady-state vowel Skoe & Kraus [1].
Speech-ABR can be recorded reliably with monaural and binaural stimulations Ahadi [18], Uppunda [19], as well as in the presence of background noise Chandrasekaran [20], Cunningham [21], Hornickel [22]. The responsiveness of speech-ABR in documenting neural plasticity following specific training procedures is also encouraging Jalaei, Filippini [23], Krishnamurti [24], Russo [25], Song [26]. Deficits in the neural encoding of speech cues have been revealed by speech- ABR in individuals with learning problems Cunningham [21], King [27], Sanfins [28], Wible [29], phonological disorders Gonçalves [30], auditory processing disorders Krishnamurti [24], attention deficit disorders Jafari [31], autism Russo [32], epilepsy Elkabariti [33] and stuttering problems Tahaei [34]. The results of numerous studies on speech-ABR have shown that this electrophysiological response is indeed useful, sensitive and reliable in providing objective documentation of temporal and spectral encoding of complex stimuli Skoe & Kraus [1]. It can be effectively elicited by stimuli that are transient or sustained, abrupt or slow and continuous or interrupted Skoe & Kraus [1]. To produce robust speech-ABR waveforms, peripheral hearing system, auditory nerve and brainstem nuclei must be intact and functioning. As such, the influence of peripheral hearing loss on the brainstem processing of speech and complex sounds has been studied but the current literature is limited and contentious. In a study by Koravand [15], speech-ABR results were compared between normally hearing children (n = 13, aged 6 to 14 years) and children with bilateral mild-to-moderately severe SNHL (n = 12, aged 6 to 14 years).
The typical speech syllable /da/ was employed to record the speech-ABR waveforms. While the onset and phoneme transition peaks were unaffected by hearing loss, significantly longer latencies of FFR peaks (D and E) were found in the hearing-impaired group. The authors then concluded that there were specific patterns of subcortical encoding of speech sounds in children with mild-to-moderately severe SNHL. In line with this, Anderson [35] recorded speech-ABR to speech syllable /da/ from normally hearing older adults (n = 15, aged 61 to 68 years) and older adults with mild-to-moderate SNHL (n = 15, aged 60 to 71 years). They then found that the neural representation of envelope-to-fine structure of the sustained segment of speech sounds was imbalanced in the hearing-impaired group. No such result was seen in the control group. On the other hand, contradictory results were reported by other studies Leite [36], Nada [37]. In a study by Nada [37], speech-ABR (elicited by syllables (/da/ and /ba/) was recorded from 40 adults with bilateral mild-to-moderate SNHL and 20 adults with normal hearing. For both types of speech syllables, statistically longer latencies of speech-ABR onset peaks (V and A) were found in the hearing-impaired group than in the normal group. When evoked by speech syllable /da/, the hearing-impaired adults also revealed longer peak C latencies than the normal group.
The sustained FFR features, on the other hand, were not affected by the peripheral hearing loss. The insignificant influence of peripheral hearing loss on FFR peaks was also reported by Leite [36] who studied speech-ABR in children with bilateral mild-to-moderate SNHL. Methodological differences (e.g., different age groups, different severity of hearing loss etc.) are the likely reasons for discrepancies between the results, in which further confirmation is required. Taken together, more research efforts are needed to enhance the current understanding on the influence of hearing loss on the brainstem encoding of speech sounds. As such, speech-ABR outcomes in children with more severe hearing loss have not been reported. The present study, thus, was carried out to compare speech-ABR findings between typically developing children and children with moderately severe SNHL. Due to the notable hearing loss, it was hypothesized that the neural encoding of transient, phoneme transition and sustained segments of speech sounds would be degraded in children with SNHL relative to their age-matched normal peers.
Materials and Methods
Participants
Thirty-two Malaysian children aged from 4 to 9 years participated in this case-control study. Of this, 17 of them were hearing-impaired children (mean age = 7.1 ± 1.8 years, 8 girls and 9 boys) and the remaining 15 were typically developing children (mean age = 6.1 ± 1.6 years, 8 girls and 7 boys). All typically developing (TD) children had no hearing, communication or academic problems and performed normally in pure tone audiometry (PTA thresholds of ≤ 20 dB HL at 250- 8000 Hz octave frequencies) and tympanometric testing (i.e.,healthy middle ear function) bilaterally. They were recruited among children of staff members of University Hospital. The hearing-impaired (HI) children, on the other hand, were selected among patients of Audiology Clinic, University Hospital using the convenience sampling method. All of them had hearing levels within a moderately severe range (sensorineural type) with downward sloping patterns bilaterally. They are all hearing aid wearers and had no other medical disabilities (as reported in their medical files). They also had no otoacoustic emissions (OAEs) in both ears, suggestive of poor functions of outer hair cells. All procedures performed in this study involving human participants were in accordance with the ethical standards of Universiti Sains Malaysia (USM/PPP/JEPeM [245.3(5)]) and with the 1975 Helsinki declaration and its later amendments. Prior to the data collection, informed consent was obtained from the parents of children included in the study.
Stimuli and Recording Parameters
The speech-ABR was recorded according to the standard protocol with a two-channel auditory evoked potential device (Natus Medical Inc., Mundelein, USA). The electrodes were positioned accordingly on each child’s head: non-inverting on the vertex, inverting on the right mastoid and ground on the forehead. The impedance of all electrodes was kept below 3 KΩ and within 1 KΩ of each other. The standard 40-ms syllable / da/ was used as the stimulus to record speech-ABR (Figure 1). It begins with an initial onset burst and followed by a CV transition. The fundamental frequency (F0) of the stimulus rises linearly from 103 to 125 Hz while the formants shift as follows: F1: 220 → 720 Hz; F2: 1700 → 1240 Hz; F3: 2580 → 2500 Hz. The fourth (3600 Hz) and fifth (4500 Hz) formants remain constant for the duration of the stimulus. The stimulus was presented to each child’s right ear at 30 dB SL (i.e., 30 dB above the averaged hearing level at 500, 1000, 2000 and 4000 Hz) using the headphones. The stimulus rate was 10.3/s with 3000 sweeps for each trial. The time window was set at 74.67 ms (including a 10-ms pre-stimulus period). The acquired responses were amplified 100,000 times and band-pass filtered at 100-2000 Hz. For eliminating the unwanted responses, the artifact rejection was set at ± 23.8 μV. To ensure good waveform replicability, the recording was repeated twice for each trial. During the speech- ABR recording, the children laid comfortably in the supine position on the provided bed while watching voiceless cartoons.
Statistical Analyses
Two experienced audiologists were involved in identifying and verifying speech-ABR peaks. For each child, amplitude and latency values of speech-ABR peaks (V, A, C, D, E, F and O) were computed. Speech-ABR composite onset measures (VA inter-peak duration, VA inter-peak amplitude and VA slope) were also recorded. Since all data were found to be normally distributed (p > 0.05 by Kolmogorov-Smirnov normality test), parametric inferential statistical analyses were carried out. To control for the possible effect of age and gender on speech-ABR data, analysis of covariance (ANCOVA, with age and gender as covariates) was conducted to compare the speech-ABR findings between TD and HI groups. Prior to this, the data were initially tested for the homogeneity of slopes to ensure the ANCOVA assumption had not been violated. Pearson analysis was used to determine the correlation between hearing level and speech- ABR results. The p values of less than 0.05 were considered statistically significant. In addition, to determine the magnitude of difference between the two groups, Cohen‘s effect size (d) was computed and interpreted as small (d = 0.20), medium (d = 0.50), large (d = 0.80) or very large (d ≥ 1.00) Kraemer [38]. All data analyses were conducted with SPSS software version 20 (SPSS Inc, Chicago, IL, USA).
Results
Age and Hearing Profiles
In the present study, even though the mean age was descriptively lower in the TD group (6.1 ± 1.6 years) than in the HI group (7.1 ± 1.8 years), this age difference was not significant (t = -1.766, df = 30, p = 0.088). For the HI group, the hearing thresholds ranged from 30 to 85 dB HL between 500, 1000, 2000 and 4000 Hz frequencies in both ears. The averaged hearing levels (at 500, 1000, 2000 and 4000 Hz frequencies) ranged from 51.3 to 71.3 dB (mean = 62.4 ± 5.7 dB) and from 46.3 to 72.5 dB (mean = 62.9 ± 7.2 dB) for the right and left ears, respectively. These averaged hearing levels were found to be comparable between the ears (t = -0.372, df = 16, p = 0.717).
Latency Analyses
Reliable speech-ABR waveforms had been successfully obtained from all children of both groups. The representative speech-ABR waveforms of a TD child and a HI child are depicted in Figure 2. Mean, standard deviation (SD) and inferential statistical analyses for speech-ABR results for both groups are shown in Table 1. As revealed, the speech-ABR latencies were descriptively longer in the HI group than in the TD group, particularly for onset peaks. This observation was then confirmed by ANCOVA results, in which statistically longer latencies were found in HI children for peaks V, A and C (p < 0.05). For these peaks, effect sizes were very large indicating that substantive 004 Global Journal of Otolaryngology How to cite this article: Bahram J, Mohd Normani Z. Speech-Evoked Auditory Brainstem Response in Children with Sensorineural Hearing Loss. Glob J Oto, 2019; 20(4): 556042. DOI: 10.19080/GJO.2019.20.556042 differences were indeed noted between the two groups (d = 1.16 - 6.10). On the other hand, the sustained FFR components (peaks D, E and F) and offset portion (peak O) of speech-ABR were not affected by hearing loss as comparable latency findings were noted between the two groups (p > 0.05). To further support the ANCOVA results, the correlations between hearing level and speech-ABR results were also reported. As shown in Table 2, strong positive correlations were noted between hearing level and latency of peak V (r = 0.90, p < 0.001) and peak A (r = 0.93, p < 0.001). A moderate but significant correlation was found between hearing level and peak C latency (r = 0.54, p = 0.001). In this regard, latency of these peaks increased with increasing the hearing level. No significant correlations were found between hearing level and latency of sustained peaks of speech-ABR (p > 0.05).
Amplitude Analyses
For peak amplitude, the significant statistical result was only found for peak A (Table 1). That is, the mean amplitude of peak A was significantly larger in the TD group (-0.19 ± 0.07 μV) than in the HI group (-0.07 ± 0.05 μV) (F = 28.374, df = 1, 28, p < 0.001). This amplitude difference was further supported by a very large effect size (d = 1.99). As revealed in Table 2, a strong correlation was also found between hearing level and peak A amplitude (r = 0.71, p < 0.001). No such statistical outcomes were seen for other speech-ABR peaks.
Composite Onset Measures
While no significant result was found for VA inter-peak duration (F = 0.131, df = 1, 28, p = 0.720), TD children revealed statistically larger VA inter-peak amplitudes and steeper VA slopes compared to HI children (p < 0.05). As shown in Table 1, the effect size was large for VA slope (d = 0.87) and very large for VA inter-peak amplitude (d = 1.36). Other than the VA interpeak duration (r = 0.01, p = 0.986), significant correlations were found between hearing level and composite onset measures of speech-ABR (p < 0.05) (Table 2).
Discussion
Recall that the present study aimed to compare the speech- ABR findings between typically developing children and children with moderately severe hearing loss. Since OAEs were absent in all tested HI children (suggestive of cochlear damage), the type of hearing loss was likely to be sensory loss. The comparisons between the two groups were considered appropriate as the age and gender distributions were comparable. Furthermore, age and gender influences (if any) were controlled with the use of ANCOVA when speech-ABR results were compared between TD and HI groups. Overall, based on the statistical results, a pronounced influence of hearing loss was observed on some speech-ABRpeaks. Different from the findings of previous studies, the peak latencies of TD children in the present study were consistently longer. This is not surprising as low stimulation levels (i.e., 30 dB SL) were used to record speech-ABR. Most of studies on speech-ABR among normally hearing participants utilized high intensity levels (≥ 80 dB SPL), which resulted in waveforms with earlier peak latencies Koravand [15], Leite [36], Sanfins [39]. This latency pattern is common and has been consistently demonstrated in many auditory electrophysiological tests Hall [4], Møller [40], Picton [41]. As reported elsewhere, at high intensity levels, the neural responses are mostly contributed by the basal part of cochlea Picton [41]. At lower intensity levels, the site of excitation moves toward the apical region of cochlear, resulting in ABR with longer latencies Picton [41]. In addition, the earlier ABR latencies at high stimulation levels may also be contributed by the temporal aspect of excitatory postsynaptic potentials (EPSPs). That is, the time taken for EPSPs to reach thresholds is faster at high intensity levels (thus, shorter latencies) than at low intensities Møller [40].
Onset peak measures of speech-ABR
In the present study, the latencies of onset peaks (V and A) were significantly prolonged in the HI group with very large effect sizes. Moreover, the hearing levels were positively correlated with the onset peak latencies. This suggests that the temporal encoding of stop consonant in the brainstem region is notably degraded in children with SNHL. This abnormal speech encoding might also contribute to the poor ability of HI children in understanding conversations in their daily life. The delayed onset latencies in HI participants have also been reported in the previous studies on speech-ABR Archana [42], Nada [37] and click-evoked ABR Keith & Greville [43], Strelcyk [44]. In this regard, the use of transient broadband signals would excite all regions of the cochlea Hall [4]. In cases of downward sloping hearing loss (where high-frequency hearing loss is more prominent), the contribution to the ABR from neurons in the basal region would be reduced, whereas the contribution from more apical neurons would be increased, resulting in longer ABR latencies Hall [4], Gorga [45]. Prolonged latencies of speech- ABR onset peaks have also been observed in normally hearing children with learning disabilities King [27] and auditory processing disorders Krishnamurti [24], suggesting abnormal encoding of speech sounds at the brainstem level.
On the other hand, Koravand [15] found no significant differences in the onset peak latencies between normal and HI children. This finding is not consistent with the present study outcome, which is likely due to the methodological difference. Unlike the present study that recruited children with greater hearing loss, Koravand [15] recorded the speech-ABR from children with mild-to-moderately-severe hearing loss, which 006 Global Journal of Otolaryngology How to cite this article: Bahram J, Mohd Normani Z. Speech-Evoked Auditory Brainstem Response in Children with Sensorineural Hearing Loss. Glob J Oto, 2019; 20(4): 556042. DOI: 10.19080/GJO.2019.20.556042 resulted in shorter onset peak latencies. It is generally agreed that as the hearing level increases, the absolute latencies of ABR increases Hall [4], Hyde [46]. It is worth noting that the onset peak latencies of HI children found in Koravand et al.’s study were comparable to normal values. This is not surprising as at high intensity levels, absolute latencies of ABR tend to fall within the normal limits, particularly for audiograms with downward sloping patterns Galambos & Hecox [47], Gorga [45], Hall [4], Keith & Greville [43]. In the field of auditory electrophysiology, delayed peak latencies would be typically followed by reduced peak amplitudes when testing participants with specific auditory disorders Aihara [48], Hall [4], Lin [49], Pushpalatha & Konadath [50].
However, due to the fluctuations in background of electroencephalography (EEG) activity, the peak amplitude is less stable and more variable than the peak latency, particularly for low-amplitude early latency responses Chiappa [51], Edwards [52]. In the present study, reduced amplitude in the HI group was found for peak A. The peak amplitude of ABR represents the sum of action potentials of neurons that are synchronously stimulated by transient auditory stimuli. This neural excitation is degraded in the presence of cochlear pathology, particularly when the basal region of cochlea is damaged Steinhoff [53]. Therefore, significantly lower amplitudes of peak A in HI children found in the present study are consistent with the reduced neural synchronization because of the sensory hearing loss. The significant influence of peripheral hearing loss on the amplitude of speech-ABR onset peak was not revealed in other studies Koravand [15], Nada [36], which could be due to amplitude variability or participants’ characteristics (e.g., milder degree of hearing loss).
Composite Onset Measures of Speech-ABR
The physiological representation of the transient segment of syllable /da/ can be further enlightened by the composite onset measures of speech-ABR including VA inter-peak duration, VA inter-peak amplitude and VA slope. In the present study, significantly lower VA inter-peak amplitudes and less steep VA slopes were noted in HI children than in TD children. Reduced VA inter-peak amplitudes in HI participants were also observed in the study by Leite [36], which are again related to the decreased neural synchronization because of the cochlear damage. As highlighted by Wible [29], the VA slope can be a useful indicator for describing the neural encoding of the onset segment of speech as it captures both the duration and the amplitude of V-to-A transition. Relative to normal children, less steep VA slopes have been consistently noted in children with learning problems Wible [29], Wible [54]. Since HI children were found to have statistically lower VA inter-peak amplitudes, the less steep VA slopes found in this group were in fact sensible. The VA inter-peak duration, on the other hand, was not influenced by the peripheral hearing loss, which is in line with the previous study outcomes Nada [37].
Phoneme Transition Measure of Speech-ABR
In speech-ABR, the consonant-to-vowel transition segment is represented by peak C. Compared to other speech-ABR peaks, peak C has the smallest amplitude and may not be present in all healthy subjects Skoe & Kraus [1]. In the present study, nevertheless, the peak C was present in all tested children. While no significant difference in amplitude was found between the groups, HI children produced statistically longer peak C latencies than TD children. This implies that the encoding of phoneme transition cue is disrupted in children with SNHL, which is consistent with the previous speech-ABR outcome Nada [37]. In view of the aforementioned finding, it is also of interest to know whether there is a link between the deficient phoneme transition encoding (as reflected by the peak C of speech-ABR) and the temporal resolution (i.e., the ability of the auditory system to detect subtle changes or “gaps” between phonemes). Behavioral gap detection tasks are typically performed to measure the temporal resolution Moore [2].
Reduced temporal resolution skills (i.e., higher gap detection thresholds) have been reported in individuals with peripheral hearing loss Fitzgibbons & Gordon-Salant [55], Glasberg [56], Mishra [57], Moore [2] and auditory processing disorders Boscariol [58], Dias [59], which may underlie their poor speech perception. In a study by Sanju [60], gap detection thresholds and speech-ABR results were compared between young and middle-aged adults with normal hearing. As reported, significant differences in the test results were found between the two age groups. On the other hand, no significant correlations were observed between gap detection thresholds and speech-ABR results (peak V latency and FFR amplitude), suggesting that the auditory stimuli might be encoded differently by the two measures. The peak C of speech-ABR, however, was not analyzed and discussed in this study. Further research is therefore warranted to determine whether there is a significant relation between the peak C of speech-ABR and the gap detection ability, particularly when testing participants with compromised auditory functions.
Sustained FFR measures of speech-ABR
In the speech-ABR testing elicited by the consonant-vowel speech syllable, the sustained FFR peaks (D, E and F) represent the phase-locked neural encoding of the steady state vowel Skoe & Kraus [1]. While the influence of peripheral hearing loss on the onset peaks of speech-ABR was robustly noted in the present study, no significant differences in FFR results were found between HI and TD children. The insignificant effect of hearing loss on the sustained peaks of speech-ABR has also been reported in the previous studies Leite [36], Nada [37] and there are several possible reasons for this. First, intact peripheral and central auditory pathways contribute to the robustness of speech-ABR. The onset, severity and type of hearing loss may affect speech-ABR results. All HI children participated in the present study had better hearing thresholds at low frequencies How to cite this article: Bahram J, Mohd Normani Z. Speech-Evoked Auditory Brainstem Response in Children with Sensorineural Hearing Loss. Glob J Oto, 007 2019; 20(4): 556042. DOI: 10.19080/GJO.2019.20.556042 Global Journal of Otolaryngology (i.e., downward sloping audiometric patterns). Therefore, if compared to the high-frequency consonant perception, the vowel segment of speech syllable would be perceived more clearly and with better audibility, resulting in discernible FFR peaks.
Second, since the transient and sustained segments of the speech syllable might be encoded differently (by distinct neural mechanisms within the brainstem region) Akhoun [61], Hoormann [62], Kraus & Nicol [63], Russo [64], the influence of peripheral hearing loss on these speech elements would also be different. Russo [64] recorded speech-ABR to a repetitive 40-ms speech syllable /da/ from normally hearing children (n = 38, aged 8 to 12 years) in quiet and in noisy conditions. In the quiet environment, while significant correlations were found between the onset peaks, the sustained FFR measures were not related to each other and had no correlation with the onset peaks. Moreover, when the speech-ABR was recorded in the presence of noise, the onset responses were severely affected but the FFR features remained unaltered. When recording speech-ABR to speech syllable /ba/ in normally hearing adults, Akhoun [61] found that the onset and FFR features of speech-ABR were affected differently by the intensity level. Taken together, these findings suggest that there are two distinct neural streams (within the same pathway) for encoding the transient and sustained segments of speech sounds Kraus & Nicol [63]. In the study by Nada [37] that compared speech-ABR results between healthy and SNHL adults, while the latencies of onset peaks were statistically longer in the SNHL group, the FFR results were comparable between the two groups.
They then concluded that the neural phase locking to fundamental frequency and its harmonics was not influenced by the peripheral hearing loss, which is in line with the present study outcomes. Different from the present study outcomes and previous reports Leite [36], Nada [37], Koravand [15] found the latencies of FFR peaks (D and E) to be significantly delayed in hearing-impaired children than in the control group. Other speech-ABR peaks (including peak F of FFR), on the other hand, were not influenced by the hearing loss. Moreover, the amplitude of peak O was found to be consistently larger in the hearing-impaired group. The exact reasons for the discrepancies between the results are unknown and further large-scale research is warranted to provide more concrete evidence of these aspects of speech-ABR. In the present study, latencies and amplitudes of peak O were found to be comparable between TD and HI children. This implies that the neural encoding of the end of vowel is unaffected by hearing loss, which is line with the previous speech-ABR findings Leite [36], Nada [37].
Stimulation level issues
It is worth noting that the present study recorded speech- ABR at a similar low sensation level (i.e., 30 dB SL) as an effort to obtain an “equal” level of stimulation for both TD and HI groups. By considering the maximum output limit of the ABR device (around 100 dB nHL) and the maximum hearing levels of hearing-impaired participants (around 70 dB HL), the stimulation level was therefore set at 30 dB SL. Previous speech- ABR studies involving hearing-impaired participants used intensity levels between 80-95 dB sound pressure level (SPL) Abd El-Ghaffar [65], Anderson [35], Leite [36], Koravand [15]. Regarding the intensity level for speech-ABR recording, there is a concern whether the stimulus level should be presented in dB SPL or in dB SL when testing hearing-impaired individuals. When comparisons are made between normal and hearing-impaired listeners, the stimulus presentation at the same SL may result in higher absolute levels perceived by HI participants, leading to unequal (i.e., better) results Lewis [66]. In a study by Lewis [66] that recorded tone-evoked ABR in normal and mild-to-moderate hearing-impaired adults, significantly earlier wave V latencies were reported in the hearing-impaired group when the tonebursts were presented at similar SLs. In contrast, both groups produced similar wave V latencies with the use of equivalent SPLs in recording ABR.
The earlier ABR latencies in sensory hearing-impaired individuals when the tone bursts were presented at similar SLs can be explained by several factors including decreased cochlear filter build-up time, reduced cochlear amplifier gain and broadened auditory filters Lewis [66]. In the present study, the stimulus presentation at a similar SL in recording speech- ABR was considered appropriate to obtain the desired results. Different from Lewis et al.’s study (that used tone bursts to record ABR), the present study recorded speech-ABR using the speech syllable /da/ and thus, the findings would be different. In fact, since the HI children in the present study had greater hearing loss, the latencies of onset peaks were in fact delayed (not earlier) compared to that of the TD group [67-70].
Study Limitations
The present study is not without limitations. The size of sample was modest and perhaps more favorable outcomes would be obtained if more children could be recruited. Furthermore, the data were gathered from children of a limited age range (4 to 9 years) and different study findings might be obtained if different age groups are tested. In this respect, it is of interest to see how the brainstem encoding of speech cues is affected by hearing loss in younger children (≤ 3 years of age) and this is subject to further research [71-75].
Conclusion
In the present study, the brainstem encoding of speech syllable /da/ was compared between children with moderately severe SNHL and their age-matched normal peers by means of speech-ABR testing. As revealed, the neural encoding of steady-state vowel was unaffected by hearing loss. In contrast, the brainstem encoding of transient and phoneme transition segments of speech syllable was significantly disrupted in HI children. Nevertheless, since the differences in speech-ABRresults between the two groups can be very large (e.g., effect size of 6.10 for peak A latency), it remains puzzling whether the encoding deficits can also be due to compromised neural functions (conditions similar to learning-impaired children) rather than solely caused by peripheral hearing loss. Further research is therefore warranted to shed light on this issue. The present study findings may also provide insight into essential speech elements that require a “special” focus when developing an effective auditory training protocol for children with moderately severe hearing loss. Moreover, since the sustained FFR components are unaffected by hearing loss, the usefulness of speech-ABR can be further expanded by studying multi-handicapped children (those with peripheral hearing loss and auditory brainstem deficits). For example, it is of interest to know whether speech-ABR is useful as an objective tool in distinguishing between hearing-impaired children, and those with peripheral hearing loss and central auditory deficits (that compromise the encoding of sustained speech segment). This diagnostic aspect of speech-ABR is subject to further research [76-78].
Acknowledgement
This work was supported by Research University (RU) Grant (1001/PPSK/812114), University Sains Malaysia. The authors were grateful to all children and their respected parents for the ongoing cooperation during the data collection process.
References
- Skoe E, Kraus N (2010) Auditory brain stem response to complex sounds: A tutorial. Ear Hear 31(3): 302-324.
- Moore BC J (1998) Cochlear Hearing Loss. Whurr Publishers Ltd, London, UK.
- Northern JL, Downs MP (2014) Hearing in Children, sixth ed. San Diego, CA, Plural Publishing, US.
- Greenberg S (1980) Neural Temporal Coding of Pitch and Vowel Quality: Human Frequency-Following Response Studies of Complex Signals. Los Angeles, CA: Phonetics Laboratory, Department of Linguistics, UCLA. Hall JW (2007). New Handbook of Auditory Evoked Responses. Boston: Pearson.
- Koravand A, Jutras B, Lassonde B (2013) Auditory event related potentials in children with peripheral hearing loss. Clin Neurophysiol 124(7): 1439-1447.
- Sharma A, Glick H, Deeves E, Duncan E (2015) The P1 biomarker for assessing cortical maturation in pediatric hearing loss: a review. Otorinolaringologia 65(4): 103-114.
- Studer Eichenberger E, Studer Eichenberger F, Koenig T (2016) Statistical learning, syllable processing, and speech production in healthy hearing and hearing-impaired preschool children: a mismatch negativity study. Ear Hear 37: 57-71.
- Nash Kille A, Sharma A (2014) Inter-trial coherence as a marker of cortical phase synchrony in children with sensorineural hearing loss and auditory neuropathy spectrum disorder fitted with hearing aids and cochlear implants. Clin Neurophysiol 125(7): 1459-1470.
- Yang FF, Mc Pherson B, Shu H, Xiao Y (2012) Central auditory nervous system dysfunction in infants with non-syndromic cleft lip and/or palate. Int J Pediatr Otorhinolaryngol 76(1): 82-89.
- Anwar A, Singleton A, Fang Y, Wang B, Shapiro W, et al. (2017) The value of intraoperative EABRs in auditory brainstem implantation. Int J Pediatr Otorhinolaryngol 101: 158-163.
- Koohi N, Vickers, DA, Lakshmanan R, Chandrashekar H, Werring DJ (2017) Hearing characteristics of stroke patients: prevalence and characteristics of hearing impairment and auditory processing disorders in stroke patients. J Am Acad Audiol 28(6): 491-505.
- Birkent ÖF, Karlıdağ T, Başar F, Yalçın Ş, Kaygusuz I, et al. (2017) Evaluation of the relationship between the air-bone gap and prolonged ABR latencies in mixed-type hearing loss. J Int Adv Otol 13(1): 88-92.
- Cobb KM, Stuart A (2016) Neonate auditory brainstem responses to CE-Chirp and CE-Chirp octave band stimuli I: versus click and tone burst stimuli. Ear Hear 37(6): 710-723.
- Prendergast G, Guest H, Munro KJ, Kluk K, Leger A, et al. (2017) Effects of noise exposure on young adults with normal audiograms I: Electrophysiology. Hear Res 344: 68-81.
- Koravand A, Al Osman R, Rivest V (2017) Speech-evoked auditory brainstem responses in children with hearing loss. Int J Pediatr Otorhinolaryngol 99: 24-29.
- Dhar S, Abel R, Hornickel J, Nicol T, Skoe E, et al. (2009) Exploring the relationship between physiological measures of cochlear and brainstem function. Clin Neurophysiol 120(5): 959-966.
- Song JH, Banai K, Russo NM, Kraus N (2006) On the relationship between speech- and nonspeech-evoked auditory brainstem responses. Audiol Neurootol 11(4): 233-241.
- Ahadi M, Pourbakht A, Jafari AH, Jalaie S (2014) Effects of stimulus presentation mode and subcortical laterality in speech-evoked auditory brainstem responses. Int J Audiol 53(4): 243-249.
- Uppunda AK, Bhat J, D'costa PE, Raj M, Kumar K (2015) Binaural interaction component in speech evoked auditory brainstem responses. J Int Adv Otol 11(2): 114-117.
- Chandrasekaran B, Hornickel J, Skoe E (2009) Context dependent encoding in the human auditory brainstem relates to hearing speech in noise: Implications for developmental dyslexia. Neuron 64(3): 311-319.
- Cunningham J, Nicol T, Zecker SG, Bradlow A, Kraus N (2001) Neurobiologic responses to speech in noise in children with learning problems: Deficits and strategies for improvement. Clin Neurophysiol 112(5): 758-767.
- Hornickel J, Skoe E, Nicol T, Zecker S, Kraus N (2009) Subcortical differentiation of stop consonants relates to reading and speech-in-noise perception. Proc Natl Acad Sci USA 106(31): 13022-13027.
- Filippini R, Befi Lopes, DM, Schochat E (2012) Efficacy of auditory training using the auditory brainstem response to complex sounds: auditory processing disorder and specific language impairment. Folia Phoniatr Logop 64(5): 217-226.
- Krishnamurti S, Forrester J, Rutledge C, Holmes GW (2013) A case study of the changes in the speech-evoked auditory brainstem response associated with auditory training in children with auditory processing disorders. Int J Pediatr Otorhinolaryngol 77(4): 594-604.
- Russo NM, Nicol TG, Zecker S, Hayes EA, Kraus N (2005) Auditory training improves neural timing in the human brainstem. Behav Brain Res 156(1): 95-103.
- Song JH, Skoe E, Wong PC, Kraus N (2008) Plasticity in the adult human auditory brainstem following short-term linguistic training. J Cogn Neurosci 20(10): 1892-1902.
- KingC, Warrier CM, Hayes E, Kraus N (2002) Deficits in auditory brainstem pathway encoding of speech sounds in children with learning problems. Neurosci Lett 319(2): 111-115.
- Sanfins MD, Borges LR, Ubiali T, Colella Santos MF (2017) Speech auditory brainstem response (speech ABR) in the differential diagnosis of scholastic difficulties. Braz J Otorhinolaryngol 83(1): 112-116.
- Wible B, Nicol T, Kraus N (2004) Atypical brainstem representation of onset and formant structure of speech sounds in children with language-based learning problems. Biol Psychol 67(3): 299-317.
- Gonçalves IC, Wertzner HF, Samelli AG (2011) Speech and non-speech processing in children with phonological disorders: an electrophysiological study. Clinics (Sao Paulo) 66(2): 293-298.
- Jafari Z, Malayeri S, Rostami R (2015) Subcortical encoding of speech cues in children with attention deficit hyperactivity disorder. Clin Neurophysiol 126(2): 325-332.
- Russo N, Nicol T, Trommer B (2009) Brainstem transcription of speech is disrupted in children with autism spectrum disorders. Dev Sci 12(4): 557-567.
- Elkabariti RH, Khalil LH, Husein R, (2014) Speech evoked auditory brainstem response findings in children with epilepsy. Int J Pediatr Otorhinolaryngol 78(8): 1277-1280.
- Tahaei AA, Ashayeri H, Pourbakht A, Kamali M (2014) Speech evoked auditory brainstem response in stuttering. Scientifica (Cairo) 2014: 328646.
- Anderson S, Parbery Clark A, White Schwoch T, Drehobl S, Kraus N (2013) Effects of hearing loss on the subcortical representation of speech cues. J Acoust Soc Am 133(5): 3030-3038.
- Leite RA, Magliaro FCL, Raimundo JC, (2016) Effect of hearing aids use on speech stimulus decoding through speech-evoked ABR. Brazilian Journal of Otorhinolaryngology.
- Nada NM, Kolkaila EA, Gabr TA, Trandil H, El Mahallawi, et al. (2016) Speech auditory brainstem response audiometry in adults with sensorineural hearing loss. Egyptian Journal of Ear, Nose, Throat and Allied Sciences 17(2): 87-94.
- Kraemer HC, Morgan G, Leech N, Gliner JA, Vaske JJ (2003) Measures of clinical significance. J Am Acad Child Adolesc Psychiatry 42(12): 1524-1529.
- Sanfins MD, Borges LR, Ubiali T, Donadon C, Diniz Hein TA, et al. (2016) Speech-evoked brainstem response in normal adolescent and children speakers of Brazilian Portuguese. Int J Pediatr Otorhinolaryngol 90: 12-19.
- Møller AR (1981) Latency in the ascending auditory pathway determined using continuous sounds: Comparison between transient and envelope latency. Brain Res 207(1): 184-188.
- Picton TW, Stapells DR, Campbell KB (1981) Auditory evoked potentials from the human cochlea and brainstem. J Otolaryngol Suppl 9: 1-41.
- Archana G, Kishan MM, Kumar V (2015) Latency-intensity function of speech evoked auditory brainstem responses in individuals with moderate sensory neural hearing loss. Indian J Otol 21(4): 254-259.
- Keith WJ, Greville KA (1987) Effects of audiometric configuration on the auditory brainstem response. Ear Hear 8(1): 49-55.
- Strelcyk O, Christoforidis D, Dau T (2009) Relation between derived-band auditory brainstem response latencies and behavioral frequency selectivity. J Acoust Soc Am 126(4): 1878-1888.
- Gorga MP, Reiland JK, Beauchaine KA (1985) Auditory brainstem responses in a case of high-frequency conductive hearing loss. J Speech Hear Disord 50: 346-350.
- Hyde M (1985) Effect of cochlear lesions on the ABR. In J Jacobson (Eds.), The Auditory Brainstem Response. San Diego: College Hill Press pp. 133-146.
- Galambos R, Hecox KE (1978) Clinical applications of the auditory brain stem response. Otolaryngol Clin North Am 11(3): 709-722.
- Aihara N, Murakami S, Takemura K (2013) Interaural difference of wave V predicting postoperative hearing in Gardner-Robertson Class II acoustic neuroma patients. J Neurol Surg B Skull Base 74(5): 274-278.
- Lin HC, Chou YC, Wang C, Hung LW, Shih CP, et al. (2017) Correlation between auditory brainstem response and hearing prognosis in idiopathic sudden sensorineural hearing loss patients. Auris Nasus Larynx 44(6): 678-684.
- Pushpalatha, ZV, Konadath S (2016) Auditory brainstem responses for click and CE-chirp stimuli in individuals with and without occupational noise exposure. Noise Health 18(84): 260-265.
- Chiappa KH, Gladstone KJ, Young RR (1979) Brainstem auditory evoked responses: studies of waveform variations in 50 normal human subjects. Arch Neurol 36(2): 81-87.
- Edwards RM, Buchwald JS, Tanguay PE (1982) Sources of variability in auditory brain stem evoked potential measures over time. Electroenceph Clin Neurophysiol 53(2): 125-132.
- Steinhoff HJ, Böhnke F, Janssen T (1988) Click ABR intensity-latency characteristics in diagnosing conductive and cochlear hearing losses. Arch Otorhinolaryngol 245(5): 259-265.
- Wible B, Nicol T, Kraus N (2005) Correlation between brainstem and cortical auditory processes in normal and language-impaired children. Brain 128(Pt2): 417-423.
- Fitzgibbons PJ, Gordon Salant S (1987) Minimum stimulus levels for temporal gap resolution in listeners with sensorineural hearing loss. J Acoust Soc Am 81(5): 1542-1545.
- Glasberg BR, Moore BC, Bacon SP (1987) Gap detection and masking in hearing-impaired and normal-hearing subjects. J Acoust Soc Am 81(5): 1546-1556.
- Mishra SK, Dey R, Davessar JL (2015) Temporal resolution of the normal ear in listeners with unilateral hearing impairment. J Assoc Res Otolaryngol 16(6): 773-782.
- Boscariol M, Guimarães CA, Hage SR, Garcia VL, Schmutzler KM, et al. (2011) Auditory processing disorder in patients with language-learning impairment and correlation with malformation of cortical development. Brain Dev 33(10): 824-831.
- Dias KZ, Jutras B, Acrani I (2012) Random Gap Detection Test (RGDT) performance of individuals with central auditory processing disorders from 5 to 25 years of age. Int J Pediatr Otorhinolaryngol 76(2): 174-178.
- Sanju H K, Bohra V, Sinha SK (2017) Speech evoked auditory brainstem response and gap detection threshold in middle-aged individual. Eur Arch Otorhinolaryngol 274(4): 2041-2048.
- Akhoun I, Gallégo S, Moulin A, Ménard M, Veuillet E, et al. (2008) The temporal relationship between speech auditory brainstem responses and the acoustic pattern of the phoneme /ba/ in normal-hearing adults. Clin Neurophysiol 119(4): 922-933.
- Hoormann J, Falkenstein M, Hohnsbein J (1992) The human frequency-following response (FFR): normal variability and relation to the click-evoked brainstem response. Hear Res 59(2): 179-188.
- Kraus N, Nicol T (2005) Brainstem origins for cortical ‘what’ and ‘where’ pathways in the auditory system. Trends Neurosci 28(4): 176-181.
- Russo N, Nicol T, Musacchia G, Nina Kraus (2004) Brainstem responses to speech syllables. Clin Neurophysiol 115(9): 2021-2030.
- Abd El Ghaffar NM, El Gharib AM, Kolkaila EA (2017) Speech-evoked auditory brainstem response with ipsilateral noise in adults with unilateral hearing loss. Acta Otolaryngol 138(2): 145-152.
- Lewis JD, Kopun J, Neely ST (2015) Tone-burst auditory brainstem response wave V latencies in normal-hearing and hearing-impaired ears. J Acoust Soc Am 138(5): 3210-3219.
- Aizawa N, Eggermont JJ (2006) Effects of noise-induced hearing loss at a young age on voice onset time and gap-in-noise representations in adult cat primary auditory cortex. J Assoc Res Otolaryngol 7(1): 71-81.
- Bones O, Hopkins K, Krishnan A, Plack CJ (2014) Phase locked neural activity in the human brainstem predicts preference for musical consonance. Neuropsychologia 58: 23-32.
- Galbraith GC, Arbagey PW, Branski R, Comerci N, Rector PM (1995) Intelligible speech encoded in the human brain stem frequency-following response. Neuroreport 6(17): 2363-2367.
- Jafari Z, Malayeri S, Rostami R (2015) Subcortical encoding of speech cues in children with attention deficit hyperactivity disorder. Clin Neurophysiol 126(2): 325-332.
- Johnson KL, Nicol T, Zecker SG, Bradlow AR Skoe E et al. (2008) Brainstem encoding of voiced consonant-vowel stop syllables. Clin Neurophysiol 119(11): 2623-2635.
- Krishnan A (2002) Human frequency-following responses: representation of steady-state synthetic vowels. Hear Res 166(1-2): 192-201.
- Krishnan A, Xu Y, Gandour J, Cariani P (2005) Encoding of pitch in the human brainstem is sensitive to language experience. Brain Res Cogn Brain Res 25(1): 161-168.
- Krishnan A, Swaminathan J, Gandour JT (2009) Experience-dependent enhancement of linguistic pitch representation in the brainstem is not specific to a speech context. J Cogn Neurosci 21(6): 1092-1105.
- Song JH, Nicol T, Kraus N (2011) Test-retest reliability of the speech-evoked auditory brainstem response. Clin Neurophysiol 122(2): 346-355.
- Shafer VL, Schwartz RG, Martin B (2011) Evidence of deficient central speech processing in children with specific language impairment: The T-complex. Clin Neurophysiol 122(6): 1137-1155.
- Wong PC, Skoe E, Russo NM, Dees T, Kraus N (2007) Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nat Neurosci 10(4): 420-422.
- Xu Y, Krishnan A, Gandour JT (2006) Specificity of experience dependent pitch representation in the brainstem. Neuroreport 17(15): 1601-1605.