Acessibilidade / Reportar erro

Encoding of speech sounds at auditory brainstem level in good and poor hearing aid performers Please cite this article as: Shetty HN, Puttabasappa M. Encoding of speech sounds at auditory brainstem level in good and poor hearing aid performers. Braz J Otorhinolaryngol. 2017;83:512-22.

Abstract

Introduction:

Hearing aids are prescribed to alleviate loss of audibility. It has been reported that about 31% of hearing aid users reject their own hearing aid because of annoyance towards background noise. The reason for dissatisfaction can be located anywhere from the hearing aid microphone till the integrity of neurons along the auditory pathway.

Objectives:

To measure spectra from the output of hearing aid at the ear canal level and frequency following response recorded at the auditory brainstem from individuals with hearing impairment.

Methods:

A total of sixty participants having moderate sensorineural hearing impairment with age range from 15 to 65 years were involved. Each participant was classified as either Good or Poor Hearing aid Performers based on acceptable noise level measure. Stimuli /da/ and /si/ were presented through loudspeaker at 65 dB SPL. At the ear canal, the spectra were measured in the unaided and aided conditions. At auditory brainstem, frequency following response were recorded to the same stimuli from the participants.

Results:

Spectrum measured in each condition at ear canal was same in good hearing aid performers and poor hearing aid performers. At brainstem level, better F0 encoding; F0 and F1 energies were significantly higher in good hearing aid performers than in poor hearing aid performers. Though the hearing aid spectra were almost same between good hearing aid performers and poor hearing aid performers, subtle physiological variations exist at the auditory brainstem.

Conclusion:

The result of the present study suggests that neural encoding of speech sound at the brainstem level might be mediated distinctly in good hearing aid performers from that of poor hearing aid performers. Thus, it can be inferred that subtle physiological changes are evident at the auditory brainstem in a person who is willing to accept noise from those who are not willing to accept noise.

KEYWORDS
Frequency following response; Acceptable noise level; Hearing aid performer

Resumo

Introdução:

Os aparelhos auditivos são prescritos para aliviar a perda de audibilidade. Tem sido relatado que 31% dos usuários rejeitam seu aparelho auditivo devido ao desconforto com o ruído de fundo. A razão para a insatisfação pode estar situada em qualquer local desde o microfone do aparelho auditivo até a integridade de neurônios ao longo da via auditiva.

Objetivos:

Medir espectros desde a saída do aparelho auditivo no nível do meato acústico externo e frequência de resposta (FFR) registrada no tronco encefálico de indivíduos com deficiência auditiva.

Método:

Foram selecionados 60 participantes com deficiência auditiva neurossensorial moderada, de 15 a 65 anos. Cada participante foi classificado como usuário bom ou mau de prótese auditiva (GHP ou PHP) com base na medida de nível de ruído aceitável (ANL). Estímulos/da/e/si/foram apresentados em alto-falante a 65 dB SPL. No meato acústico externo, os espectros foram medidos nas condições sem aparelho e com aparelho. No tronco encefálico auditivo, FFR foram registradas para os mesmos estímulos dos participantes.

Resultados:

Os espectros medidos em cada condição no meato acústico externo foram os mesmos em GHP e PHP. No nível do tronco cerebral, melhor codificação F0; energias de F0 e F1 foram significativamente maiores em GHP do que em PHP. Embora os espectros do aparelho auditivo fossem quase os mesmos entre GHP e PHP, existem variações fisiológicas sutis no tronco encefálico auditivo.

Conclusão:

O resultado do presente estudo sugere que a codificação neural do som da fala no nível do tronco encefálico pode ser mediada distintamente em GHP em comparação com PHP. Assim, pode-se inferir que mudanças fisiológicas sutis são evidentes no tronco encefálico em uma pessoa que está disposta a aceitar o ruído em comparação com aqueles que não estão dispostos a aceitar o ruído.

PALAVRAS-CHAVE
Frequência seguida de resposta; Nível de ruído aceitável; Usuário de aparelho auditivo

Introduction

Hearing aid is one of the common rehabilitative measures for individuals with permanent hearing impairment. In some cases of hearing losses, hearing aids can be used transitory. However, hearing aid users often complain of background noise resulting in rejection of hearing AID.11 Bentler RA, Niebuhr DP, Getta JP, Anderson CV. Longitudinal study of hearing aid effectiveness. II: Subjective measures. J Speech Hear Res. 1993;36:820-31. Kochkin22 Kochkin S. On the issue of value: hearing aid benefit, price, satisfaction, and brand repurchase rates. Hear Rev. 2003;10:12-26. has reported that about 31% of the hearing aid users reject their hearing aid because of background noise. Various outcome measures are available that consider background noise as one of the factors to have an effect on satisfaction with the hearing aid. Unfortunately, these outcome measures are to be administered after a period of experience with hearing aid. Besides measures such as speech in noise test, quick speech in noise test, competing noise test, and hearing in noise test are being used to predict the hearing aid benefit.33 Nemes J. Despite benefits of outcomes measures, advocates say they're under used. Hear J. 2003;56:19-25. Though these tests are sensitive to measure speech performance in noise and are administered at the time of fitting hearing aid; they fail to predict real-world benefit and/or satisfaction from hearing Aids.44 Nabelek AK. Acceptance of background noise may be key to successful fittings. Hear J. 2005;58:10-5. This issue is addressed partly by acceptable noise level (ANL) measure introduced by Nabelek et al.,55 Nabelek AK, Tucker FM, Letowski TR. Toleration of background noises: relationship with patterns of hearing aid use by elderly persons. J Speech Hear Res. 1991;34:679-85. in which the client rates annoyance due to background noise in the presence of speech.

Nabelek et al.66 Nabelek AK, Freyaldenhoven MC, Tampas JW, Burchfiel SB, Muenchen RA. Acceptable noise level as a predictor of hearing aid use. J Am Acad Audiol. 2006;17:626-39. demonstrated that value of ANLs predict good and poor hearing aid performers with 85% accuracy. The ANL is not affected by the type of background noise,55 Nabelek AK, Tucker FM, Letowski TR. Toleration of background noises: relationship with patterns of hearing aid use by elderly persons. J Speech Hear Res. 1991;34:679-85. preference of background sounds,77 Freyaldenhoven MC, Smiley DF. Acceptance of background noise in children with normal hearing. J Educ Audiol. 2006;13:27-31. primary language of the listener,88 von Hapsburg D, Bahng J. Acceptance of background noise levels in bilingual (Korean-English) listeners. J Am Acad Audiol. 2006;17:649-58. speech presentation levels,77 Freyaldenhoven MC, Smiley DF. Acceptance of background noise in children with normal hearing. J Educ Audiol. 2006;13:27-31. age, hearing sensitivity and language99 Brannstrom KJ, Lantz J, Nielsen LH, Olsen SO. Acceptable noise level with Danish, Swedish, and non-semantic speech materials. Int J Audiol. 2012;51:146-56. content of speech signal and speaker gender.1010 Plyler PN, Alworth LN, Rossini TP, Mapes KE. Effects of speech signal content and speaker gender on acceptance of noise in listeners with normal hearing. Int J Audiol. 2011;50:243-8. Harkrider,1111 Harkrider AW, Tampas JW. Differences in responses from the cochleae and central nervous systems of females with low versus high acceptable noise levels. J Am Acad Audiol. 2006;17:667-76. studied the physiological correlate of ANL involved at higher auditory centres utilizing electrophysiological measurement. In individuals with low ANLs (i.e., greater background noise acceptance), amplitudes of wave V of auditory brainstem response (ABR), all components of middle latency response (MLR), and late latency response (LLR) were noted to be significantly prolonged when compared to individuals who obtained high ANLs (lower background noise acceptance). This is due to stronger efferent mechanism, such that sensory inputs are suppressed and/or central afferent mechanism is less active.1212 Tampas JW, Harkrider AW, Hedrick MS. Neurophysiological indices of speech and nonspeech stimulus processing. J Speech Lang Hear Res. 2005;48:1147-64. Thus, ANL proved to be physiologically sensitive measure. However, it is interesting to know the way in which amplified speech is represented physiologically in good and poor hearing aid performers.

Despite advancement in hearing aid technology, some individuals accept hearing aid and others may reject in spite of fact that they have a similar hearing loss in terms of degree, type, and configuration. The variability in satisfaction from rehabilitative device might probably be due to the processing parameters of hearing aid, and/or at interaction between output of hearing aid and its acoustic parameters relayed through different parts of auditory pathway.1313 Souza PE, Tremblay KL. New perspectives on assessing amplification effects. Trends Amplif. 2006;10:119-43. In the present study, output of hearing aid is investigated at ear canal and at auditory brainstem.

Over the decades, researchers have used probe tube microphone (PTM) system to measure the effect of hearing aid processing on acoustics of speech. The PTM measurement reflects the acoustic effect of the factors such as pinna, ear canal, head and torso.1414 Mueller HG. Probe microphone measurement: unplugged. Hear Rev. 2005;48:10-36. Primarily, the PTM is used to optimize/verify the hearing aid gain to match with the target gain at different frequencies as prescribed by the fitting formula.1515 Cunningham DR, Lao-Davila RG, Eisenmenger BA, Lazich RW. Study finds use of Live Speech Mapping reduces follow-up visits and saves money. J Hear. 2002;55:43-6. It is well established that output of hearing aid at the ear canal will alter amplitudes of formants leading to misperception. An experiment was conducted by Stelmachowicz et al.1616 Stelmachowicz PG, Kopun J, Mace A, Lewis DE, Nittrouer S. The perception of amplified speech by listeners with hearing loss: acoustic correlates. J Acoust Soc Am. 1995;98:1388-99. who recorded output of the hearing aid at the ear canal using linear and non-linear hearing aids on three listeners with mild-moderate sensorineural hearing loss. They carried out spectral analysis on these recorded stimuli. The results revealed a precipitous roll-off in the high frequency response, thus limiting the information of a consonant cues. On similar line, Souza and Tremblay1717 Souza PE, Tremblay KL. Combining acoustic, electrophysiological and behavioral measures of hearing aids. Scottsdale, AZ: American Auditory Society; 2005. conducted a study to correlate consonant errors to acoustic analysis of amplified speech in subjects with mild to moderate sensorineural hearing loss. They observed that /da/ stimulus was consistently misperceived as /ga/. This was attributed to the amplitude of aided burst spectrum of /da/ which was found to be similar to the unprocessed burst spectrum amplitude of /ga/. Kewley-Port1818 Kewley-Port D. Time-varying features as correlates of place of articulation in stop consonants. J Acoust Soc Am. 1983;73:322-35. reported that identification of stop consonants in the initial position requires the spectrum of burst as the primary cue for speech recognition. Thus, after amplification the stop consonants are more likely to have place error.1717 Souza PE, Tremblay KL. Combining acoustic, electrophysiological and behavioral measures of hearing aids. Scottsdale, AZ: American Auditory Society; 2005. However, amplified consonant-vowel combination of fricative or affricative tends to show manner errors, as consistent misperception of /ʓi/ for /dʓi/ speech sounds1919 Souza PE, Tremblay KL, Davies-Venn E, Kalstein L. Explaining consonant errors using short-term audibility. Salt Lake City, UT: Am Acad Audiol; 2004. was noted. When the acoustic output of hearing aid was analyzed, it was revealed that amplitude spectrum of fricative /ʓi/ was similar to the unprocessed affricative spectrum amplitude of /dʓi/. Hence, performing spectral analysis of the output of hearing aid recorded at the ear canal throws light on the processing parameters of hearing aid. There are instances in which acoustic cues are distorted but a listener still recognizes correctly. This could be due to redundancy or from the contextual cues of speech. In some other instances acoustic cues are preserved but a listener fails to recognize speech sound. This may possibly be because of insufficient sensitivity in cochlea and/or concomitant changes at different levels of auditory pathway. Hence, an evoked potential recorded to speech stimuli should be used to validate perception registered at different levels of auditory pathway. In the present study, evoked response at the level of auditory brainstem of good and poor hearing aid performers is investigated.

The frequency following responses (FFR) has been extensively studied to understand the physiological processing of speech at the auditory brainstem level. The FFR is a phase-locked response to periodic aspects of stimuli, including speech, up to 1000 Hz 2020 Kraus N, Nicol T. Brainstem origins for cortical 'what' and 'where' pathways in the auditory system. Trends Neurosci. 2005;28:176-81.,2121 Young ED, Sachs MB. Representation of steady-state vowels in the temporal aspects of the discharge patterns of populations of auditory-nerve fibers. J Acoust Soc Am. 1979;66:1381-403. from the neural population of inferior colliculus of rostral brainstem.2222 Worden FG, Marsh JT. Frequency-following (microphonic-like) neural responses evoked by sound. Electroencephalogr Clin Neurophysiol. 1968;25:42-52. The FFR has been reliably recorded to consonant-vowel sounds /da/. 2323 Russo N, Nicol T, Musacchia G, Kraus N. Brainstem responses to speech syllables. Clin Neurophysiol. 2004;115:2021-30.

24 Russo NM, Nicol TG, Zecker SG, Hayes EA, Kraus N. Auditory training improves neural timing in the human brainstem. Behav Brain Res. 2005;156:95-103.

25 Johnson KL, Nicol T, Zecker SG, Kraus N. Developmental plasticity in the human auditory brainstem. J Neurosci. 2008;28:4000-7.
-2626 Chandrasekaran B, Kraus N. The scalp-recorded brainstem response to speech: neural origins and plasticity. Psychophysiology. 2010;47:236-46. Further, FFR to /da/ stimulus has been investigated in monaural2727 Cunningham J, Nicol T, Zecker SG, Bradlow A, Kraus N. Neurobiologic responses to speech in noise in children with learning problems: deficits and strategies for improvement. Clin Neurophysiol. 2001;112:758-67. and binaural2828 Parbery-Clark A, Skoe E, Lam C, Kraus N. Musician enhancement for speech-in-noise. Ear Hear. 2009;30:653-61. conditions; in the presence of background noise;2727 Cunningham J, Nicol T, Zecker SG, Bradlow A, Kraus N. Neurobiologic responses to speech in noise in children with learning problems: deficits and strategies for improvement. Clin Neurophysiol. 2001;112:758-67. and stimulation of either right or left ear.2929 Hornickel J, Skoe E, Kraus N. Subcortical laterality of speech encoding. Audiol Neurootol. 2009;14:198-207. The FFR was successfully recorded using loudspeaker as a transducer to deliver the stimuli /da/ and /si/.3030 Hemanth N, Manjula P. Representation of speech syllables at auditory brainstem. J Indian Speech Hear Assoc. 2012;26:1-13. From this, it is clear that FFR is a stimulus contingent response that is most robust for mid- and low-frequencies. Though the frequency response of hearing aid is up to 6500 Hz, the FFR is a sensitive tool to any change in processing in the auditory brainstem, such that it answers the question on how amplified speech sounds are encoded in mid- and low-frequencies from good and poor hearing aid performers.

From the existing literature, it can be inferred that spectral analyses of hearing aid output obtained using the PTM give information on hearing aid processing at the ear canal level. Further, stimulus from the ear canal is relayed to the auditory brainstem level and is measured using FFR. The FFR will help in inferring the neural encoding of the ongoing speech. These measures give insight into the way in which the speech is neurally encoded at the brainstem level, in individuals with sensorineural hearing impairment who are classified as good and poor hearing aid performers having comparable type, degree and configuration of hearing loss. Therefore, the present study intends to investigate hearing aid output at the ear canal to determine extent of alteration caused by the hearing aid on the spectral parameters. In addition, way in which amplified speech is represented at the brainstem level in good and poor hearing aid performers also is being investigated. The objectives formulated for the study were to compare: (1) spectral changes between GHP and PHP in unaided and aided conditions at the ear canal using the PTM; and (2) neural encoding of speech sounds at auditory brainstem level in GHP and PHP.

Methods

Participants

A total of 60 participants who had bilateral moderate sensorineural hearing loss with a flat configuration were involved in the study. Flat configuration was operationally defined as the difference between least and highest air-conduction thresholds being less than 20 dB in the range from 0.25 to 8 kHz.3131 Pittman AL, Stelmachowicz PG. Hearing loss in children and adults: audiometric configuration, asymmetry, and progression. Ear Hear. 2003;24:198-205. The age range of the participants was from 15 to 65 years. They had speech identification scores (SIS) that was greater than or equal to 75% at 40 dB SL (re: speech reception threshold, SRT). The test ear had normal middle ear status as indicated by 'A' type tympanogram with middle ear peak pressure ranging from +50 daPa to −100 daPa, and admittance ranging from 0.5 mL to 1.75 mL. The auditory brainstem response (ABR) was recorded at two repetition rates of 11.1 s and 90.1 s at 90 dB nHL to ensure that there was no retro cochlear pathology. The latency difference of V peak of ABR was found to be less than 0.8 ms between the two repetition rates. All participants were naïve hearing aid users and there was no self-reported history of other otological and neurological problems. The participants were further classified into Good or Poor Hearing aid Performers (GHP or PHP) using the acceptable noise level (ANL).66 Nabelek AK, Freyaldenhoven MC, Tampas JW, Burchfiel SB, Muenchen RA. Acceptable noise level as a predictor of hearing aid use. J Am Acad Audiol. 2006;17:626-39. Those participants who obtained an ANL score of ≤7 were considered as good hearing aid performers and a score of ≥13 were considered as poor hearing aid performers.66 Nabelek AK, Freyaldenhoven MC, Tampas JW, Burchfiel SB, Muenchen RA. Acceptable noise level as a predictor of hearing aid use. J Am Acad Audiol. 2006;17:626-39. The demographic data of each participant in clinical group are tabulated in Table 1. The hearing thresholds at each audiometric frequency of the test ear of the good and poor hearing aid performers are depicted in Fig. 1. The study was approved by the All India Institute of Speech and Hearing Ethics Committee for Research in Humans. Informed consent was obtained from each participant.

Table 1
Demographic data of study participants.

Figure 1
Audiograms of good and poor hearing aid performers.

Stimuli

Naturally produced consonant vowel (CV) tokens were utilized as target test stimuli. An adult male with normal voice was used to record the CV tokens. The duration of /da/ and /si/ stimuli was 94 ms and 301 ms, respectively. For /da/, the voice onset time was 18 ms, burst duration was 5 ms, transition duration was 37 ms, and vowel duration was 34 ms. For /si/, the fricative duration was 159.3 ms, transition duration was 47.1 ms and the vowel duration was 94.6 ms. Both the stimuli were converted from '.wav' to '.avg' format using wavtoavg m-file of Brainstem tool box. The '.avg' format of both the stimuli were band pass filtered from 30 to 3000 Hz using Neuroscan (Scan 2-version 4.4) to know the functional relationship between the acoustic structure of speech and the brain stem response to speech. The 'stimulus.avg', waveforms and spectrograms of the two CV tokens are depicted in Fig. 2. Table 2 summarizes the fundamental frequency (F0) and the first two formant frequencies (F1 and F2) of the vowel component of /da/ and /si/ stimuli. The onset to steady state F0, F1 and F2 within the transition duration (37 ms) of /da/ stimulus, and frequency components within the transition duration (42 ms) of /i/ portion of /si/ stimulus were measured using Praat (version 5.1.29) software.

Table 2
Fundamental frequency and the two formant frequencies (in Hz) at the transition duration of original and filtered version of /ɖa/ and /si/ stimuli.

Figure 2
(A) and (D) are the spectrograms of /da/ and /si/ stimuli. The dark black solid line in both stimuli indicates the F0, which has a falling pattern. The formant frequencies are represented by white lines. The F1 and F2 of /da/ is flat in pattern. The F1 of /si/ stimulus is falling in pattern and F2 is raising in pattern. (C) and (F) are the waveforms of /da/ and /si/ stimuli. For sound /da/, the voice onset time was 18 ms, the burst duration was 5 ms, the transition duration was 37 ms and vowel duration of 34 ms. For the sound /si/, the fricative duration was 159.3 ms, transition duration was 47.1 ms and vowel duration was 94.6 ms.

Further, Kannada passage developed by Sairam and Manjula3232 Sairam VVS, Manjula P. Long term average speech spectrum in Kannada. University of Mysore; 2002. was read out in normal vocal effort by a female speaker was recorded using Adobe Audition (version 3) software. This recorded passage was used to determine the acceptable noise level (ANL). A goodness test was performed in order to verify the quality of the recorded Kannada passage, in which ten listeners with normal hearing rated the passage for naturalness.

Hearing aid

Digital Behind The Ear (BTE) hearing aid was used to record the output at the ear canal and at the auditory brainstem response from each participant. According to the technical specifications, frequency range of test hearing aid extended from 0.210 to 6.5 kHz. The peak full-on gain was 58 dB and high-frequency average full-on gain was 49 dB. The functioning of the hearing aid was ensured at the beginning of the data collection and repeated periodically during data collection.

Procedure

Each participant was classified into good and poor hearing aid performer using the behavioural ANL test. The test hearing aid with custom ear mould was fitted to each participant and its gain was optimized. To optimize the hearing aid gain, six Ling's syllables were presented at a calibrated level of 65 dB SPL through the audiometer in a sound field. The gain and the frequency response of the hearing aid were manipulated for the audibility of each of six Ling's syllables, through fine tuning option. To know the extent to which spectral feature are preserved by the hearing aid, the output of the hearing aid to each stimulus was recorded at the ear canal using the probe tube microphone measure. Further, the FFR at the brainstem level was recorded to each stimulus, in both unaided and aided conditions.

Acceptable noise level

Acceptable noise level (ANL) evaluates the reaction of the listener to background noise while listening to speech. For the measurement of ANL, the method given by Nabelek et al.55 Nabelek AK, Tucker FM, Letowski TR. Toleration of background noises: relationship with patterns of hearing aid use by elderly persons. J Speech Hear Res. 1991;34:679-85. was adopted. Each study participant was made to sit comfortably on a chair in front of the loudspeaker of the audiometer that was located at 1 m distance and 45° Azimuth. To compute the ANL, most comfort level (MCL) and background noise level (BNL) were measured.

The recorded Kannada passage was routed through the auxiliary input to the loudspeaker of the audiometer. The presentation level set at the level of SRT. Gradually, the level was adjusted in 5 dB-steps to establish the most comfortable level (MCL) and then in smaller steps size of +1 and −2 dB, until the MCL was established reliably. After the MCL was established, speech noise was introduced at 30 dB HL. The level of the speech noise was increased in 5 dB-steps initially, and then in 2 dB-steps, to a point at which the participant was willing to accept the noise without becoming tired or fatigued while listening to and following the words in the story. The maximum level at which he/she could accept or put up with the speech noise without becoming tired was considered as the background noise level (BNL). The level of the speech noise was adjusted until participant was able to 'put-up-with' the noise while following the story. The resultant level was the BNL. The ANL quantifies the acceptable level of background noise and is calculated as the difference between MCL (dB HL) and BNL (dB HL).44 Nabelek AK. Acceptance of background noise may be key to successful fittings. Hear J. 2005;58:10-5. Based on the ANLs, each participant was classified as good (ANL of ≤7 dB) or poor (ANL of ≥13 dB) hearing aid performers.44 Nabelek AK. Acceptance of background noise may be key to successful fittings. Hear J. 2005;58:10-5. The procedure of ANL was repeated twice and the average of the two values was considered as the ANL for each participant.

Hearing aid gain optimization

Each participant was fitted with the digital BTE test hearing aid using a custom made soft shell mould. The hearing aid was programmed using NAL-NL1 prescriptive formula. The real ear measurement was carried out to match the gain of hearing aid with the target gain objectively. Further, the Ling's six speech sounds were presented at 65 dB SPL to optimize the hearing aid gain. Through fine tuning option, the gain and the frequency shaping of the hearing aid were optimized for the audibility of Ling's six sounds.

Hearing aid processed speech at ear canal

The level of the each signal (stored in personal computer) was varied in the audiometer so that the intensity measured was 65 dB SPL in sound level meter. Larson Davis 824 sound level meter (SLM) was positioned at the test ear of the participant. The SLM was set at fast weighting function, and it was ensured that the stimuli /da/ and /si/ were presented at 65 dB SPL, based on peak amplitude level read on the SLM. After the calibration of stimulus was ensured, output spectrum at the ear canal was recorded using the probe tube microphone measurement, in both unaided and aided conditions. The probe tube microphone in the ear canal picks up the spectral energies at approximately half-octave bands from 0.25 kHz to 8 kHz for each speech stimulus. The levels as a function of frequency from 0.25 kHz to 8 kHz, in octaves, were noted down for each stimulus, in the unaided and aided conditions.

Acquisition of the frequency following response

Each participant was seated comfortably in a reclining chair with arm. The electrode sites were cleaned up with skin preparing gel. Disc type silver coated electrodes were placed using conduction gel at the test sites. The FFR was recorded using vertical montage. The non-inverting electrode (+) was placed on the vertex (Cz), the ground electrode was on upper fore head (Fpz) and the inverting electrode (−) was placed on nose. It was ensured that the electrode impedance was less than 5 kΩ for each of the electrodes and that the inter-electrode impedance was less than 2 kΩ.

Prior to recording, calibration of stimuli was ensured using Larson Davis System 824 SLM. The SLM was positioned at reference point. It is the point where the test ear of the participant would be positioned at the time of testing. The SLM was set at fast weighting function for the measurement. It was ensured that both stimuli /da/ and /si/ were presented at 65 dB SPL, based on peak amplitude level read on the SLM.

The loudspeaker of the Auditory Evoked Potential equipment was placed at 45° Azimuth from the participant test ear, located at the calibrated position of 12 inch distance. The height of loudspeaker was adjusted to the level of participant test ear. The participant was instructed to ignore the stimulus and to watch a movie that was muted and played through a battery operated laptop computer. He/she was also asked to minimize the eye and head movement.

For recording the unaided and the aided FFR, the stimulus /da/ was presented through loud speaker at the presentation level of 65 dB SPL to the test ear. The PC-based evoked potential system, Neuroscan 4.4 (Stim 2-version 4.4), controlled the timing of stimulus presentation and delivered an external trigger to the evoked potential recording system, Neuroscan (Scan 2-version 4.4). To allow for a sufficient refractory period within the stimulus sweep, while minimizing the total recording time, an inter-stimulus interval (ISI) of 93 ms. from offset to onset of the next stimulus was included for recording FFR to /da/ stimulus. A similar procedure was repeated to record the unaided and aided FFR for /si/ stimulus. However, for recording unaided and aided FFR to /si/ stimulus, an ISI of 113 ms was used. The order of stimuli while testing on each participant was counter balanced. The FFR was recorded from 1500 sweeps each in condensation and rarefaction polarities, delivered in a homogenous train using the stimulus presentation software Neuroscan 4.4 (Stim 2-version 4.4).

The FFR recording was initiated once a stable electroencephalogram (EEG) was obtained. The ongoing EEG was converted from analogue-to-digital with the rate of 20,000 Hz. The continuous EEG was online band-pass filtered from 30 to 3000 Hz with 12 dB/octave roll-off. This was stored to disc for offline analysis.

Data analyses

The output of the hearing aid in the ear canal for each stimulus in the unaided and aided conditions were analyzed for spectra. Further, the FFR recorded was analyzed for F0, F0 energy and F1 energy obtained for each stimulus. The continuous EEG data were epoched over a window of 160 ms for /da/ stimulus (which included a 30 ms pre-stimulus period and a 130 ms post-stimulus time). The response for /si/ stimulus was epoched over a window of 360 ms (which included a pre-stimulus period of 30 ms and a post-stimulus period of 330 ms). The epoched waveforms were corrected for baseline. The responses were averaged and filtered off-line from 0.030 kHz (high-pass filter, 12 dB/octave) to 3 kHz (low-pass filter, 12 dB/octave). All artefacts exceeding ±35 µV were rejected while averaging the response for each averaged response, in rarefaction and condensation polarity. A minimum of 1450 artefact-free epochs was ensured. The averaged waveforms of rarefaction and condensation polarities were added. Further, the added waveforms were created by averaging two trials recorded for each stimulus, in unaided and aided conditions.

For all the participants, the unaided responses were absent for both the stimuli. From FFR recorded for /da/ stimulus in aided condition, the latency of 'V' peak was identified by visual inspection. The default MATLAB-code of autocorrelation was utilized, in which a range for latency was specified to obtain F0 in the FFR i.e., from noted 'V' peak latency till transition duration. Whereas, the latency of 'a1' corresponding to CV boundary3030 Hemanth N, Manjula P. Representation of speech syllables at auditory brainstem. J Indian Speech Hear Assoc. 2012;26:1-13. in the FFR was identified for /si/ stimulus is shown in Fig. 3. The latency of 'a1' till transition duration was specified in autocorrelation MATLAB code to obtain F0 in the FFR. Further, F0 energy and F1 energy were determined, using 'Brainstem Toolbox' which utilizes the FFT technique (Fig. 4), from the transient response ('V' peak for /da/ stimulus; and 'a1' for /si/ stimulus) till specified transition duration (37 ms for /ɖa/ stimulus; and 47.1 ms for /si/ stimulus).3333 Skoe E, Nicol T, Kraus N. Phase coherence and the human brainstem response to consonant-vowel sounds. In: Association for research in otolaryngology symposium. 2008.

Figure 3
Transition response of FFR obtained from /si/ stimulus in aided condition.

Figure 4
(A) Response corresponding to stimulus at transition portion of /da/ stimulus; (B) showing grand average spectrum of GHP and PHP sub-group. (B) Fundamental frequency and frequency of first formant in FFR for /da/ stimulus.

Results

The spectral data obtained at the ear canal using probe tube measurement and FFR at brainstem level were analyzed in good and poor hearing aid performers. Statistical Package for the Social Sciences (SPSS for window, version 17) software was used to perform the statistical analyses. The results obtained are discussed with respect to each objective.

Hearing aid output at the ear canal

Spectral energy at frequencies from 0.25 to 8 kHz (in octaves) in the unaided and aided conditions, for both the stimuli was analyzed. It was performed to determine representation of energy across frequencies at the ear canal, in good and poor hearing aid performers. The data of spectral energy met the assumption of normal distribution on Kolmogorov-Smirnov normality test (p > 0.05) and homogeneity on Levene's test (F < 2). The spectral energy (0.5-8 kHz in octaves) for both stimuli obtained from both groups, in unaided and aided conditions, was subjected to MANOVA. The result revealed that there was no significant difference between groups in the spectral energy at each octave frequency, in both the unaided and the aided conditions, for /ɖa/ and /si/ stimuli. Thus, the data of spectral energy was combined between groups. Descriptive analysis was carried out separately in the unaided and aided conditions. For /ɖa/ stimulus (Fig. 5), at extreme low frequency (0.25 kHz) and at extreme high frequencies (4 kHz and 8 kHz) the energy in both unaided and aided conditions is relatively minimal than at other frequencies (0.5 kHz, 1 kHz and 2 kHz). For /si/ stimulus (Fig. 6), at extreme low frequencies (0.25 kHz) and at extreme high frequency (8 kHz) the energy in both unaided and aided conditions is relatively minimal compared to other frequencies (1 kHz, 2 kHz and 4 kHz).

Figure 5
Mean and standard deviation of intensity of /si/ stimulus as a function of frequency in unaided and aided conditions.

Figure 6
Mean and standard deviation of intensity of /ɖa/ stimulus as a function of frequency in unaided and aided conditions.

Comparison of FFR in terms of F0, F0 energy and F1 energy in good hearing aid performers and poor hearing aid performers

The F0, F0 energy and F1 energy of FFR between the groups for each stimulus met the assumption of normal distribution on Kolmogorov-Smirnov normality test (p > 0.05) and homogeneity on Levene's test (F < 2) was also performed. Hence, an independent samples t-test was conducted on each data of FFR between GHP (n = 34) and PHP (n = 24) groups. From the mean value of F0 of FFR (Fig. 7), it can be inferred that the F0 was represented better in GHP than in PHP, for each stimulus. Further, the F0 of FFR was compared between GHP and PHP using independent samples test. The result showed that there was a significant better F0 encoding in GHP than PHP for /da/ stimulus (t = 3.41, p = 0.001) and /si/ stimulus (t = 2.84, p = 0.006).

Figure 7
Mean, standard deviation and p-value of independent samples t-test on F0 of FFR for each stimulus, in GHP and PHP groups.

In addition, the F0 energy and the F1 energy of FFR was compared between GHP and PHP groups to each stimulus. From Fig. 8, it was noted that the mean and standard deviation of F0 energy and F1 energy of FFR to each stimulus were higher in GHP than in PHP. Further, to know if there was any significant difference between GHP and PHP in the mean the F0 energy and the F1 energy of FFR for each stimulus, independent samples t-test was performed. The result revealed a significant higher F0 energy in GHP than in PHP for /da/ stimulus (t = 6.80, p = 0.000) and /si/ stimulus (t = 6.20, p = 0.000). Further, a significant higher F1 energy was observed in GHP than PHP for /da/ (t = 3.11, p = 0.002) and /si/ stimulus (t = 5.20, p = 0.000).

Figure 8
Mean and standard deviation of F0 energy and F1 energy for each stimulus in GHP and PHP groups.

Discussion

The aim of the study was to investigate the representation of amplified speech at the ear canal and at the auditory brainstem from the good and the poor hearing aid performers.

Effect of hearing aid processing on spectral parameters of speech stimuli

In the unaided condition for /da/ and /si/ stimuli, energy measured at 2 kHz was relatively larger than at other frequencies (in octave). Further, there was a decline in energy after 4 kHz i.e., an approximately 10 dB per octave for /da/ stimulus and 12 dB per octave for /si/ stimulus (Figs. 5 and 6). This pattern of energy representation as a function of frequencies, for both the stimuli, could be because of frequency response of microphone used in recording the target test stimuli. In the aided condition for /da/ stimulus and /si/ stimulus, the energy measured was relatively less at the two extreme cut off frequencies, which is at low frequency below 0.25 kHz and high frequencies above 4 kHz. Thus, at extreme frequencies, the mean energy in both unaided and aided conditions was less. At low frequency, reduced energy could be because of less gain in that frequency region provided by the prescriptive formula.3434 Dillon H, Katsch R, Byrne D, Ching T, Keidser G, Brewer S. The NAL-NL1 prescription procedure for non-linear hearing aids. National Acoustics Laboratories Research and Development, Sydney: National Acoustics Laboratories; 1998. Additionally, low energy noted in the low frequency region of /da/ and /si/ stimulus could also be because of frequency response of the hearing aid. The low frequency cut-off of the frequency response of the test hearing aid was 0.21 kHz. At high frequencies i.e., above 4 kHz, the energy reduced approximately at the rate of 10 dB per octave for /da/ and 14 dB per octave for /si/ stimulus. This could be the frequency response of /da/ and /si/ stimulus had energy till 4 kHz as noted from unaided condition. Yet another reason could be though the frequency response of hearing aid had 0.216 to 6.5 kHz, energy after 4 kHz gradually reduced per octave. Thus, remarkable energy was noted in the frequency range from 0.5 kHz to 4 kHz. It can be inferred that there is a relatively high amplification in the mid-frequency region of the hearing aid than other two extreme cut-off frequencies (low and high). Informally, participants were instructed to repeat the syllables which were randomly presented for three times. In unaided condition, the participants were unable to identify the CV tokens as the presentation level was 65 dB SPL, which failed to reach audibility range. However, in aided condition, all the participants consistently identified syllables. Further, on spectral analysis, it was noted that the amplitude of aided burst spectrum of /da/ was similar to the unprocessed burst spectrum amplitude of /da/. It was also observed that amplitude spectrum of fricative /si/ was similar to the unprocessed fricative spectrum amplitude of /si/. This infers that hearing aid preserves inherent speech cues at the ear canal.

Comparison of F0 of FFR, F0 energy and F1 energy in good and poor hearing aid performers

The FFR in both unaided and aided conditions were obtained from all the participants. In the unaided condition, the brainstem responses were absent, as the stimuli (/da/ and /si/) were presented at 65 dB SPL, which failed to reach audibility. In the aided condition, the F0 representation in the FFR to each stimulus (/da/ and /si/) remained robust and similar to that of the unprocessed filtered raw stimulus. This indicated that preserved spectral content from hearing aid is relayed to the auditory brainstem level. For /da/ stimulus, the mean F0 of FFR was higher in GHP (133.46 Hz) than in PHP (128.84), such that the difference was found to be significantly different. This was true for F0 of FFR for /si/ stimulus between GHP (134.42 Hz) and PHP (130.84 Hz). Further, the F0 of the aided stimulus of /da/ was 134.95 Hz and that for /si/ was 144.74 Hz. The difference in F0 (in Hz), between encoding of F0 at brainstem level and F0 of aided test stimulus was 1 Hz in GHP and 6 Hz in PHP for /da/ stimulus. Similarly, the difference noted was 10 Hz in GHP and 14 Hz in PHP for /si/ stimulus.

The mean difference between the GHP and PHP in the encoding of F0 was 5 Hz for /da/ and 4 Hz for /si/ stimulus. Though this difference was significant in the encoding of F0 between GHP and PHP for both stimuli, this may not bring a change in speaker identity. This is because, according to Iles3535 Iles MW. Speaker identification as a function of fundamental frequency and resonant frequencies. University of Florida; 1972. a change of up to ±25 Hz in the F0 will not bring about a change in speaker identity. The finding of the study is in accordance with the research report by Horii3636 Horii Y. Fundamental frequency perturbation observed in sustained phonation. J Speech Hear Res. 1979;22:5-19. who reported that a difference of greater than 25 Hz in the F0 between the same two stimuli does not cause difference in speaker identity. Additionally, the intra-subject variability of F0 in a normal vocal effort ranged between ±9.6 Hz.3737 Abberton E. A laryngographic study of voice quality. University of London; 1976. Thus, it can be inferred that the mean F0 of FFR to /da/ and /si/ stimuli was neurally well represented in GHP than PHP, and that both the groups were able to recognize the identity of the speaker.

Further, it was noted that the F0 energy and the F1 energy of FFR to each stimulus were significantly higher in GHP than PHP. The higher energies of F0 and F1 in GHP might be due to stronger efferent fibres that inhibit other harmonics that do not correspond to fundamental frequency and formant frequencies. This is in accordance with the research reports by Ashmore3838 Ashmore JF. A fast motile response in guinea-pig outer hair cells: the cellular basis of the cochlear amplifier. J Physiol. 1987;388:323-47. and Knight.3939 Knight RT. Distributed cortical network for visual attention. J Cogn Neurosci. 1997;9:75-91. To be more specific, central afferent mechanism is stronger in the group of GHP such that neurons at inferior colliculus fire precisely to the harmonics corresponding to F0 and F1. In addition, the efferent mechanism might be stronger such that the efferent fibres inhibit the other harmonics which do not correspond to the fundamental frequency and formant frequencies, thereby fine tuning the auditory input. The excitatory and inhibitory mechanisms of neurons of the underlying neural generator of the inferior colliculus in GHP fire more or less precisely to the corresponding F0 and F1 components of the stimulus. The inference of the present study supports the findings reported by Krishnan.4040 Krishnan A. Human frequency-following responses: representation of steady-state synthetic vowels. Hear Res. 2002;166:192-201. He demonstrated that efferent auditory pathway suppresses energies adjacent to the harmonics corresponding to the F0 and the F1 of FFR. Along with an active afferent pathway, the afferent auditory nerve generates the electrical activity more precisely corresponding to the F0 and the F1 of the stimulus. This involves the release of neurotransmitter, thereby reducing the trans-membrane threshold and increase in neural firing. In poor hearing aid performers, though similar physiological activity was present, probably a lack of precision in neural activity due to less sensitive afferent and weak efferent auditory pathway, might have failed to provide higher energy at harmonics corresponding to F0 and F1 of each stimulus. Thus, it is can be inferred from the present study that subtle physiological variations might be present at the inferior colliculus of the auditory pathway in the poor hearing aid performers with reference to that in good hearing aid performers.

Conclusion

Though the hearing aid preserved inherent cues in speech syllables, an effect of annoyance towards noise alters the neural encoding at auditory brainstem level. It infers that acoustic cues transferred by hearing aid are successfully relied at auditory brainstem level but subtle physiological alterations were present at auditory brainstem in those individuals who are annoyed from those who are not by noise.

Implication

The study presents an evidence to use objective approaches to validate the hearing aid output at ear canal and at auditory brainstem level. Utilization of the real ear measurement for analyzing the hearing aid output in the ear canal will help in knowing the representation of inherent speech cues. Studying the encoding of amplified speech in individuals with hearing impairment with their annoyance level demonstrates a critical role of stimulus contingent response in the assessment of hearing aid algorithms. It solves some of the practical problems faced by the audiologists regarding setting of amplification parameters in providing the maximum usable information. Findings of the present study help the audiologist in counselling a hearing aid user regarding extent of benefit derived with best hearing aid prescribed.

  • Peer Review under the responsibility of Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial.
  • Please cite this article as: Shetty HN, Puttabasappa M. Encoding of speech sounds at auditory brainstem level in good and poor hearing aid performers. Braz J Otorhinolaryngol. 2017;83:512-22.

Acknowledgements

The authors would like to thank the Director and the HOD-Audiology, All India Institute of Speech and Hearing, for granting permission to carry out the study. The authors would also thank all the participants of the study for their co-operation.

References

  • 1
    Bentler RA, Niebuhr DP, Getta JP, Anderson CV. Longitudinal study of hearing aid effectiveness. II: Subjective measures. J Speech Hear Res. 1993;36:820-31.
  • 2
    Kochkin S. On the issue of value: hearing aid benefit, price, satisfaction, and brand repurchase rates. Hear Rev. 2003;10:12-26.
  • 3
    Nemes J. Despite benefits of outcomes measures, advocates say they're under used. Hear J. 2003;56:19-25.
  • 4
    Nabelek AK. Acceptance of background noise may be key to successful fittings. Hear J. 2005;58:10-5.
  • 5
    Nabelek AK, Tucker FM, Letowski TR. Toleration of background noises: relationship with patterns of hearing aid use by elderly persons. J Speech Hear Res. 1991;34:679-85.
  • 6
    Nabelek AK, Freyaldenhoven MC, Tampas JW, Burchfiel SB, Muenchen RA. Acceptable noise level as a predictor of hearing aid use. J Am Acad Audiol. 2006;17:626-39.
  • 7
    Freyaldenhoven MC, Smiley DF. Acceptance of background noise in children with normal hearing. J Educ Audiol. 2006;13:27-31.
  • 8
    von Hapsburg D, Bahng J. Acceptance of background noise levels in bilingual (Korean-English) listeners. J Am Acad Audiol. 2006;17:649-58.
  • 9
    Brannstrom KJ, Lantz J, Nielsen LH, Olsen SO. Acceptable noise level with Danish, Swedish, and non-semantic speech materials. Int J Audiol. 2012;51:146-56.
  • 10
    Plyler PN, Alworth LN, Rossini TP, Mapes KE. Effects of speech signal content and speaker gender on acceptance of noise in listeners with normal hearing. Int J Audiol. 2011;50:243-8.
  • 11
    Harkrider AW, Tampas JW. Differences in responses from the cochleae and central nervous systems of females with low versus high acceptable noise levels. J Am Acad Audiol. 2006;17:667-76.
  • 12
    Tampas JW, Harkrider AW, Hedrick MS. Neurophysiological indices of speech and nonspeech stimulus processing. J Speech Lang Hear Res. 2005;48:1147-64.
  • 13
    Souza PE, Tremblay KL. New perspectives on assessing amplification effects. Trends Amplif. 2006;10:119-43.
  • 14
    Mueller HG. Probe microphone measurement: unplugged. Hear Rev. 2005;48:10-36.
  • 15
    Cunningham DR, Lao-Davila RG, Eisenmenger BA, Lazich RW. Study finds use of Live Speech Mapping reduces follow-up visits and saves money. J Hear. 2002;55:43-6.
  • 16
    Stelmachowicz PG, Kopun J, Mace A, Lewis DE, Nittrouer S. The perception of amplified speech by listeners with hearing loss: acoustic correlates. J Acoust Soc Am. 1995;98:1388-99.
  • 17
    Souza PE, Tremblay KL. Combining acoustic, electrophysiological and behavioral measures of hearing aids. Scottsdale, AZ: American Auditory Society; 2005.
  • 18
    Kewley-Port D. Time-varying features as correlates of place of articulation in stop consonants. J Acoust Soc Am. 1983;73:322-35.
  • 19
    Souza PE, Tremblay KL, Davies-Venn E, Kalstein L. Explaining consonant errors using short-term audibility. Salt Lake City, UT: Am Acad Audiol; 2004.
  • 20
    Kraus N, Nicol T. Brainstem origins for cortical 'what' and 'where' pathways in the auditory system. Trends Neurosci. 2005;28:176-81.
  • 21
    Young ED, Sachs MB. Representation of steady-state vowels in the temporal aspects of the discharge patterns of populations of auditory-nerve fibers. J Acoust Soc Am. 1979;66:1381-403.
  • 22
    Worden FG, Marsh JT. Frequency-following (microphonic-like) neural responses evoked by sound. Electroencephalogr Clin Neurophysiol. 1968;25:42-52.
  • 23
    Russo N, Nicol T, Musacchia G, Kraus N. Brainstem responses to speech syllables. Clin Neurophysiol. 2004;115:2021-30.
  • 24
    Russo NM, Nicol TG, Zecker SG, Hayes EA, Kraus N. Auditory training improves neural timing in the human brainstem. Behav Brain Res. 2005;156:95-103.
  • 25
    Johnson KL, Nicol T, Zecker SG, Kraus N. Developmental plasticity in the human auditory brainstem. J Neurosci. 2008;28:4000-7.
  • 26
    Chandrasekaran B, Kraus N. The scalp-recorded brainstem response to speech: neural origins and plasticity. Psychophysiology. 2010;47:236-46.
  • 27
    Cunningham J, Nicol T, Zecker SG, Bradlow A, Kraus N. Neurobiologic responses to speech in noise in children with learning problems: deficits and strategies for improvement. Clin Neurophysiol. 2001;112:758-67.
  • 28
    Parbery-Clark A, Skoe E, Lam C, Kraus N. Musician enhancement for speech-in-noise. Ear Hear. 2009;30:653-61.
  • 29
    Hornickel J, Skoe E, Kraus N. Subcortical laterality of speech encoding. Audiol Neurootol. 2009;14:198-207.
  • 30
    Hemanth N, Manjula P. Representation of speech syllables at auditory brainstem. J Indian Speech Hear Assoc. 2012;26:1-13.
  • 31
    Pittman AL, Stelmachowicz PG. Hearing loss in children and adults: audiometric configuration, asymmetry, and progression. Ear Hear. 2003;24:198-205.
  • 32
    Sairam VVS, Manjula P. Long term average speech spectrum in Kannada. University of Mysore; 2002.
  • 33
    Skoe E, Nicol T, Kraus N. Phase coherence and the human brainstem response to consonant-vowel sounds. In: Association for research in otolaryngology symposium. 2008.
  • 34
    Dillon H, Katsch R, Byrne D, Ching T, Keidser G, Brewer S. The NAL-NL1 prescription procedure for non-linear hearing aids. National Acoustics Laboratories Research and Development, Sydney: National Acoustics Laboratories; 1998.
  • 35
    Iles MW. Speaker identification as a function of fundamental frequency and resonant frequencies. University of Florida; 1972.
  • 36
    Horii Y. Fundamental frequency perturbation observed in sustained phonation. J Speech Hear Res. 1979;22:5-19.
  • 37
    Abberton E. A laryngographic study of voice quality. University of London; 1976.
  • 38
    Ashmore JF. A fast motile response in guinea-pig outer hair cells: the cellular basis of the cochlear amplifier. J Physiol. 1987;388:323-47.
  • 39
    Knight RT. Distributed cortical network for visual attention. J Cogn Neurosci. 1997;9:75-91.
  • 40
    Krishnan A. Human frequency-following responses: representation of steady-state synthetic vowels. Hear Res. 2002;166:192-201.

Publication Dates

  • Publication in this collection
    Sep-Oct 2017

History

  • Received
    5 Feb 2016
  • Accepted
    20 June 2016
Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Sede da Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico Facial, Av. Indianópolia, 1287, 04063-002 São Paulo/SP Brasil, Tel.: (0xx11) 5053-7500, Fax: (0xx11) 5053-7512 - São Paulo - SP - Brazil
E-mail: revista@aborlccf.org.br