Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A Cochlear Implant Performance Prognostic Test Based on Electrical Field Interactions Evaluated by eABR (Electrical Auditory Brainstem Responses)

  • Nicolas Guevara ,

    guevara.n@chu-nice.fr

    Affiliation University Head and Neck Institute, CHU de Nice, 31 Avenue de Valombrose, 06107 Nice cedex 2, France

  • Michel Hoen,

    Affiliation Oticon Medical, Clinical and Scientific Research Department, 2720 chemin St Bernard, 06220 Vallauris, France

  • Eric Truy,

    Affiliation Department of Audiology and Otorhinolaryngology, Edouard Herriot Hospital, 5 Place d'Arsonval, 69437 Lyon, France

  • Stéphane Gallego

    Affiliation Institute for Readaptation Sciences and Techniques, 8 Avenue Rockefeller, 69373 Lyon Cedex 08, France

Abstract

Background

Cochlear implants (CIs) are neural prostheses that have been used routinely in the clinic over the past 25 years. They allow children who were born profoundly deaf, as well as adults affected by hearing loss for whom conventional hearing aids are insufficient, to attain a functional level of hearing. The “modern” CI (i.e., a multi-electrode implant using sequential coding strategies) has yielded good speech comprehension outcomes (recognition level for monosyllabic words about 50% to 60%, and sentence comprehension close to 90%). These good average results however hide a very important interindividual variability as scores in a given patients’ population often vary from 5 to 95% in comparable testing conditions. Our aim was to develop a prognostic model for patients with unilateral CI. A novel method of objectively measuring electrical and neuronal interactions using electrical auditory brainstem responses (eABRs) is proposed.

Methods and Findings

The method consists of two measurements: 1) eABR measurements with stimulation by a single electrode at 70% of the dynamic range (four electrodes distributed within the cochlea were tested), followed by a summation of these four eABRs; 2) Measurement of a single eABR with stimulation from all four electrodes at 70% of the dynamic range. A comparison of the eABRs obtained by these two measurements, defined as the monaural interaction component (MIC), indicated electrical and neural interactions between the stimulation channels.

Speech recognition performance without lip reading was measured for each patient using a logatome test (64 "vowel-consonant-vowel"; VCV; by forced choice of 1 out of 16). eABRs were measured in 16 CI patients (CIs with 20 electrodes, Digisonic SP; Oticon Medical ®, Vallauris, France).

Significant correlations were found between speech recognition performance and the ratio of the amplitude of the V wave of the eABRs obtained with the two measurements (Pearson's linear regression model, parametric correlation: r2 = 0.26, p<0.05).

Conclusions

This prognostic model allowed a substantial amount of the interindividual variance in speech recognition scores to be explained. The present study used measurements of electrical and neuronal interactions by eABR to assess patients' bio-electric capacity to use multiple information channels supplied by the implant. This type of prognostic information may be valuable in several ways. On the patient level, it allows customizing of individual treatments.

ClinicalTrials.gov Identifier: NCT01805167

Introduction

Cochlear implants (CI) represent a bona fide 20th century medical revolution. Used in clinical practice since the 1990s, this technology has allowed functional hearing to be restored in individuals with severe to profound hearing loss [1,2]. It has furthermore transformed the lives of many born-deaf children, providing them access to the world of oral communication and allowing them to develop intellectual and educational abilities often comparable to those of children with normal hearing [3]. CIs function by delivering direct electrical stimulations to the auditory nerve (cranial nerve VIII) using multi-electrode arrays that are surgically implanted in the cochlea. The “modern” CI (i.e., a multi-electrode implant using sequential coding strategies) [1] has yielded good speech comprehension outcomes. The recognition level for monosyllabic words is approximately 50% to 60%, and sentence comprehension is close to 90% (after 12 month use, without lip reading) [4]. These good average results however hide a very important interindividual variability as scores in a given patients’ population often vary from 5 to 95% in comparable testing conditions. In their 2004 study for example, Firszt and colleagues observed perception scores ranging from approximately 2 to 87% (average of 42%) for monosyllabic words in quiet [5]. This large interindividual variability can certainly be explained by many different factors identified at several levels of the auditory and speech processing chain. Identified candidates range among others, from developmental factors or age at implantation [6,7], to the quality of the electrode-neuron interface obtained after implantation [8], up to cognitive [9] or psycho-social aspects [10]. Improving our understanding of these factors and developing new, individualized approaches for using this neural prosthesis is mandatory to reduce outcome variability and optimize individual benefits. In this context, some parameters of the electrical stimulation delivered through CIs such as the number of stimulating electrodes, the speed or the frequency of stimulation, and the type of electrical stimulation used, can somewhat be customized for each patient. In practice however, customization often remains limited, due to difficulties in understanding and diagnosing the integration of the signal at the individual level.

The existence of electrical interactions between individual channels inside the cochlea [11,12] could explain important amounts variation in CI outcomes. Moreover if interactions could be assessed more efficiently, this could lead to interesting individual fine-tuning options. These interactions are responsible for an important loss in efficiency of CIs as systems today offer electrode carriers with up to 22 electrodes, when it was shown that individual patients in the end often benefit from only reduced numbers of independent spectral channels, down to four to eight [13, 14]. Although this could have a limited impact on speech recognition in silent backgrounds [15], it could become a crucial issue for the perception of speech in noise [13,16] or music [17]. The potential origins of channel interactions can be manifold and their effects could be of two opposite consequences, often difficult to disentangle. Residual polarization effects will have facilitating effects on consecutive stimulations, while refractory effects will tend to diminish subsequent neural responses (see [18] for a recent review on interaction mechanisms). CI channel interactions were studied since the introduction of the multichannel CI. Early studies quantified channel interactions using psychophysical methods evaluating loudness summation or forward masking effects across channels [1921]. Results from these works were at the origin of the introduction of sequential stimulation strategies, avoiding to stimulate spatially adjacent electrodes at the same time. This decreased the strong interactions created by simultaneous stimulation of electrodes observed in early versions of multichannel CIs [1]. More recent studies systematically characterized channel interactions using the forward masking psychophysical paradigm in different CI models [22] or the spatial spread of excitation obtained using different pulse shapes or stimulation modes [2327]. Even if they happen to be powerful research tools for comparing spatial spreading induced by different pulse shapes, grounding schemes or stimulation modes, such psychophysical methods for evaluating channel interactions are difficult to apply in clinical settings. These approaches are often very time-consuming and rely on the feedback of patients, which cannot always be obtained in a detailed and reliable fashion. Moreover, psychophysical methods as spatial tuning curves were shown to be relatively poor predictors of speech perception outcomes in CI [28,29], although some studies reported significant correlations with speech perception scores, see for example [30].

Faster, more reproducible and potentially more clinically-relevant evaluations may be obtained using electrophysiological evaluations of channel interactions. Early studies demonstrated the feasibility of evaluating the spatial spread of neural excitation using electrically evoked compound action potentials (ECAPs) and showed that this measure could provide robust predictions of perceptual dimensions (C and T-levels) in CI users [31]. Further experiments also suggested that an ECAP derived channel separation index could be correlated with pitch ranking [32]. ECAPs were further used to study different stimulation strategies in CIs [3335] or to obtain detailed spectral resolution maps along electrode arrays [36]. To our knowledge, few works directly correlated ECAP derived interaction measures with speech perception outcomes. Won and colleagues (2014) reported evidence that vowel perception could be predicted by the global spectral resolution of entire electrode arrays estimated through ECAP measures but did not observe any effect at the single-electrode level [37]. This is limiting potential clinical applications as efficiently limiting channel interactions requires to modify parameters of individual electrodes. ECAPs thus provided up to now limited predictions of speech perception outcomes.

Because they constitute very peripheral measures, ECAPs may provide a good estimation of the quality of local electrode-neuron interfaces but can only poorly predict more integrated perceptual mechanisms reflecting activity in further structures of the auditory system, beyond the cochlea. Along this line, two recent papers reported evidence that combining peripheral measures via ECAPs to more central measures using the acoustic change complex (ACC) could provide more reliable estimates of spectral selectivity, significantly correlated with speech perception [38,39]. The contribution of central auditory pathways can be studied electrophysiologically by auditory evoked potentials, with an early component (known as the electrically auditory brainstem response–eABR) and a late component (corresponding to the late cortical responses). We have previously shown that the relationship between electrophysiological and speech recognition variables is more pronounced when early (brainstem) rather than late (cortical) evoked responses are considered [4042].

To date, no quantitative method based on eABRs evaluating electrical interactions was described. Former studies reported the measurement of higher-level binaural interaction (i.e., at the level of the superior olivary complex, the core of the lateral lemniscus and the inferior colliculus) in patients with bilateral CI. These measures were performed by subtracting the eABR obtained after bilateral stimulation from the sum of the two eABRs obtained in a monaural fashion [43]. The objective was to measure the interaction between the right side and the left side. This methodology, called the binaural interaction component (BIC), has seen renewed success given the number of patients with bilateral implants. Nonetheless, due to the length of the tests and the lack of real clinical usefulness, this technique is rarely applied outside of research programs [44]. The aim of the current experiment was to test a novel eABR-based method of measuring interactions and to correlate the results of this testing with speech recognition performances of CI patients. Similar to the BIC, our measure is based on a comparison of eABRs, although it is obtained using a single implant. It assesses the electrical and neural interactions occurring with unilateral stimulation. By analogy, we therefore defined this test as the monaural interaction component (MIC).

Materials and Methods

Participants

Sixteen adults (11 female; aged between 19 and 81 years) who had undergone CI surgery (Digisonic SP; Oticon Medical, Vallauris, France) participated in this study. All were native speakers of French with post-lingual deafness onset and had used the implant for at least 10 months prior to the experiment. No specific inclusion criteria were defined, particularly in terms of etiology or duration of deafness and the time since implantation. This approach provided us access to a wide range of word recognition performances. Demographic data are presented in Table 1. This study and all experimental procedures were approved by the Institutional Review Board at the University Hospital in Nice. Each participant provided written consent and the study and all experimental procedures were approved by the South Mediterranean V Ethical Review Board (Comité de Protection des Personnes Sud Méditérrannée V, authorization n°12,075).

The Digisonic SP CI system

The Digisonic SP is a transcutaneous multi-electrode CI system that uses the main peak interleaved sampling (MPIS) coding strategy [45]. The implanted part of the device is composed of a receptor/transducer and an electrode array that is 25 mm long, with 20 active platinum-iridium electrodes. The stimulation in this system employs a hybrid grounding scheme resulting from the simultaneous use of two types of ground electrodes. One ground electrode is positioned outside of the cochlea, as in a typical monopolar scheme, and all non-active electrodes on the electrode array are placed into a ground mode, as in a common-ground configuration. This leads to the coexistence of two possible current return paths, one intra- and one extra-cochlear. The balance between the two current return paths is modulated by their respective impedances. The monopolar pathway is favored by high impedances. As a result, the intra-cochlear pathway will be favored only for electrodes very close to the stimulation electrode. Another particularity of this system is related to the shape of the electrical pulse used for stimulation. In this device, electrical pulses have a single active phase (anodic), the current equalization being realized passively by capacitive coupling. At the time of the experiment, all patients wore the same sound processor, incorporating a post-spectral decomposition logarithmic output compression function [46].

Measurement of speech recognition performance without lip reading

Speech recognition was assessed using a vowel-consonant-vowel (VCV), 16-alternatives forced-choice logatome test ran in quiet. VCVs were extracted from standardized lists recorded on a CD (Listes de logatomes de Dodelé, Collège National d'Audioprothèse, Paris, France) and embedded in an ad-hoc developed Matlab® (The MathWorks, Natick, Mass., United-States) program. Only consonant comprehension was tested, to both increase the difficulty of the test and increase the matching between the speech perception evaluation and the frequency areas tested by the eABR (from 1,107 Hz to 4,883 Hz). All patients were tested four times on each 16 stimuli, with a random selection, for a total of 64 trials. Stimuli were delivered through a loudspeaker placed approximately one meter in front of the patient. This test was performed in a sound treated room, at a comfortable sound level that was individually determined for each patient in order to account for the important interindividual variability in dynamic range observed in CI patients. Keeping clinical mapping settings constant, and fixing the output level to an arbitrary value could have led to very different percepts in different CI patients, equalization of listening comfort was therefore preferred. Presentation levels ranged from 65 dB SPL to 75 dB SPL (detailed in Table 1). A 1/16 comprehension level (6.25%) corresponded to random recognition.

Electrical Auditory Brainstem Response (eABR) measurements

EABRs were recorded using surface electrodes that were placed on the forehead (+), the lower part of the chin (-), and the contralateral mastoid (common). Each testing session lasted three to four hours, including breaks. During the recordings, participants were encouraged to relax during the recording sessions to limit artifacts due to muscle activity.

The full-scale range used for eABR recording was set to 100 μV using the ABR recording system Centor USB (Racia-Alvar®). Responses were filtered using a wideband (1.6–3.6 kHz) analog bandpass filter. The averaging involved 1,500 sweeps. To estimate and confirm reproducibility, eABRs were recorded twice, reported data correspond to the first recording obtained for each participant. The recording duration was 10 milliseconds per sweep. The stimuli used to record the eABRs were generated by a PC using a portable Neurelec-Digistim® stimulator that was attached through a USB port. Stimuli were asymmetric charge-balanced biphasic pulses (phase amplitude: 0.9 mA, total pulse duration: from 20 μs to 250 μs, as a function of loudness perception, pulse-rate: 47 Hz, pulse onset asynchrony adjusted depending on pulse duration). Stimuli were delivered to four different electrodes distributed along the cochlea (electrode numbers: 4, 7, 10, and 13, from base to apex) at a unique, individually adjusted stimulation level. Tested electrodes were selected because they correlated with the primary area required for word comprehension. Indeed the articulation index tells us that the frequencies between 1 and 3 kHz contribute most to speech intelligibility [47]. Electrodes 4, 7, 10, and 13 are situated approximately at 4.6, 8.2, 11.8, and 15.4 mm from the base of the cochlea (estimated from the electrode-array length of 25 mm; electrode length: 0.5 mm; inter-electrode distance: 0.7 mm). In acoustic terms these electrodes code for the frequency bands: 4,120–4,883 Hz, 2,409–2,799 Hz, 1,497–1,758 Hz, and 1,107–1,237 Hz, respectively (default parameters of frequency-mapping in the Digisonic SP CI system). Given that the threshold (T) and maximal comfort (M) levels for the electrical stimulation can vary considerably across individual patients and electrodes, these levels were specified as a percentage of the electrical dynamic range of the tested electrode, rather than an absolute intensity. For each electrode tested, we calculated the difference between T- and M- Levels and set the stimulation level to the 70% value of this electrical dynamic range. This ensured robust and reliable eABR recordings, maximizing the chance of obtaining well-defined wave-V responses, without the risk of reaching the discomfort threshold during the stimulation of the four electrodes in a same-pulse-train.

Two types of measurements were performed: i) Individual stimulations were obtained by recording four different eABRs from each single electrode among the four test electrodes (4, 7, 10, and 13). The sum of these four eABRs was subsequently calculated. ii) Multi-electrode stimulations were obtained by recording one eABR trace using a sequential pulse- train stimulation from the same four electrodes. Using the MPIS strategy, the delay between the first pulse occurring at electrode number 4 and the last pulse of the pulse train at electrode number 13 is 84 μs.

eABR analyses and assessment of the monaural interaction component (MIC)

For each participant, the eABR recordings obtained from the four individual stimulations were summed, leading to one eABR trace per participant in this recording condition and one second eABR trace in the multi-electrode stimulation condition. We then derived the MIC, mathematically defined as the ratio between the peak amplitudes of the wave-V components obtained from the summed individual and multi-electrode stimulations. The amplitude peak of wave-V was selected both for a practical and theoretical issue. In practice, wave-V is more robustly observed in patients than wave-III, which can be sometimes harder to identify. In theory, wave-V is generated higher in the auditory system (inferior colliculus) than wave-III is (superior olivary complex) and may represent a better candidate to observe a correlation with integrated auditory processes such as speech perception. Several works could indeed report that the inferior colliculus would represent sparse speech cues [48,49].

We hypothesized that the MIC would represent an approximation of the interaction between the concerned electrodes:

  • In the theoretical scenario where there would be no electrical and/or neural interactions, we would assume that the neural response triggered by the stimulation at each individual electrode would be completely independent from each other in both recording modes, individual or multi-electrode. In this case, the sum of the four eABRs obtained by individual stimulations should be very similar to the eABR obtained with the multi-electrode stimulation.
  • On the opposite, in the theoretical case where electrical and/or neural interactions would be total, the eABR obtained from the multi-electrode stimulation would actually tend to be equal to the eABR obtained from a pseudo-monoelectrode implant. Upon multi-electrode stimulation, the eABR obtained would tend to be equal to the eABR obtained from a single electrode. The sum of the four eABRs obtained from individual stimulations should then be four times larger than the eABR obtained from the multi-electrode stimulation.

Results

Good quality eABRs were obtained in most cases, with a clear wave V, while wave III could only be identified in 13/16 cases. To better explain the protocol and serve as an example, Fig 1 shows and illustration of the low- (top-left) and high (top-right) interaction cases and the eABR recordings obtained from two individual patients using our method. Fig 1 bottom-left, the patient appears to show little interaction. The sum of the eABRs for individual stimulations was similar to the multi-electrode stimulation eABR and the observed MIC was close to 1. On the right side of Fig 1, the other patient showed a higher level of cross-channel interaction. The sum of the eABRs for individual stimulations was much higher than that of the multi-electrode stimulation eABR, leading to a MIC value of 2.6. Stimulation of the same four electrodes in the same pulse-train generated an important overlap. These first observations suggested that the MIC could be a good indicator of channel interaction in CI patients.

thumbnail
Fig 1. Schematic representation of the electrical interaction assessment method by eABR.

Left: low-interaction case. The sum of the four eABR measures obtained from individual stimulations (red) equals the eABR obtained with the multi-electrode stimulation (blue), the MIC value tends to 1. Right: high-interaction case: the eABR amplitude obtained from the multi-electrode stimulation tends to equal that of one eABR measure in the individual recording, the MIC tends to rise and constitutes a metric of the electrical interaction.

https://doi.org/10.1371/journal.pone.0155008.g001

We then analyzed the main characteristics of waves III and V for the 16 patients, these observations are summarized in Table 2. eABR traces obtained for the multi- and individual-electrode recordings were averaged across participants in order to generate grand-average traces. As can be seen on Fig 2, the grand-average eABR reveals differences between the components obtained from the sum of individual-electrodes stimulations and from the multi-electrode stimulations. The first main difference concerned latencies of wave-III and wave-V components, which both appeared earlier in the multi-electrode condition. Both peak-latencies were significantly different, respectively: 1.93 ms +/- 0.18 vs. 2.08 ms +/- 0.12, difference = 0.154 ms (t (12) = -2.90, p = 0.013) for wave-III and 3.63 ms +/- 21 vs. 3.83 ms +/- 0.21, difference = 0.200 ms (t (15) = -3.56, p = 0.003) for wave-V. As expected, wave III could only be characterized precisely in 13 out of the 16 included participants, the amplitude analyses were therefore restricted to wave-V. The amplitudes of wave-V observed for both recording conditions were also significantly different, the average amplitude being lower in the multi-electrode condition: 3.01 μV +/- 0.18 than in the individual electrode condition 4.41 μV +/- 0.26 (t (15) = -3.48, p = 0.003). These results show that, on average, the multi-electrode stimulation does not generate an ABR response equivalent to the sum of four individual stimulations, presumably reflecting the effect of across channel interactions. These interactions are characterized by a shorter latency and diminished amplitude in the multi-electrode stimulation condition.

thumbnail
Fig 2. Grand-averaged eABRs obtained in the multi- and individual-electrode conditions.

Red: grand-averaged eABR (N = 16) obtained by summing the traces from the individual-electrodes recordings. Blue: grand-averaged eABR (N = 16) obtained from the same pulse-train, multi-electrodes stimulation. III: peak of wave-III. V: peak of wave-V.

https://doi.org/10.1371/journal.pone.0155008.g002

In order to further characterize the effects of interactions as evaluated by our ABR recording paradigm, we calculated differences between the two recording conditions (sum of individual vs. multi-electrode) for each participant for the latencies of wave-III and wave-V. For the amplitudes, we quantified by the MIC calculated as the ratio of wave-V amplitudes. Pearson’s parametric correlations were used to test for a relationship between VCV scores and these different parameters. The results of this analysis are displayed in Table 3. All factors led to non-significant correlations with speech scores, except for MIC (n = 16, r2 = 0.26, p<0.05). Data reflecting the relationship between individual MIC values and VCV perception scores are shown in Fig 3. In order to evaluate if a simple mathematical model could better fit data, we modeled the relationship using a decay exponential with the following criteria: i) MIC values constrained between 4: maximal interaction and 1: no interaction; and a chance-level of 6.25% (1/16) in the VCV perception test. A decreasing exponential non-linear regression was performed with Microsoft® Excel 2000/XLSTAT©-Pro (Version 4.07, 2013, Addinsoft, Inc., Brooklyn, NY, USA), which matched the model Eq (1): (1)

In (1), MIC being the MIC value and VCV score the percentage of correct responses obtained with the speech perception test. The resulting function was mapped on the individual data as shown in Fig 3. The fact that the data are well explained by an exponential decay points to an interesting property of our dataset. As the speech score increases from low to medium values, (in our experiment for participants 10-CM with 12.5% to 6-MJ with 37.5% recognition score (triangles in Fig 3) the MIC value decreases exponentially from high values around 3.5 to around 1.5. This is suggesting that the participants with very low speech score (<40% VCV recognition score closed-set in quiet) in general have high MIC values while patients with higher speech recognition score (>47.5% in our sample) have low MIC values. This difference remained however non-significant in our study as reflected by a non-significant p-value for a two independent heteroscedastic samples t-test (t(14) = 1.80, p = 0.09). Given the size of the two compared sample (n = 8) this observation still suggests that MIC values rapidly decrease with the augmentation of speech perception scores.

thumbnail
Fig 3. Prognostic model of the speech recognition performance as a function of the electrical interaction measured by EABR.

Individually calculated MIC values compared to VCV scores. Triangles: the eight low speech performing patients (VCV score < 40%). Circles the eight best performing patients (VCV scores > 40%). Orange curve: the decreasing exponential non-linear regression matching the data and following equation: MIC = 1 + 2.8e−0.04(VCVscore−6.25).

https://doi.org/10.1371/journal.pone.0155008.g003

thumbnail
Table 3. Difference analyses and correlations with speech-scores.

https://doi.org/10.1371/journal.pone.0155008.t003

Finally, and in order to evaluate the effect of the possible confounding factor of the level of presentation used in the VCV test, which was individually adjusted for each subject in order to account for individual variations in comfort levels across CI patients, we also calculated the correlation between scores at the VCV test expressed in % and the presentation level expressed in dB SPL. This correlation remained non-significant (n = 16, r2 = 0.16, p = 0.12). Finally, both factors were entered into a multiple-regression analysis in order to estimate their respective contributions to speech perception scores. This observation suggested that the VCV scores obtained in the present experiment were not significantly influenced by the individually adjusted comfortable level of presentation of the VCV stimuli.

Discussion

The substantial variability in functional results with CIs, combined with a lack of reliable prognostic factors, limit the individual optimization of this neural prosthesis. We have developed a prognostic test that uses eABRs to measure global electrical and neural interactions at the level of the electrode array. It assesses the biological capacity of the patients to use several information channels supplied by the implant and can explain a substantial amount of the interindividual variability found in speech perception scores among CI patients. We also observed that the speech perception data were well explained by an exponential decrease of the electrical interactions as evaluated by the MIC. Given the number of participants included in the present study it was however not possible to reach significance when studying the difference between low- and high- speech proficient participants, this limit of the current project suggests the interest of running a large multicentric study involving a larger population of CI patients in order to confirm and extend the present observations. Another limitation of the present approach concerns the fact that, for the clinical relevance of the approach, participants were tested using a clinical processor in clinical settings, imposing some compromises. For example, we decided to keep the mapping parameters of the patients and run the tests with these parameters, thereby preventing us from harmonizing the level of presentation of the speech stimuli across participants and choosing to ask participants to indicate a comfortable level of presentation. The range of presentation levels was in the end relatively restricted, between 65 and 75 dB SPL, but this could have introduced a bias in the speech perception measures. Further studies will be dedicated to reproducing the present observations with a larger population to confirm the present observations. A caveat associated to the MIC approach as introduced in the present article is associated to the simultaneous measure of ABR responses from multiple electrodes. In the hypothetical case where the amplitude of wave V on one of the electrodes would completely dominate both the sum of the single electrode measures as well as the multi-electrode measure, we would observe a small MIC not necessarily reflecting low channel interactions. Future developments will therefore concentrate on developing measures based in individual electrode measures. Nevertheless, despite these compromises, we were able to identify a significant relationship between an eABR derived estimation of channel interactions, the MIC, and speech perception outcomes in CI users. Moreover, the MIC measure showed some interesting distribution properties. Patients with high MIC values had poor speech perception levels, suggesting that high electrical interactions are detrimental to speech perception. However, low electrical interactions as evaluated by the MIC seem to be a necessary but not sufficient condition to observe medium to high speech perception scores, as participants with low MIC values >1.5, showed a large range of speech perception scores, approximately from 23 to 75%. This result suggests that other dimensions also play a crucial role in observing good speech perception outcomes with CIs, the number of independent spectral channels is important but other factors come into play. Numerous predictive factors for post-CI outcomes were described in former studies, including large multicentric analyses of prognostic factors in adults, performed in 1996 and updated in 2013 [50,51]. These analyses confirmed the importance of the following four main factors: the duration of severe deafness prior to placement of the CI, the age at onset of severe deafness, the etiology of the deafness, and the duration of CI use. In 1996, a model defined by these four factors accounted for 21% of interindividual variance in CI outcomes. However, with the widening of CI indications, and improvements in signal processing allowing always more participant to reach higher perception scores, the overall CI performance improved as did the variance of outcomes, causing these initial prognostic factors to explain only 10% of the variance in 2003. Despite the introduction of nine new significant factors in this predictive model (including the duration of moderate hearing loss and the use of hearing aids), only 22% of the variance could ultimately be explained. These disparities in post-CI outcomes are likely influenced strongly by the physiology of the auditory nerve, as well as auditory pathways in the brainstem and auditory cortex of the wearer of the implant, up to the influence of cognitive factors. These components are highly variable in terms of functional integrity and capacity to adapt (functional plasticity), this variability may at least partly explain the differences in results that persist to date among individuals who have received CIs.

Our results strongly suggest that, at least for patients with poor speech perception, a significant amount of their speech perception variability can be explained by cross-channel electrical interactions.

Channel interactions can be measured using different approaches, including electrical field imaging (EFI), evaluating interactions at the physical level, ECAPs, allowing an estimation of physiological interactions and psychophysics measures evaluating the effect of interactions at the perceptual level, see [18] or [36] for reviews. In 2011, Tang, Benitez and Zeng have explored interactions at the three levels, combining the three approaches in the same CI patients. Comparing the interaction estimation obtained at the three levels revealed that interindividual variability increased from physical measures with EFI to physiological and behavioral measures, suggesting that higher-order interactions as revealed by ECAPs or behavioral measures were better reflecting individual variability in CI outcomes than electrical field imaging, and that reducing these higher-order interactions could be helpful to reduce variability in CI outcomes. In the present work, we introduced a novel approach to evaluating higher-order interactions by using an electrophysiological measure (amplitude of the eABR wave-V component), which we believed would be closer to behavioral variability as physical EFI or ECAP measurements would be. Due to the complexity of recording eABRs, the majority of electrophysiological studies regarding electrode-neuron interactions have used ECAP measurements, which are fast and easy to perform. However, ECAP measurements have several limitations. This measurement of the auditory nerve response to electrical stimulation is recorded by another intracochlear electrode. It therefore only reflects the cochlear nerve response (and only a part of it). Additionally, ECAP is affected by its close interface with the intracochlear recording electrode. eABRs instead allow the measurement of the progressive auditory brainstem elaboration of the auditory information, leading to complex percepts as speech for example. In particular, using the amplitude wave-V component reveals interactions observed at the level of the inferior colliculus, a neural relay that was shown to contain elaborated auditory representations relevant to speech perception [4042, 40, 49]. As a result, if ECAP measures can be used to predict low-level percepts (thresholds or comfort levels) [31], their application to predict speech outcomes was up to now limited [37]. As suggested by the results from the present study, eABRs may reflect speech perception scores more accurately. Further studies will be dedicated to extend this observation to the individual electrode level in order to improve the clinical relevance of the approach.

Four different neural interaction phenomena were described, which may be characterized as temporal, that participate to cross-channel interactions in CI: refractoriness, facilitation, accommodation and spike-rate adaptation [18]. The two latter ones are mostly related to temporal integration factors which are certainly minored in the paradigm used in the present experiment using a low-rate stimulation (47Hz) and a sequential stimulation on distant electrodes. The first two ones could however be observed, especially in the multi-electrode same pulse-train stimulation condition. The fact that this condition was associated to a significant decrease in the latency of the wave-III and V components suggests a facilitating effect due to residual polarization consecutive to the short delay pulses. A decrease in latency of the eABR components is usually associated with an increase in the amplitude of the stimulation [52]. However, the fact that this decrease in latency was in our study accompanied by a decrease in amplitude relative to the sum of the wave-V amplitude observed with individual recordings suggests that our method also identified interactions due to refractory effects between channels. The reduced number of usable spectral channels is due to diffusion of the signal from the stimulated electrodes and the resultant interactions with the overlapping neural population. ECAP measurements allow observation of these overlaps [34], but they do not allow an analysis of how these peripheral interactions are integrated and processed in the first auditory relays. In addition, there is variability among the electrode-neuron interface within a cochlea, and these variations probably explain the lack of correlation between ECAPs and performance as measured by a VCV test, which is related to the treatment of multiple electrodes activation throughout the cochlea. Auditory performance can therefore be predicted in part by the characteristics of the eABRs. Nonetheless, due to the complexity of the neural processes involved speech processing, many other factors (including more central ones) might contribute to the variability in CI outcomes.

Outlook of the methods

Our eABR-based assessment of interactions could be used to develop novel ways of improving CI mapping in a more individualized way. It is likely that an actual decrease in neuronal interactions allows optimizing the central coding of auditory information provided by the implant. Nonetheless, before identifying ways to decrease the interactions, methods to measure interactions that correlate with speech recognition performance are necessary.

For any given patient, interactions beyond the cochlea (and the electrode-neuron interface) are linked to intrinsic factors (e.g., the number of residual spiral ganglion cells and the temporal synchrony of neuronal discharges). Currently, there is no easy way to directly modulate the nervous system (e.g., with the use of growth factors). However, it may be possible to act on information coding strategies sent by the implant. Techniques to focus and direct the electrical current using different modes of stimulation (“current steering and current focusing”) have been proposed to reduce its diffusion [53]. Tripolar stimulations were shown to improve speech in noise perception compared to monopolar stimulations [54] (but see [55] for inconsistent results) and current focusing can provide better spectral resolution [56,57]. However, other limitations (e.g., power consumption related to threshold increase and the use of multiple current sources), as well as difficulties in adapting proper programming to the right patient have limited the routine use of these stimulation modes. Another option would be to alter the algorithm for selecting the “N-of-M” peaks (peak-picking algorithm); deactivating specific channels according to the measured interactions could also potentially be of benefit, as already shown by Noble and colleagues using a CT-derived imaging method [58] or as implemented in the PACE/MP3000 strategy, which selects channels based on psychoacoustic simultaneous masking trying to minimize channel interaction [59]. Lastly, decreasing the stimulation frequency for patients in whom many interactions have been measured might help to limit these interactions. Although a decrease in transmitted information may result, speech recognition performance might also improve.

Acknowledgments

This work was supported by the Nice University Hospital.

Author Contributions

Conceived and designed the experiments: NG SG. Performed the experiments: NG. Analyzed the data: NG MH SG ET. Contributed reagents/materials/analysis tools: NG SG. Wrote the paper: NG MH SG ET.

References

  1. 1. Wilson BS, Finley CC, Lawson DT, Wolford RD, Eddington DK, Rabinowitz WM. Better speech recognition with cochlear implants. Nature. 1991 Jul 18;352(6332):236–8. pmid:1857418
  2. 2. Gates GA, Mills JH. Presbycusis. Lancet. 2005 Sep 24–30;366(9491):1111–20. pmid:16182900
  3. 3. Svirsky MA, Robbins AM, Kirk KI, Pisoni DB, Miyamoto RT. Language development in profoundly deaf children with cochlear implants. Psychol Sci. 2000 Mar;11(2):153–8. pmid:11273423
  4. 4. Wilson BS, Dorman MF. Cochlear implants: a remarkable past and a brilliant future. Hear Res. 2008 Aug;242(1–2):3–21. pmid:18616994
  5. 5. Firszt JB, Holden LK, Skinner MW, Tobey EA, Peterson A, Gaggl W, et al. Recognition of speech presented at soft to loud levels by adult cochlear implant recipients of three cochlear implant systems. Ear Hear. 2004 Aug;25(4):375–87. pmid:15292777
  6. 6. Holden LK, Finley CC, Firszt JB, Holden TA, Brenner C, Potts LG, et al. Factors affecting open-set word recognition in adults with cochlear implants. Ear Hear. 2013 May-Jun;34(3):342–60. pmid:23348845
  7. 7. Boons T, Brokx JP, Dhooge I, Frijns JH, Peeraer L, Vermeulen A, et al. Predictors of spoken language development following pediatric cochlear implantation. Ear Hear. 2012 Sep-Oct;33(5):617–39. pmid:22555184
  8. 8. Lenarz T, Pau HW, Paasche G. Cochlear implants. Curr Pharm Biotechnol. 2013;14(1):112–23. pmid:23092261
  9. 9. Harris MS, Kronenberger WG, Gao S, Hoen HM, Miyamoto RT, Pisoni DB. Verbal short -term memory development and spoken language outcomes in deaf children with cochlear implants. Ear Hear. 2013 Mar-Apr;34(2):179–92. pmid:23000801
  10. 10. Holt RF, Beer J, Kronenberger WG, Pisoni DB, Lalonde K. Contribution of family environment to pediatric cochlear implant users' speech and language outcomes: some preliminary findings. J Speech Lang Hear Res. 2012 Jun;55(3):848–64. pmid:22232387
  11. 11. White MW, Merzenich MM, Gardi JN. Multichannel cochlear implants. Channel interactions and processor design. Arch Otolaryngol. 1984 Aug;110(8):493–501. pmid:6547597
  12. 12. Bierer JA, Middlebrooks JC. Cortical responses to cochlear implant stimulation: channel interactions. J Assoc Res Otolaryngol. 2004 Mar;5(1):32–48. Epub 2003 Oct 20. pmid:14564662
  13. 13. Friesen LM, Shannon RV, Baskent D, Wang X. Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. J Acoust Soc Am. 2001 Aug;110(2):1150–63. pmid:11519582
  14. 14. Cooper WB, Tobey E, Loizou PC. Music perception by cochlear implant and normal hearing listeners as measured by the Montreal Battery for Evaluation of Amusia. Ear Hear. 2008 Aug;29(4):618–26. pmid:18469714
  15. 15. Shannon RV, Zeng FG, Kamath V, Wygonski J, Ekelid M. Speech recognition with primarily temporal cues. Science. 1995 Oct 13;270(5234):303–4. pmid:7569981
  16. 16. Shannon RV, Fu QJ, Galvin J 3rd. The number of spectral channels required for speech recognition depends on the difficulty of the listening situation. Acta Otolaryngol Suppl. 2004 May;(552):50–4. pmid:15219048
  17. 17. Spahr AJ, Litvak LM, Dorman MF, Bohanan AR, Mishra LN. Simulating the effects of spread of electric excitation on musical tuning and melody identification with a cochlear implant. J Speech Lang Hear Res. 2008 Dec;51(6):1599–606. pmid:18664681
  18. 18. Boulet J, White M, Bruce IC. Temporal Considerations for Stimulating Spiral Ganglion Neurons with Cochlear Implants. J Assoc Res Otolaryngol. 2015 Oct 26.
  19. 19. Shannon RV. Multichannel electrical stimulation of the auditory nerve in man. II. Channel interaction. Hear Res. 1983 Oct;12(1):1–16. pmid:6689326
  20. 20. Tong YC, Blamey PJ, Dowell RC, Clark GM. Psychophysical studies evaluating the feasibility of a speech processing strategy for a multiple-channel cochlear implant. J Acoust Soc Am. 1983 Jul;74(1):73–80. pmid:6688434
  21. 21. Tong YC, Clark GM. Loudness summation, masking, and temporal interaction for sensations produced by electric stimulation of two sites in the human cochlea. J Acoust Soc Am. 1986 Jun;79(6):1958–66. pmid:3722606
  22. 22. Cohen LT, Lenarz T, Battmer RD, Bender von Saebelkampf C, Busby PA, Cowan RS. A psychophysical forward masking comparison of longitudinal spread of neural excitation in the Contour and straight Nucleus electrode arrays. Int J Audiol. 2005 Oct;44(10):559–66. pmid:16315446
  23. 23. Bierer JA, Faulkner KF. Identifying cochlear implant channels with poor electrode-neuron interface: partial tripolar, single-channel thresholds and psychophysical tuning curves. Ear Hear. 2010 Apr;31(2):247–58 pmid:20090533
  24. 24. Landsberger DM, Padilla M, Srinivasan AG. Reducing current spread using current focusing in cochlear implant users. Hear Res. 2012 Feb;284(1–2):16–24. pmid:22230370
  25. 25. Undurraga JA, Carlyon RP, Macherey O, Wouters J, van Wieringen A. Spread of excitation varies for different electrical pulse shapes and stimulation modes in cochlear implants. Hear Res. 2012 Aug;290(1–2):21–36. pmid:22583921
  26. 26. Fielden CA, Kluk K, McKay CM. Place specificity of monopolar and tripolar stimuli in cochlear implants: the influence of residual masking. J Acoust Soc Am. 2013 Jun;133(6):4109–23. pmid:23742363
  27. 27. Marozeau J, McDermott HJ, Swanson BA, McKay CM. Perceptual interactions between electrodes using focused and monopolar cochlear stimulation. J Assoc Res Otolaryngol. 2015 Jun;16(3):401–12. pmid:25742726
  28. 28. Anderson ES, Nelson DA, Kreft H, Nelson PB, Oxenham AJ. Comparing spatial tuning curves, spectral ripple resolution, and speech perception in cochlear implant users. J Acoust Soc Am. 2011 Jul;130(1):364–75. pmid:21786905
  29. 29. Azadpour M, McKay CM. A psychophysical method for measuring spatial resolution in cochlear implants. J Assoc Res Otolaryngol. 2012 Feb;13(1):145–57. pmid:22002609
  30. 30. Boëx C, Kós MI, Pelizzone M. Forward masking in different cochlear implant systems. J Acoust Soc Am. 2003 Oct;114(4 Pt 1):2058–65. pmid:14587605
  31. 31. Cohen LT, Richardson LM, Saunders E, Cowan RS. Spatial spread of neural excitation in cochlear implant recipients: comparison of improved ECAP method and psychophysical forward masking. Hear Res. 2003 May;179(1–2):72–87. pmid:12742240
  32. 32. Hughes ML. A re-evaluation of the relation between physiological channel interaction and electrode pitch ranking in cochlear implants. J Acoust Soc Am. 2008 Nov;124(5):2711–4. pmid:19045758
  33. 33. Busby PA, Battmer RD, Pesch J. Electrophysiological spread of excitation and pitch perception for dual and single electrodes using the Nucleus Freedom cochlear implant. Ear Hear. 2008 Dec;29(6):853–64. pmid:18633324
  34. 34. Hughes ML, Stille LJ, Baudhuin JL, Goehring JL. ECAP spread of excitation with virtual channels and physical electrodes. Hear Res. 2013 Dec;306:93–103. pmid:24095669
  35. 35. Goehring JL, Neff DL, Baudhuin JL, Hughes ML. Pitch ranking, electrode discrimination, and physiological spread of excitation using current steering in cochlear implants. J Acoust Soc Am. 2014 Dec;136(6):3159. pmid:25480063
  36. 36. Tang Q, Benítez R, Zeng FG. Spatial channel interactions in cochlear implants. J Neural Eng. 2011 Aug;8(4):046029. pmid:21750370
  37. 37. Won JH, Humphrey EL, Yeager KR, Martinez AA, Robinson CH, Mills KE, et al. Relationship among the physiologic channel interactions, spectral-ripple discrimination, and vowel identification in cochlear implant users. J Acoust Soc Am. 2014 Nov;136(5):2714–25. pmid:25373971
  38. 38. Scheperle RA, Abbas PJ. Peripheral and Central Contributions to Cortical Responses in Cochlear Implant Users. Ear Hear. 2015 Jul-Aug;36(4):430–40. pmid:25658747
  39. 39. Scheperle RA, Abbas PJ. Relationships Among Peripheral and Central Electrophysiological Measures of Spatial and Spectral Selectivity and Speech Perception in Cochlear Implant Users. Ear Hear. 2015 Jul-Aug;36(4):441–53. pmid:25658746
  40. 40. Maurer J, Collet L, Pelster H, Truy E, Gallégo S. Auditory late cortical response and speech recognition in Digisonic cochlear implant users. Laryngoscope. 2002 Dec;112(12):2220–4. pmid:12461344
  41. 41. Guiraud J, Gallego S, Arnold L, Boyle P, Truy E, Collet L. Effects of auditory pathway anatomy and deafness characteristics? (1): On electrically evoked auditory brainstem responses. Hear Res. 2007 Jan;223(1–2):48–60. pmid:17157463
  42. 42. Guiraud J, Gallego S, Arnold L, Boyle P, Truy E, Collet L. Effects of auditory pathway anatomy and deafness characteristics? Part 2: On electrically evoked late auditory responses. Hear Res. 2007 Jun;228(1–2):44–57. pmid:17350776
  43. 43. Pelizzone M, Kasper A, Montandon P. Binaural interaction in a cochlear implant patient. Hear Res. 1990 Oct;48(3):287–90. pmid:2272938
  44. 44. Van Yper LN, Vermeire K, De Vel EF, Battmer RD, Dhooge IJ. Binaural interaction in the auditory brainstem response: a normative study. Clin Neurophysiol. 2015 Apr;126(4):772–9. pmid:25240247
  45. 45. Di Lella F, Bacciu A, Pasanisi E, Vincenti V, Guida M, Bacciu S. Main peak interleaved sampling (MPIS) strategy: effect of stimulation rate variations on speech perception in adult cochlear implant recipients using the Digisonic SP cochlear implant. Acta Otolaryngol. 2010;130(1):102–7. pmid:19424919
  46. 46. Lorenzi C, Sibellas J, Füllgrabe C, Gallégo S, Fugain C, Meyer B. Effects of amplitude compression on first- and second-order modulation detection thresholds in cochlear implant listeners. Int J Audiol. 2004 May;43(5):264–70. pmid:15357409
  47. 47. Amlani AM, Punch JL, Ching TY. Methods and applications of the audibility index in hearing aid selection and fitting. Trends Amplif. 2002 Sep;6(3):81–129. pmid:25425917
  48. 48. Warrier CM, Abrams DA, Nicol TG, Kraus N. Inferior colliculus contributions to phase encoding of stop consonants in an animal model. Hear Res. 2011 Dec;282(1–2):108–18. pmid:21945200
  49. 49. Garcia-Lazaro JA, Belliveau LA, Lesica NA. Independent population coding of speech with sub-millisecond precision. J Neurosci. 2013 Dec 4;33(49):19362–72. pmid:24305831
  50. 50. Blamey P, Artieres F, Başkent D, Bergeron F, Beynon A, Burke E, et al. Factors affecting auditory performance of postlinguistically deaf adults using cochlear implants: an update with 2251 patients. Audiol Neurootol. 2013;18(1):36–47. pmid:23095305
  51. 51. Blamey P, Arndt P, Bergeron F, Bredberg G, Brimacombe J, Facer G, et al. Factors affecting auditory performance of postlinguistically deaf adults using cochlear implants. Audiol Neurootol. 1996 Sep-Oct;1(5):293–306. pmid:9390810
  52. 52. Gallégo S, Garnier S, Micheyl C, Truy E, Morgon A, Collet L. Loudness growth functions and EABR characteristics in Digisonic cochlear implantees. Acta Otolaryngol. 1999 Mar;119(2):234–8. pmid:10320083
  53. 53. Bonham BH, Litvak LM. Current focusing and steering: modeling, physiology, and psychophysics. Hear Res. 2008 Aug;242(1–2):141–53. pmid:18501539
  54. 54. Srinivasan AG, Padilla M, Shannon RV, Landsberger DM. Improving speech perception in noise with current focusing in cochlear implant users. Ear Hear. 2013 May;299:29–36.
  55. 55. Mens LH, Berenstein CK. Speech perception with mono- and quadrupolar electrode configurations: a crossover study. Otol Neurotol. 2005 Sep;26(5):957–964. pmid:16151343
  56. 56. Berenstein CK, Mens LH, Mulder JJ, Vanpoucke FJ. Current steering and current focusing in cochlear implants: comparison of monopolar, tripolar, and virtual channel electrode configurations. Ear Hear. 2008 Apr;29(2):250–60. pmid:18595189
  57. 57. Smith ZM, Parkinson WS, Long CJ. Multipolar current focusing increases spectral resolution in cochlear implants. Conf Proc IEEE Eng Med Biol Soc. 2013;2013:2796–9. pmid:24110308
  58. 58. Noble JH, Gifford RH, Hedley-Williams AJ, Dawant BM, Labadie RF. Clinical evaluation of an image-guided cochlear implant programming strategy. Audiol Neurootol. 2014;19(6):400–11. pmid:25402603
  59. 59. Nogueira W, Büchner A, Lenarz T, Edler B. A psychoacoustic "NofM"-type speech coding strategy for cochlear implants. EURASIP J. Appl. Signal Process. 2005; Jan:3044–3059.