ORIGINAL RESEARCH article
Sec. Auditory Cognitive Neuroscience
Volume 14 - 2020 | https://doi.org/10.3389/fnins.2020.00787
Combined Brain-Perfusion SPECT and EEG Measurements Suggest Distinct Strategies for Speech Comprehension in CI Users With Higher and Lower Performance
- 1Department of Nuclear Medicine, Hannover Medical School, Hanover, Germany
- 2Cluster of Excellence Hearing4all, Hannover Medical School, University of Oldenburg, Oldenburg, Germany
- 3Department of Otorhinolaryngology, Hannover Medical School, Hanover, Germany
- 4Department of Otorhinolaryngology, University of Cologne, Cologne, Germany
- 5Department of Medical Physics and Radiation Protection, Hannover Medical School, Hanover, Germany
- 6Department of Otorhinolaryngology, Faculty of Medicine Carl Gustav Carus, Saxonian Cochlear Implant Center, Technical University Dresden, Dresden, Germany
Cochlear implantation constitutes a successful therapy of inner ear deafness, with the majority of patients showing good outcomes. There is, however, still some unexplained variability in outcomes with a number of cochlear-implant (CI) users, showing major limitations in speech comprehension. The current study used a multimodal diagnostic approach combining single-photon emission computed tomography (SPECT) and electroencephalography (EEG) to examine the mechanisms underlying speech processing in postlingually deafened CI users (N = 21). In one session, the participants performed a speech discrimination task, during which a 96-channel EEG was recorded and the perfusions marker 99mTc-HMPAO was injected intravenously. The SPECT scan was acquired 1.5 h after injection to measure the cortical activity during the speech task. The second session included a SPECT scan after injection without stimulation at rest. Analysis of EEG and SPECT data showed N400 and P600 event-related potentials (ERPs) particularly evoked by semantic violations in the sentences, and enhanced perfusion in a temporo-frontal network during task compared to rest, involving the auditory cortex bilaterally and Broca’s area. Moreover, higher performance in testing for word recognition and verbal intelligence strongly correlated to the activation in this network during the speech task. However, comparing CI users with lower and higher speech intelligibility [median split with cutoff + 7.6 dB signal-to-noise ratio (SNR) in the Göttinger sentence test] revealed for CI users with higher performance additional activations of parietal and occipital regions and for those with lower performance stronger activation of superior frontal areas. Furthermore, SPECT activity was tightly coupled with EEG and cognitive abilities, as indicated by correlations between (1) cortical activation and the amplitudes in EEG, N400 (temporal and occipital areas)/P600 (parietal and occipital areas) and (2) between cortical activation in left-sided temporal and bilateral occipital/parietal areas and working memory capacity. These results suggest the recruitment of a temporo-frontal network in CI users during speech processing and a close connection between ERP effects and cortical activation in CI users. The observed differences in speech-evoked cortical activation patterns for CI users with higher and lower speech intelligibility suggest distinct processing strategies during speech rehabilitation with CI.
Cochlear implantation is an established and effective method of treating sensorineural hearing loss (Wilson and Dorman, 2008a, b; Gaylor et al., 2013). Cochlear implants (CIs) bypass the damaged structures of the inner ear by electrical stimulation of the auditory nerve (Wilson and Dorman, 2008a, b). Although cochlear implantation allows open-set speech perception in most of the cases, there is a high variability in CI outcomes (Heydebrand et al., 2007). This variability cannot be completely explained so far (Lazard et al., 2012; Blamey et al., 2013) but seems to be at least partially related to individual differences in the auditory nerve, the position of the implant electrodes, cognitive abilities, and neuronal plasticity (Nadol, 1997; Drennan and Rubinstein, 2008; Lazard et al., 2012; Rönnberg et al., 2013; Sandmann et al., 2015; Finke et al., 2016a). Neuroimaging can help improve the understanding of the individual differences in speech comprehension by providing important insights into the sensory and cognitive processes underlying speech perception in CI users.
Previous studies using electroencephalography (EEG) and event-related potentials (ERPs) in particular have shown differences in cortical speech processing between CI users and normal-hearing (NH) listeners, both at initial sensory and at later higher-level cognitive processing stages (e.g., Hahne et al., 2012; Finke et al., 2016a). In particular, CI users have shown smaller amplitudes of N1 ERPs to speech sounds, indicating smaller assembly or reduced synchronization of activated neurons in the auditory cortex of CI users when compared with NH listeners (Groenen et al., 2001; Aggarwal and Green, 2012). Regarding the later cognitive processing stages, ERPs in response to semantic anomalies (N400) and syntactic violations (P600) have been rarely examined in CI users (Hahne et al., 2012; Henkin et al., 2014; Kallioinen et al., 2016; Vavatzanidis et al., 2018). The N400, reflecting semantic memory use during language comprehension (Kutas and Federmeier, 2000), has been shown to be prolonged in adult CI users when compared with NH listeners (Hahne et al., 2012; Henkin et al., 2014), suggesting a delayed and a more effortful speech processing with the limited CI input (Finke et al., 2016a). However, it is currently unknown whether the N400 can distinguish between CI users who have good versus poor speech recognition, although such a distinctiveness has been previously shown for other auditory ERPs (Soshi et al., 2014; Turgeon et al., 2014).
Positron emission tomography (PET) and single-photon emission computed tomography (SPECT) enable the precise spatial assignment of neuronal activity that underlies speech perception in CI users (Giraud et al., 2001a,b,c; Wong et al., 2002). In general, the spatial resolution of both PET and SPECT enables investigations of neuronal activity (changes) in the auditory cortex and associated brain regions (Abraham and Feng, 2011). Regarding CI users, SPECT has been shown previously to be a suitable tool to objectively evaluate speech comprehension performance (Allen et al., 2004). In particular, different cortical activations were observed for higher and lower CI performers during speech comprehension (Tobey et al., 2004). SPECT has also been suggested to be considered for the presurgical evaluation of prelingually deaf adults being candidates for cochlear implantation (Di Nardo et al., 2013). Furthermore, previous PET studies with CI users have revealed a positive correlation between speech recognition ability and activation in the primary and association auditory cortices (Green et al., 2005). They have also shown a different network recruited for speech processing in proficient and non-proficient CI users, suggesting that activation in both the temporal cortices and the left inferior prefrontal cortex are a prerequisite for successful speech comprehension (Mortensen et al., 2006). On the other hand, in NH listeners, a stronger activation of inferior frontal regions has been related to enhanced listening effort (Davis and Johnsrude, 2003) and to poorer cognitive abilities, in particular lower working memory capacity (Zekveld et al., 2012). Although previous results point to a remarkable influence of cognitive abilities on speech recognition with a CI (Heydebrand et al., 2007), no study so far has examined how individual differences in cognitive abilities and listening effort relate to the different cortical response patterns in proficient and non-proficient CI users.
The combination of SPECT/PET and EEG measurements allows a synergistic examination of speech processing, as it provides the excellent temporal resolution of the EEG and the good spatial resolution of the emission tomography. A few studies so far have used combined sequential SPECT or PET and EEG measurements to study auditory processing in different groups of patients, in particular in patients with mild and moderate Alzheimer’s disease (O’Mahony et al., 1996; Gungor et al., 2005), Schizophrenia (Blackwood et al., 1994; Shajahan et al., 1997; Medved et al., 2001) or obsessive–compulsive disorders (Molina et al., 1995). The results have revealed correlations between ERPs and regional cerebral perfusion, suggesting a connection between disease-related alterations in auditory ERPs and cortical activation (O’Mahony et al., 1996; Gungor et al., 2005).
The principal aim of this study was to contribute to the better understanding of the high variability in CI outcomes. We used, for the first time, a synchronized multimodal SPECT-ERP approach in CI users to thoroughly examine the neuronal activation patterns underlying speech comprehension and their relation to cognitive abilities. The study also aimed to prove the suitability of a typical EEG paradigm for SPECT imaging. CI users with higher and lower speech performance were tested with a semantic-anomaly paradigm to study the N400 ERP in response to sentences with semantic violations (Kutas and Hillyard, 1980a; Lau et al., 2008). It has been previously shown for NH listeners that the left posterior middle temporal gyrus (MTG) is critically involved in N400 generation (for a review, see Lau et al., 2008). Given this finding and previous PET results about speech processing in CI users (Green et al., 2005; Mortensen et al., 2006), we predicted positive correlations between the N400 response and activation in the MTG. We also expected differences in cortical activation in the (pre)frontal, the superior temporal, and the posterior middle temporal regions, i.e., in temporo-frontal networks, between CI users with higher and lower speech recognition ability (Friederici, 2012). In addition, we expect that CI users with higher speech comprehension have stronger activation in regions and networks representing cognitive functions, such as the temporal cortex (memory) and parietal cortex (attention) (Coez et al., 2014). Finally, we tested the hypothesis that cross-modal (i.e., auditory) activation of the occipital (i.e., visual) cortex is beneficial in terms of speech understanding with a CI (Giraud et al., 2001b). Indeed, our results suggest a close connection between ERP effects and cortical activation in CI users and different activation patterns during speech processing between higher and lower performers, pointing to different neural resource allocation and strategies used for speech processing.
Materials and Methods
In this section, the following methodological issues are described in subchapters: the patient characteristics (Patients), the sequence of procedures (Sequence of Procedures), the audiometric and cognitive tests (Audiometric and Neurocognitive Testing), the applied EEG/SPECT paradigm (Speech Condition Stimuli for Combined EEG-SPECT Measurement), and details on EEG (EEG Recording) and SPECT (SPECT – Acquisition and Reconstruction) acquisition. For the latter two methods, Data Analysis gives details on data analysis.
Twenty-one postlingually deafened CI users [mean age, 62.1 years; standard deviation (SD), 11.5 years; range, 30–80 years; 10 female) participated in the present study, with 18 CI users being consistent right-handers and three being consistent left-handers (Annett, 1970). Eight CI users were implanted unilaterally (5 left), and 13 were implanted bilaterally. In case of bilateral implantation, the “better” ear, according to the performance in the Freiburg monosyllabic word test (Hahlbrock, 1970), was used for stimulation (5 left). All CI users were native German speakers, had at least 11 months of CI experience (mean, 99.4 months; SD, 68.5; range, 11.0–346.0 months) and achieved a word recognition score of at least 20% in the Hochmair–Schulz–Moser (HSM) sentence test in quiet (Hochmair-Desoyer et al., 1997) with the tested CI. None of the CI users reported using sign language for communication. Details concerning the subject’s implant system and demographics can be obtained in Table 1. None of the CI users reported neurological or psychiatric disorders or used medications affecting the central nervous system.
All participants gave informed written consent before the experiment. The study was approved by the Ethics Committee of the Hannover Medical School (vote number 6678) and the German Federal Office for Radiation Protection (reference number Z5-22461/2-2014-012) and was carried out in accordance with the Declaration of Helsinki.
Sequence of Procedures
In this subchapter, the sequence of procedures is described. Details on the procedures are given in the following subchapters. The participants underwent two individual sessions, separated by 13.0 ± 6.5 days. In one session, CI users completed the neuropsychological testing, except for the size-comparison span test (SICSPAN) (Sorqvist et al., 2010). Subsequently, a SPECT scan was performed after application of 729.1 ± 8.1 MBq Technetium-99m (99mTc) labeled HMPAO without stimulation (“rest condition”). Injection of the substance took place in a quiet room with dimmed light, where participants stayed for 15 min before and 5 min after application for uptake. The SPECT scan itself was performed ∼1.5 h postinjection (p.i.). In the other session, participants underwent the audiometric testing and the SICSPAN (Sorqvist et al., 2010), which was followed by combined sequential EEG and SPECT measurements (“speech condition”). The participants were seated comfortably in a dimly lit, as well as electrically and acoustically shielded cabin, 100 cm in front of a computer screen. Before the actual start of the EEG experiment, participants performed a training block with seven sentences. Subsequently, the participants listened to 80 different sentences (40 semantically correct/40 semantically incorrect), which were presented within the course of two experimental blocks, with 55 sentences being presented in block 1 (∼7 min) and 25 sentences presented in block 2 (∼3 min). The length of the respective blocks was adapted so that – without interrupting the paradigm – application (2 min after the start of block 1) and subsequent radiopharmaceutical uptake phase (for another 5 min) was possible. The order of the sentences was pseudo-randomized between participants. The CI users were instructed in written form to listen to each sentence while focusing on a black screen. A white fixation point, appearing 1,000 ms after the offset of each sentence, signaled the participants to provide a response via a button press on whether the sentence was semantically correct or not. Assignment of the buttons to the two answer possibilities was counterbalanced across participants. The fixation point remained on the screen for 3,000 ms, which constituted the response window. The delayed response window ensured decoupling the N400 and P600 ERPs from the motor response. After the task, subjects were asked to evaluate the subjective listening effort during the task with a 5-point rating scale (1.0 = not demanding, 5.0 = too demanding; the words could not be understood). During the first block, 2 min after the start of the task, 731.5 ± 6.8 MBq 99mTc-labeled HMPAO was applied intravenously via medical tubing from outside the shielded cabin. Approximately 1.5 h after injection, a SPECT scan was acquired, reflecting cortical activity during the sentence discrimination task. In general, the “rest condition” was performed first, and the “speech condition” (combined sequential SPECT and EEG measurements) took place in the second session. Due to organizational issues, however, the “speech condition”’ was carried out first in a few cases (n = 4).
Audiometric and Neurocognitive Testing
Speech recognition abilities obtained with the CI used in the experimental session were assessed using three frequently applied German speech tests: (1) the Freiburg monosyllabic word test in quiet (Hahlbrock, 1970), (2) the HSM sentence test (Hochmair-Desoyer et al., 1997) in quiet and in noise [10 dB signal-to-noise ratio (SNR)], and (3) the Göttinger sentence test (GÖSA; adaptive noise; Kollmeier and Wesselkamp, 1997). The GÖSA is a widely used audiometric test, which contains complete German sentences that reflect the everyday speech situation. It uses an adaptive procedure to measure the signal-to-noise ratio at which 50% of the speech signal is correctly understood. In the current study, the speech material of all of the three speech tests was presented at a sound intensity of 65 dB SPL. Participants were instructed to report all words perceived.
Our study aimed to compare brain activation patterns between CI users with different performance levels. However, the present study did not enable to compare markedly poor and good performers, since all CI users had to at least be able to perform the sentence discrimination task to a certain extent, allowing for an adequate number of correct EEG trials for analysis. Groups, therefore, rather represent CI users with higher or lower performance levels. Accordingly, the group assignment was based on a median split procedure, for which we relied on the GÖSA, resulting in a cutoff of +7.6 dB SNR. This procedure was not based on previous studies, but rather exploratory, with the aim of obtaining groups of CI users with different performance levels. In the following, the groups are referred to as “lower” (50% speech reception threshold at > 7.6 dB SNR; note that more positive values indicate worse performance) and “higher” CI performers (50% speech reception threshold at < 7.6 dB SNR). In the present study, the GÖSA was chosen to be the most appropriate one for the group selection based on the median split procedure, for the following reasons: (1) It contains meaningful sentences from everyday life (ecological validity of stimulus material), (2) it allows the measurement of an individual speech reception threshold by means of an adaptive procedure (speech test with high accuracy), and (3) it is highly demanding and provides performance scores with fair variability in performance scores (no problem of floor and ceiling effects). Thus, with regards to group assignment, the GÖSA test is preferable to the Freiburg monosyllabic word test and the HSM sentence test, given that the GÖSA test uses a more appropriate stimulus material (some of the words in the Freiburg monosyllabic word test are outdated), and the test results are not confounded by floor and ceiling effects (non-adaptive speech tests, for instance the HSM sentence test with a fixed SNR, provide a risk for these boundary effects).
To control for residual hearing, the contralateral device was detached at the time of testing and the ear was closed by means of an earplug. Beforehand, to assess residual hearing in the non-tested contralateral ear, a pure-tone audiometry (unaided; range, 0.25–8 kHz) was performed.
Beside the audiometric tests, participants completed four different cognitive tests, assessing working memory capacity and verbal abilities: (1) The “Mehrfachwahl-Wortschatz-Intelligenz-Test” (MWT-B, Lehrl, 1977) was applied to measure verbal intelligence. Here, participants had to identify a real word among four pseudowords. According to the official guidelines provided with the test material, individual percentiles (in relation to a normative sample) were used for the statistical analyses. (2) The lexical verbal fluency subtest of the “Regensburger Wortflüssigkeits-Test” (RWT, Aschenbrenner et al., 2001; Harth et al., 2004) was used to test verbal fluency. During this test, subjects were asked to report as many words as possible with the initial letter “s” within 2 min. Here, likewise, individual percentiles in relation to a normative sample provided with the test manual were used for the statistical analyses. (3) A German version of the SICSPAN (Sorqvist et al., 2010; Finke et al., 2016a) was used to assess the verbal working-memory capacity. The SICSPAN was analyzed using the total percentage of correctly remembered words. (4) We used two subtests (verbal learning, verbal recall) of the CERAD-Plus test battery (Memory Clinic Basel)1 to assess verbal abilities. Z-scores in relation to age-specific normative data were used for statistical analysis.
Speech Condition Stimuli for Combined EEG-SPECT Measurement
The stimulus material consisted of 87 sentences in German language, constructed out of 6 words each (determinative, subject, the auxiliary “hat/haben”/“has/have,” determinative, object, past participle). The sentences were clearly pronounced by a female speaker, spoken at a moderate pace (213 ± 92 ms between the words). The sentences’ final word was either semantically correct (e.g., “Die Mutter hat den Kuchen gebacken”/“The mother has baked the cake”) or incorrect (e.g., “Der Junge hat das Radio gebadet”/“The boy has bathed the radio”) with regards to the previous sentential context. All critical sentence final words appeared in a correct as well as in an a semantically incorrect sentence, thereby guaranteeing that the integration of the word in the semantic context is responsible for different ERP effects rather than the word itself. All sentences were spoken by a trained female native German speaker. Audio files had a sampling frequency of 44 kHz with a 32-bit resolution. Sentence duration ranged from 4.01 to 5.85 s. The onset of each final word was carefully identified by auditory and visual inspection to ensure an accurate time locking of the N400 and P600 ERPs in response to the final word. Stimuli were delivered using the Presentation software (version 16.5; Neurobehavioral Systems, Inc., Berkeley, CA, United States) running on a personal computer. Sentences were presented via two loudspeakers (HECO victa 301) located at 50° azimuth. In the case of a second CI or a conventional hearing aid at the contralateral side, the device was detached for the duration of the experiment, and the respective ear was closed with a wax earplug. In total, 10 CI users were stimulated on the left and 11 on the right side. Similar to previous studies (e.g., Sandmann et al., 2015; Schierholz et al., 2017), the participants used a 7-point loudness-rating scale, which allowed adjusting the perceived loudness of the sentences to a comfortable level, equivalent to 60–70 dB (Allen et al., 1990; Zeng, 1994).
EEG data were recorded using 94 Ag/AgCl electrodes, integrated in an infracerebral electrode cap with an equidistant electrode layout (Easycap, Herrsching, Germany). To record an electrooculogram, two additional electrodes were placed below the two eyes. The reference electrode was positioned on the nose tip. A midline electrode, placed anterior to the frontocentral scalp region (AFz), served as ground. Data were recorded by means of three linked 32-channel BrainAmp amplifiers (BrainProducts, Gilching, Germany), with a sampling rate of 1,000 Hz and an online analog filter from 0.02 to 250 Hz. For data acquisition, electrode impedances were kept below 10 kΩ.
SPECT – Acquisition and Reconstruction
For the scan, participants were positioned as comfortably as possible on the patient bed, and their head was carefully fixed with a special headband with Velcro straps. The participants were instructed to avoid head movements during the scan. Acquisition was performed using a dual-head SPECT camera (Discovery 670 NM/CT, GE Healthcare, Haifa, Israel) equipped with low-energy high-resolution (LEHR) parallel-hole collimators. In total, 180 projections, that is, 90 projections for each of the two detectors, were acquired using a step and shot mode with circular orbit (rotation around the head of the patient with the smallest possible distance, normally 15 cm). With this setup, typically, a total number of counts in the order of 8–9 million could be achieved per acquisition. The required projection time was individually determined before starting the scan on the basis of the count rate detected with the patients’ head in the camera field of view. Typically, the count rate was between 1.4 and 1.6 kCts, and the according total recording time was about 55 min. Projections were acquired with a 128 × 128 matrix size and a zoom factor of 2.0 (pixel size, 2.23 × 2.23 mm2). The quality of unprocessed projection data was assessed visually in cine mode and in the form of sinograms, e.g., with respect to motion artifacts, immediately after the recording. Data were reconstructed iteratively, using an ordered-subset expectation maximization (OSEM) algorithm with 5 iterations, 10 subsets, and a Butterworth filter with a cutoff frequency of 0.55 cycles/cm, power of 10 (Hudson and Larkin, 1994), and a dual window scatter correction (scaling 1.1) (Jaszczak et al., 1984) including attenuation correction according to Chang (threshold for boundary detection of 5% and attenuation coefficient of 0.11/cm) (Chang, 1978).
EEG data were preprocessed using custom scripts in MATLAB 22.214.171.1246344 (R2017a; Mathworks, Natick, MA) and EEGLAB (version 13.6.5b, Delorme and Makeig, 2004). Raw data were imported, down-sampled to 500 Hz, and low-pass filtered (40 Hz) using a Hann-windowed zero-phase finite impulse response (FIR) filter implemented in EEGLAB (pop firws.m; Widmann and Schröger, 2012). Electrodes covering the CI speech processor as well as the transmitter coil were omitted for recording and accordingly removed for the analysis. Subsequently, the continuous data were segmented into 2-s segments and pruned for unique, non-stereotype artifacts. The remaining data were high-pass filtered (1 Hz; Hann-windowed FIR filter; pop_firws.m) and subjected to an extended infomax-independent component analysis (ICA, Bell and Sejnowski, 1995). The resulting ICA weights were applied to the raw data that were filtered (0.1–30 Hz; Hann-windowed FIR filter; pop_firws.m) and epoched (-250–7,950 ms) relative to sentence onset. The prestimulus interval (-250–0 ms) was used for baseline correction. ICA components representing eye blinks, horizontal eye movements, heartbeat activity, and CI artifacts were identified and removed (mean, 36.7%; SD = 14.5%; Jung et al., 2000a, b; Debener et al., 2008). Regarding the latter, we identified the components representing the CI artifact by the centroid on the side of the implanted device and by the pedestal artifact in the time course of the respective component (Sandmann et al., 2009, 2010, 2015). Missing channels were interpolated using a spherical spline (mean, 5.8; SD, 1.6; range, 2–10 electrodes). Additional triggers were set, marking the onset of the final word (correct, semantic violation) in each sentence. Based on these triggers, additional epochs, time locked to the onset of the final words (−250–950 ms) were created. Data were corrected using the time interval of −250–0 ms relative to the onset of the critical word.
EEG Data Analysis
Single-subject ERPs were computed to the onset of the sentences by averaging over all correctly categorized trials, irrespective of the condition (correct, semantic violation) of the sentence (ERPonset). Additionally, ERPs to the onset of the final word were computed for each participant, separately for semantically correct sentences (ERPcritCorr) and sentences with a semantic violation (ERPcritViol), including only the correctly identified trials (critCon: mean, 87.3%; SD, 6.6%; critViol: mean, 87.5%; SD, 7.3%). Furthermore, a difference wave was computed relative to the onset of the final word (ERPcritDiff = ERPcritViol - ERPcritCorr). The single-subject ERPs to sentence onset were analyzed using a fronto-central region-of-interest (ROI), including seven electrodes around FCz (see Figure 1A), and a time window of the auditory N1 and P2 (N1, 80–200 ms; P2, 160–280 ms), determined by visual inspection of the grand average ERP and based on previous studies with CI users (see, e.g., Finke et al., 2015, 2016b). Regarding the final word onset, the N400 and the P600 ERPs were analyzed by considering a centroparietal ROI around CPz for both the N400 (time window, 300–900 ms; seven electrodes) and the P600 (time window, 750–940 ms; seven electrodes, see Figure 1B). Both time windows were defined by visual inspection of the grand average ERPs and based on previous studies with CI users (Hahne et al., 2012). For the quantification of the evoked responses, we determined the local minimum (N1, N400) or the local maximum (P2, P600), respectively, of the ERP amplitudes in the respective time window and the respective ROI (peakdet.m)2. The mean amplitude was computed for ± 10 ms around the local minimum/maximum. The latency of the respective peaks was determined by detecting the time of the local peak minimum (N1, N400) or maximum (P2, P600). Amplitude and latency measures were subjected to correlation analyses with the SPECT data. Moreover, ERPs were compared between a group of higher and a group of lower performing CI users by means of independent t-tests, separately for each ERP components.
Figure 1. Grand averages of the event-related potentials (ERPs), once for all subjects (N = 21; solid lines) and once separately for the groups of higher (N = 11; dotted lines) and lower (N = 10; dashed lines) cochlear-implant (CI) performers. (A) Shows the grand average ERPs in response to the sentence onset for a centrofrontal electrode region of interest (ROI). A clear negative (∼150 ms) and a clear positive (∼240 ms) deflection, referred to as the N1 and P2, respectively, can be observed for all three ERPs. (B) Shows the average difference waves (ERPcritViol - ERPcritCorr), time locked to the onset of the critical words, using a centroparietal electrode ROI. A negative (N400) and positive deflection (P600) can be observed for all three ERP waves around 650 ms. At the bottom of each panel, the topographical voltage maps are displayed for the time of the respective peak latencies of the different components [N1, P2 (A), N400 and P600 (B), separately for the different average ERPs (all, higher performers, lower performers).
SPECT Data Analysis
SPECT images were analyzed using the statistical parametric mapping software (SPM8, Wellcome Trust Center for Neuroimaging, Institute of Neurology, University College London, London, United Kingdom), running within MATLAB 126.96.36.1996344 (2017a; Mathworks, Natick, MA, United States). First baseline (“rest condition”) and stimulation (“speech condition”) images of each patient were realigned and transformed into a standard stereotaxic anatomical space according to the Montreal Neurological Institute (MNI) employing the default brain perfusion SPECT template provided in SPM8. Further preprocessing included scaling of the images before statistical testing. In order to compare the speech and the rest condition, images were scaled to the 75th percentile (Buchert et al., 2006). Then, effects of stimulus presentation were assessed using a paired t-test. For further group comparisons and correlations to speech audiometry and EEG, difference images were generated based on procedures included in subtraction ictal SPECT coregistered to MRI (SISCOM) analysis (Huberfeld et al., 2006; Apostolova et al., 2008), in particular a two-step scaling procedure. First, images were scaled to the global average, using a gray matter mask excluding the cerebellar voxels. Then, preliminary difference images (speech condition minus rest condition) were created. Proceeding from these, a mean value of voxels with a low difference between speech condition and rest condition (i.e., <2 times the standard deviation of the mean) was calculated. This mean value was used for rescaling the speech condition study, which avoids an impact of voxels from activated areas on scaling. Thereafter, final difference images were calculated by subtracting rest condition images scaled to global average from rescaled speech condition images. These final difference images (speech condition minus rest condition) were used for further group comparisons and correlations. Moreover, for the assessment of group difference and correlations based on baseline images (rest condition), these images were scaled to the 75th percentile.
For image-based statistics, SPM8 was used as well. First of all, results of statistical tests presented here were generated using a significance level of p < 0.001 (uncorrected for multiple comparisons) for inferences. The threshold was chosen with respect to previous studies, in particular studies of central auditory processing via auditory implants, where this threshold has been successfully employed (Giraud et al., 2001c; Coez et al., 2009; Berding et al., 2015; Mamach et al., 2018). Furthermore, it has been proposed as a good compromise compensating for the limited sensitivity of brain perfusion SPECT, whereby it is still protecting from false positive results (Signorini et al., 1999). Results are listed without extent voxel threshold (k = 0) in the tables. For displaying results, however, two different voxel thresholds were employed. This was done in order to account for different magnitudes in cluster sizes observed across test results.
For test results with a relatively small size of the largest cluster, an extent voxel threshold of k = 19 was used. Corresponding results are displayed using the so-called glass brain visualization. For all other test results including relatively large cluster sizes, an operational extent voxel threshold of k = 50 was used. Corresponding results are presented using surface rendered MRI image in MNI space. In general, the “modern design” option provided by SPM was employed to include some information on the depth of localized activations. Moreover, locations of significant differences were spatially assigned by automated anatomical labeling, specifically by overlaying the statistical parametric map with a Brodmann volume of interest (VOI) atlas (Rorden and Brett, 2000; Tzourio-Mazoyer et al., 2002). Additionally, statistical analyses were performed including correction for multiple comparisons based on the family-wise error (FWE) rate procedure, together with a cutoff of p < 0.05 and an extent threshold of k = 0 for statistical inferences (Flandin and Friston, 2019). However, this is a quit conservative approach for the correction for multiple comparison, and nevertheless, analyses without correction for multiple comparison as described above can be regarded as justified particularly in the context of preexisting according a priori hypotheses.
To assess potential effects of handedness, paired t-tests comparing speech and rest condition were performed twice, once including and once excluding the three left-handed participants. As no differences were obvious in the resulting statistical parametric maps, all left-handers were included in the further analyses. Another potentially influencing factor is the side of stimulation (right = 11; left = 10). Therefore, an additional comparison of speech and rest condition was performed, with images from patients with left-ear stimulation being flipped in the mid-sagittal plane. Only minor differences for flipped vs. non-flipped images were observed in primary and secondary auditory cortices. Therefore, the original, non-flipped images were used for all other analyses in order to avoid confusion with primarily unilateral components of brain networks related to speech processing (like, e.g., Broca’s area).
Further analyses included the comparison of the cortical baseline activity (baseline images) and stimulated activation (difference images) between CI users with higher and lower performance in speech comprehension. Similar to the EEG data analysis, the subgroups of higher and lower CI performance were compared by means of two-sample t-tests in consideration of two different contrasts [(1) lower performer > higher performers and (2) higher performers > lower performers]. Additionally, correlation analyses were performed using SPM to explore for the relationship between difference images on the one hand and EEG, audiometric, as well as neuropsychological data on the other hand.
Speech Comprehension Ability and Cognitive Functions
Results for speech audiometry, cognitive tests, and speech task performance are listed in Table 2 with separate means for patients with higher and lower speech comprehension as well as the results from group comparisons across the different tests. In the context of speech audiometry, the CI users showed overall high performance for speech recognition in quiet, with average scores of 81.7 ± 12.1% (Freiburg monosyllabic word test) and 93.0 ± 11.0% (HSM sentence test without background noise). As expected, speech recognition in noise was remarkably lower, as revealed by the average recognition score of 48.5 ± 19.1% in the HSM sentence test (10 dB SNR) and the average 50% speech reception threshold of + 8.6 ± 4.2 dB SNR in the GÖSA.
Table 2. Results of speech audiometry, cognitive tests, and sentence-discrimination task during EEG recording.
Performance scores in the cognitive tests showed for the MWT-B (verbal intelligence) an average percentile rank of 68.5 ± 24.5; for the SICSPAN (working memory capacity), an average of 50.2 ± 11.3 (total sum); and for the RWT (verbal fluency), an average percentile rank of 54.8 ± 28.9. Z-scores for verbal learning were on average of −0.2 ± 1.2 and for verbal recall of 0.0 ± 1.0.
The performance in the semantic-anomaly paradigm was generally high, with mean hit rates of 92.6 ± 6.9% (correct identification of semantically correct sentences) and mean correct rejection rates of 93.8 ± 6.6% (correct identification of sentences with semantic violations). Comparing the performance in the semantic-anomaly paradigm between CI users with lower and higher speech reception thresholds (median split with cutoff + 7.6 dB SNR in the Göttinger sentence test) revealed reduced hit rates in speech comprehension ability for the lower as compared to the higher CI performers [88.7 ± 6.3% vs. 96.2 ± 5.3%; t(19) = 2.8, p < 0.05, r = 0.5]. The number of correct rejections in contrast was not different between groups [92.4 ± 7.2% vs. 95.0 ± 5.6%; t(19) = -0.9, p = 0.39, r = 0.2].
The subjective rating of the listening effort during the EEG task was relatively low, as indicated by the ratings of 1.6 ± 0.6 on average. Comparison of scores for cognitive tests (tested in an unpaired t-test) and listening effort between groups of lower and higher performers (using the Mann–Whitney U-test) did not reveal any statistical significant differences [SICSPAN: t(19) = 0.3, p = 0.74, r = 0.1; MWT-B: t(19) = 0.4, p = 0.67, r = 0.1; RWT: t(19) = 0.2, p = 0.87, r = 0.1; CERAD verbal learning: t(19) = -0.5, p = 0.64, r = 0.1; CERAD verbal recall: t(19) = -0.5, p = 0.64, r = 0.1 and listening effort, U = 39, p = 0.28, r = 0.24].
A clear negative peak can be observed around 150 ms (N1 peak), followed by a positive deflection around 240 ms (P2 peak). On a group level, the N1 showed a mean amplitude of -4.6 ± 1.6 μV with a mean peak latency of 154.1 ± 14.7 ms. Average values for the P2 mean amplitude and peak latency were 4.1 ± 2.0 μV and of 244.7 ± 19.3 ms, respectively. Independent t-tests revealed no significant differences between groups with lower and higher CI performance for N1/P2 amplitudes [N1: t(19) = 0.8, p = 0.45, r = 0.2; P2: t(19) = -1.1, p = 0.27, r = 0.3] and latencies [N1: t(19) = 0.8, p = 0.45, r = 0.2; P2: t(19) = 0.8, p = 0.45, r = 0.2]. Sequential two-tailed t-tests, using a sliding window of 2 ms at α = 0.5% were used to compare ERPs on sentence onset between groups of lower and higher CI performers, whereby an interval was considered as significantly different between groups if at least 10 consecutive data points reached a p < 0.05 (Guthrie and Buchwald, 1991). To control for multiple comparisons, p-values were adjusted using the false discovery rate (FDR; Benjamini and Hochberg, 1995). Results showed no statistical difference for time windows of the N1 and the P2, although on the descriptive level, the N1 and P2 peaks of lower performers were reduced compared with the higher performers (lower vs. higher performer; N1: 80–200 ms, p ≥ 0.99, corrected; P2: 160–280 ms, p ≥ 0.99, corrected).
Difference waves on the onset of the final word (ERPcritViol - ERPcritCorr) are shown in Figure 1B, using a centroparietal ROI, respectively. A slow negative deflection can be observed between 500 and 800 ms, referred to as the N400. The average mean amplitude on a group level was -3.6 ± 1.6 μV with mean peak latencies of 593.5 ± 132.7 ms. Around 900 ms, a positive deflection (P600) with an average mean amplitude of 3.0 ± 1.5 μV and a mean peak latency of 867.8 ± 41.7 ms can be detected. Independent t-tests revealed no significant differences between groups with lower and higher CI performance for N400 amplitudes [t(19) = 0.9, p = 0.36, r = 0.2] and latencies [t(19) = 0.2, p = 0.85, r = 0.1]. For the P600, results, however, revealed a significant difference for P600 amplitudes [t(19) = 2.1, p = 0.05, r = 0.4], but not for latencies [t(19) = -0.1, p = 0.89, r = 0.02]. Sequential two-tailed t-tests, using a sliding window of 2 ms at α = 0.5% were used to compare ERPs on the onset of the critical word between groups of lower and higher CI performers, whereby an interval was considered as significantly different between groups if at least 10 consecutive data points reached a p < 0.05 (Guthrie and Buchwald, 1991). To control for multiple comparisons, p-values were adjusted using the FDR (Benjamini and Hochberg, 1995). Results showed no statistical difference for the time window of the N400, although on the descriptive level, the N400 of lower performers was reduced compared with the higher performers (300–900 ms, p ≥ 0.64, corrected). For the time window of the P600, lower performers compared to higher ones, descriptively, revealed elevated P600 peaks, which however could not be verified by sequential two-tailed t-tests, when correcting for multiple comparisons (750–940 ms, p ≥ 0.64, corrected).
Comparison of Brain Perfusion: Rest Condition vs. Speech Condition
The activation pattern induced by the sentence discrimination task (speech condition > rest condition) is displayed in Figure 2 (employing a threshold of p < 0.001 uncorrected for multiple comparisons). In Figure 2A, the results of a voxel-wise paired t-test are shown using the original, non-flipped data, whereas Figure 2B shows the results of the paired t-test including flipped images (left to right reversed) for patients stimulated on the left side.
Figure 2. Statistical parametric maps (SPMs) reflecting relative increases of perfusion due to performing the speech condition in comparison to the rest condition. SPMs are overlaid to a surface rendered MRI data set in the Montreal Neurological Institute (MNI) space. There are some areas displayed more transparent than others, which refers to their distance to the projection surface. (A) The results obtained using the original data without left/right flipping are displayed. (B) To explore for potential effects of the side of stimulation, each data set with left-sided stimulation was flipped in the median sagittal plane, while data sets with right-sided presentations remained unflipped. This resulted uniformly in data sets with stimulation from the “right” side in relation to the images for analysis. For both comparisons, the contrast speech condition > rest condition is displayed. Strong significant perfusion increases due to the task are similarly visible in (A,B), showing bilateral activation in the superior and middle temporal cortices and the inferior prefrontal cortex. Note the minor difference between unflipped and flipped images, with a lack of perfusion increase in the ipsilateral (right) primary auditory cortex (BA 41; yellow arrow) for the flipped in contrast to the unflipped image (see also Supplementary Table A.2).
Regarding the original, non-flipped data (Figure 2A), the paired t-test between the speech and rest condition revealed a strong bilateral activation in the temporal lobe, including the superior [STG; Brodmann areas (BAs) 22, 41, 42], the middle (MTG; BA 21), and the inferior temporal gyrus (ITG; BA 20), as well as the temporo-polar (BA 38) area.
Furthermore, the paired t-test showed significant activations in Broca’s area (BA 45 left), bilaterally in the pars orbitalis (BA 47), and the orbitofrontal cortex (BA 11) (Figure 2A) as well as in smaller areas of the left premotor cortex (BA 6) (Supplementary Table A.1.1).
Regarding the paired t-test for the flipped data, we observed significant activations in the bilateral STG (BA 22, 42), left BA 41, bilateral temporo-polar (BA 38), and left frontal areas (BAs 10, 11, 46, 47), as well as in left BA 45 (Figure 2B), which is bilaterally activated without using an extent voxel threshold (p < 0.001, k = 0, Supplementary Table A.2.1). Here, also small area activations for the left motor cortex (BA 6) have been detected. However, no activation could be detected at that level of significance in the right primary auditory cortex (BA 41) with stimulation always from the (ipsilateral) right side. The analyses of either data (non-flipped and flipped) including correction for multiple comparisons (FWE) revealed significant activations induced by stimulation only in temporal regions at a level of p < 0.05, specifically, on the right (BAs 22, 21) and left side (BA 42) for non-flipped data and the left (BAs 42, 48, 22) and the right side (BAs 22, 21) for the flipped data (Supplementary Tables A.1.2, A.2.2).
Contrasting Groups of Higher and Lower CI Performance in Speech Comprehension
Group comparisons were performed based on the difference images (speech condition - rest condition) (Table 3 and Figure 3) and the rest condition images (Table 4 and Figure 4). The following results were obtained employing a threshold of p < 0.001 uncorrected for multiple comparisons. Group comparisons including correction for multiple comparisons described in Contrasting Groups of Higher and Lower CI Performance in Speech Comprehension did not reveal any suprathreshold voxel with p < 0.05. During the speech comprehension task, lower compared to higher performers showed a significantly higher activation in the left frontal BA 9 (Figure 3A). Smaller areas of activation were seen in the left ITG (BA 20), as well as in the right frontal BA 8 (Table 3). In contrast, higher compared to lower performers showed significantly higher activation in the left occipital area (BA17), as well as in the right parietal (BA 3) and temporal (BA 20) areas (Figure 3B and Table 3).
Figure 3. Statistical parametric maps (SPMs) reflecting group differences between higher and lower cochlear-implant (CI) performers with regards to speech-related activation in the context of a semantic-anomaly paradigm. Relative perfusion increases (activations) due to performing the sentence discrimination task are shown. Note a pattern of a (A) prefrontal perfusion increase in lower compared to higher performers and (B) an increased occipital and parietal perfusion in higher compared to lower performers.
Figure 4. Statistical parametric maps (SPMs) reflecting relative differences in baseline perfusion between groups of cochlear-implant (CI) users with higher and lower performance in speech comprehension according to the GÖSA test (median-split procedure). SPMs are overlaid to a surface rendered MRI data set in the Montreal Neurological Institute (MNI) space. There are some areas displayed more transparent than others, which refers to their distance to the projection surface. Note a pattern of higher baseline perfusion in right parietal areas and motor cortex in CI users with lower performance compared to higher performers (A), while a pattern of higher baseline perfusion in hippocampal and inferior frontal areas is seen in higher as compared to lower CI performers (B).
Table 3. Results of unpaired t-tests comparing the difference images (speech condition - rest condition) between CI users with lower and higher speech comprehension (significance level used for inferences at a voxel level p < 0.001, extent voxel threshold k = 0).
Table 4. Results of unpaired t-test comparing the rest condition image between cochlear-implant (CI) users with lower and higher speech comprehension (significance level used for inferences at a voxel level p < 0.001, extent voxel threshold k = 0).
Rest Condition Image
During the rest condition, CI users with lower performance demonstrated significantly higher activity in the right motor and premotor cortex (BAs 4, 6) as well as the right parietal regions (BAs 2, 3, 5) (Figure 4A and Table 4). However, the group of CI users with higher performance showed significantly higher baseline activity in the left hippocampal area (BA 48) and left inferior frontal areas (BA 11, 25) as shown in Figure 4B. Additionally, a higher resting-state perfusion in higher compared to the lower performers was detected in smaller areas of right temporo-polar area (BA 38), the frontal cortex BA 47 (bilateral), and the left inferior temporal cortex (BA 20) (Table 4).
Brain Activation in SPECT vs. Audiometric and Cognitive Performance
We observed widespread positive correlations between activation in the difference image (speech condition minus rest condition) and the results of the Freiburg monosyllabic word test (Figure 5A; see also Supplementary Table A.3.1) and the MWT-B, assessing the verbal intelligence (Figure 5B; see also Supplementary Table A.4.1). Both correlations showed a similar distribution pattern of significantly activated regions, including broad bilateral temporal (BAs 20, 21, 22, 38, 41, 42), frontal (BAs 9, 10, 11, 44, 45, 46, 47), and parietal areas (BA 1, 2, 3, 40), as well as the bilateral motor cortex (BAs 4, 6) (Figures 5A,B). These results for the Freiburg monosyllable test and the MWT-B test were obtained employing a threshold of p < 0.001 uncorrected for multiple comparisons. Including correction for multiple comparisons (FWE) and using a threshold of p < 0.05 restrained the observed significances to the temporal cortices. Specifically, significant correlations were detected with the Freiburg test on the right (BA 22, 21) and left side (BA 42, 48, 22) and the MWT-B on the left (BA 42, 48, 22, 21, 20) and the right side (BA 48, 22, 21) (Supplementary Tables A.3.2, A.4.2).
Figure 5. Statistical parametric maps (SPMs) reflecting correlations between activation during the speech-discrimination task (semantic-anomaly paradigm) and the result of the Freiburg monosyllabic word test (speech recognition; A) and the MWT-B (verbal intelligence; B), respectively. SPMs are overlaid to a surface rendered MRI data set in the Montreal Neurological Institute (MNI) space. There are some areas displaying more transparency than others, which refers to their distance to the projection surface. Note the extended activation of bilateral areas in temporal, frontal, and parietal cortices, showing significant relationships with speech processing (during task) and verbal intelligence.
Furthermore, we found smaller areas of positive correlations [only in testing without correction for multiple comparisons and not in tests including correction (FWE)] between the difference image (speech condition minus rest condition) and the results of the SICSPAN test for working memory (Figure 6A), as well as with verbal learning (Figure 6B) and verbal recall (Figure 6C). Specifically, higher capacity in working memory (SICSPAN test) correlated with enhanced perfusion in the left STG (BA 22), MTG (21), and ITG (BA 20), as well as in left parietal (BAs 2, 3) and right occipital regions (BAs 18, 19) (Figure 6A, see also Supplementary Table A.5). With regards to verbal learning, higher Z-scores were related to higher perfusion in the left STG (BA 22), MTG (BA 21), and the temporopolar area (BA 38) (Figure 5B; see also Supplementary Table A.6). Finally, a better performance in the verbal recall was associated with a higher activation in the left STG (BA 22) MTG (BA 21), and ITG (BA 20) (Figure 6C, see also Supplementary Table A.7). In the Supplementary Tables A.5–7, the results are listed without application of an extent voxel threshold (k = 0).
Figure 6. Statistical parametric maps (SPMs) reflecting correlations between activation during the speech-discrimination task (semantic-anomaly paradigm) and cognitive tests, in particular (A) working-memory capacity, (B) verbal learning and (C) verbal recall. Note vastly predominant left temporal correlations for all three tests.
Brain Activation in SPECT vs. EEG Components
Brain activation (speech condition minus rest condition) correlated (based on analyses without correction for multiple comparisons) in specific small brain regions negatively with the mean peak amplitude values of the N400 ERP (Figure 7A). An enhanced N400 ERP was related to a higher regional brain activation during the speech comprehension task in the following brain areas: the right temporal BA 37 and occipital in BA 19 (Figure 7A). Further correlations in smaller areas were seen in the left temporal (BA 37) and occipital areas (BA 17, 19; see Table 5). With regard to the P600 ERP, larger amplitudes were correlated (based on analyses without correction for multiple comparisons) with enhanced activation in the left parietal cortex (BA 39; Figure 7B). Additionally, smaller regions of positive correlation were observed in occipital regions (left and right BA 18 and left BA 19; see Table 6). None of the correlations performed with correction for multiple comparisons (FWE) revealed any suprathreshold voxels with p < 0.05.
Figure 7. Statistical parametric maps (SPMs) reflecting correlations between relative increases of perfusion during the speech-discrimination task (semantic-anomaly paradigm) and the (A) N400 and the (B) P600 amplitudes, respectively. SPMs are shown as glass brain images. Note the predominant correlations in the temporal cortex (BA 37, A) and the parietal–occipital areas (in B).
Table 5. Results of negative correlations: difference image (speech condition - rest condition) vs. N400 EEG component (significance level used for inferences at a voxel level: p < 0.001, extent voxel threshold k = 0).
Table 6. Results of positive correlations: difference image (speech condition - rest condition) vs. P600 EEG component (significance level used for inferences at a voxel level p < 0.001, extent voxel threshold k = 0).
Audiometric and Cognitive Performance vs. EEG Components
Individual linear regression analyses between measures of audiometric/cognitive performance (Freiburg monosyllabic word test, HSM sentence test in quiet and in noise, GÖSA, verbal intelligence and fluency, working memory, and verbal learning) and EEG components (N1, P2, N400 and P600) were performed. None of these reached significance (p always > 0.05) and squares of the correlation coefficients were always below 0.15 excluding a relevant correlation between the respective data.
The present study aimed to better understand the high variability in CI outcomes. We used an innovative multimodal diagnostic approach, including brain-perfusion SPECT with tracer injection during EEG measurement to examine speech processing in CI users. Three main findings were obtained: First, the CI users activated a temporo-frontal network for speech processing and showed correlations between activation in the temporal gyrus and occipital regions on the one hand and cognitive ERP amplitudes on the other hand. This demonstrates a close connection between ERP effects and cortical activation in CI users. Second, the CI users with lower and higher speech comprehension showed different activation patterns for baseline brain activity (“rest condition”) as well as for activation during speech processing (“speech condition”), pointing to differential allocation of neural resources and strategies used for speech processing. Third, we observed strong correlations between the brain networks activated during speech processing and specific cognitive abilities, in particular working memory capacity and verbal memory functions, implying that these cognitive functions play a crucial role for speech comprehension in CI users.
Brain Regions Recruited for Speech Processing in NH Listeners and in CI Users
The speech processing cascade, comprising the primary acoustic analysis, the identification of phonemes and words, and the integration of semantic and syntactic information, includes a complex system of interacting brain areas (Hickok and Poeppel, 2000; Friederici, 2002; Scott and Johnsrude, 2003). The traditional view of speech processing proposed the dominant involvement of the left-hemisphere inferior frontal and temporal cortices (Broca, 1861; Wernicke, 1874). However, more recent findings have suggested a bilateral involvement and the existence of two dorsal and two ventral pathways (Friederici et al., 2006; Saur et al., 2008; Saur et al., 2010; Friederici, 2012), with the dorsal streams connecting the superior temporal gyrus (STG) with the premotor cortex and BA 44, respectively, whereas the ventral streams connect, on the one hand, the STG to BA 45/47 and, on the other hand, the anterior temporal cortex to the frontal operculum. Regarding semantic sentence processing, temporal as well as inferior frontal cortical areas are involved (for a review, see, e.g., Friederici, 2002), in particular BA 45/47 and the left middle temporal gyrus (MTG). Further MEG and functional MRI (fMRI) findings with NH listeners support the essential role of the left MTG in semantic processing (see, e.g., Lau et al., 2008).
CI users have been shown to recruit similar brain regions and circuits for speech processing when compared with NH listeners, although they reveal lower activation in the temporal voice area (Giraud et al., 2000; Coez et al., 2008). Furthermore, the CI users show a compensatory increase in activation in the left inferior prefrontal cortex (Broca’s region) (Giraud and Truy, 2002), the anterior superior temporal phonologic region (Giraud et al., 2000), temporo-occipital visual areas (Giraud et al., 2001b; Giraud and Truy, 2002), parietal attentional regions (Giraud et al., 2000; Coez et al., 2014), and parahippocampal memory areas (Giraud et al., 2001c). Consistent with previous observations with CI users and NH listeners, in our study, the comparison of speech task versus rest revealed large activations in the bilateral temporal cortex (BA 41, 42, 22, 21, and 20) as well as the inferior prefrontal cortex (BA 45, 47) (Kutas and Hillyard, 1980a, b; Friederici, 2002; Van Petten and Luka, 2006; Hahne et al., 2012). Thus, our results are in line with previous observations by demonstrating the recruitment of a temporo-frontal network of brain areas during speech processing in CI users. Moreover, they reveal that a sentence comprehension task typically employed to study ERP effects is likewise suitable to achieve synchronously brain activations detectable in emission tomography.
Interestingly, we did not observe an activation of the inferior parietal cortex and the dorsal part of Broca’s area in the inferior prefrontal cortex (BA 44). This suggests that the dorsal pathway – known to be particularly involved in processing of syntactically complex sentences (Friederici, 2012) – was not considerably activated in our patients. This is plausible due to the fact that the sentences used in our discrimination paradigm were syntactically simple and their processing may have relied rather on the ventral than the dorsal pathways. Regarding the parietal areas, the lack of activation in these regions might be related to the fact that participants were stimulated unilaterally, while previous studies reporting parietal recruitment during speech processing are restricted to CI users with bilateral stimulation (Coez et al., 2014). However, our CI users showed supplementary activation in the hippocampus (BA 48), pointing to memory functions involved in performing the semantic-anomaly paradigm.
Interestingly, when correcting our data sets for the side of stimulation (left-sided stimulation flipped, right-sided stimulation unflipped), we did not observe activation in the ipsilateral primary auditory cortex (BA 41). Similarly, previous fMRI studies with NH listeners have reported that monaural presentation of speech results in a stronger contralateral activation of the primary auditory cortex (Jäncke et al., 2002; Stefanatos et al., 2008). Contralaterally predominant activation has also been observed with unilateral as opposed to bilateral stimulation in CI users (Green et al., 2011; Coez et al., 2014).
Different Patterns of Brain Activity at Rest and Activation Related to Speech Processing in CI Users With Lower and Higher Performance
There is high interindividual variability in speech comprehension abilities across CI users (Lazard et al., 2012; Blamey et al., 2013). Differences between proficient and non-proficient CI users seem to exist already at the time before implantation, as indicated by the finding of distinct preimplantation activation patterns during a (written) word rhyming task between (prospective) lower and higher performance after implantation (Lazard et al., 2010). After implantation, proficient and non-proficient CI users have been shown to recruit the auditory cortex to a different degree, with reduced recruitment of the temporal voice area (with regard to extent and only unilaterally) in the poor performers when compared with good performers (Coez et al., 2008). However, in the present study, the activation of temporal regions (BA 20) was comparable between the two groups of CI users. This might be explained by the fact that the current study compared CI users with high and moderate speech comprehension (referred here as higher and lower CI performers), while the two aforementioned studies compared CI users with clearly different good and bad performance, leaving out a broad spectrum of patients with intermediate performance.
Despite the lack of a group difference in the temporal regions, we observed for the higher performing CI users additional activations of parietal and occipital regions, whereas for the group of lower performing subjects, we found a stronger activation of superior frontal areas. Similar to our results, previous studies have reported for good CI performers increased activity in temporo-occipital visual areas (Giraud et al., 2001b; Giraud and Truy, 2002) and parietal attentional regions (Giraud et al., 2000; Coez et al., 2014), suggesting compensatory networks of speech processing in these proficient individuals. The auditory-evoked activation in the visual cortex might be related to functional, cross-modal reorganization of the visual cortex that is used to compensate for the degraded auditory input via the CI. Accordingly, it has been shown in previous studies that activation of the visual cortex by auditory stimulation is positively related to the CI performance (Giraud et al., 2001b, c; Strelnikov et al., 2013; Chen et al., 2016). Thus, it can be speculated that, in the present study, the two groups of CI users used different compensatory strategies that may be related to differences in cross-modal reorganization of the visual and auditory cortex. CI speech performance seems to be good as long as the (beneficial) auditory-evoked activation in the visual cortex is higher than the (maladaptive) visual-evoked activation in the auditory cortex (Chen et al., 2016). This is in line with the current findings, showing that specifically the higher performers showed an enhanced beneficial cross-modal reorganization in the visual cortex.
Regarding the lower performing CI users, we observed increased superior frontal activations (BA 9) compared with those with higher performance. This is in contrast to a previous study reporting that activation in non-proficient CI users is restricted to temporal areas (Mortensen et al., 2006). The discrepancy of results is likely attributable to variations in methodology, in particular in terms of the experimental task (active vs. passive task) and the speech comprehension of the lower performing group (open-set speech comprehension: ≤ 60 vs. ≥ 65%). Furthermore, our observation of increased superior frontal activation particularly in the poorer performing CI users is meaningful, as it might reflect enhanced neural resource allocation due to limitations in electrical hearing. Indeed, peripheral factors, for instance the distance of the CI electrode arrays to the modiolar wall and the number of surviving spiral ganglion cells have been shown to affect speech comprehension with the CI (Nadol, 1997; Holden et al., 2013). In case of suboptimal peripheral conditions, the resulting strong(er) mismatch between the CI input and the attributes stored in the long-term memory may require additional explicit processing and involve cognitive resources, in particular working memory functions (Rönnberg et al., 2013; Finke et al., 2015), and may cause enhanced listening effort (Berding et al., 2015). Indeed, it has been shown previously that NH listeners recruit additional prefrontal regions specifically in difficult listening conditions in which the listening effort is enhanced (Davis and Johnsrude, 2003; Peelle, 2018). Thus, it is likely that during speech processing, the lower CI performers rely on a different processing strategy compared with the higher performers by particularly allocating executive cognitive resources located in specific frontal regions. However, this compensatory strategy seems to be limited, as indicated by the fact that the lower CI performers did not reach the speech comprehension performance levels of higher CI performers.
Additionally, we observed CI-outcome-related distinct brain activity patterns at rest: lower performers showed higher resting state perfusion in motor cortex and parietal areas, whereas higher performers showed higher perfusion in (inferior) frontal (BA47), inferior temporal (BA20), temporopolar (BA38), and hippocampal (memory) regions. The prognostic relevance of resting activity has been demonstrated before (Lee et al., 2007; Giraud et al., 2011; Strelnikov et al., 2015a; Suh et al., 2015). It has been shown, for example, that a low (resting-state) activity in the primary auditory/superior temporal cortex – indicating that no maladaptive cross-modal visual take-over has taken place – is related to a better CI outcome (Lee et al., 2007; Strelnikov et al., 2015a; Suh et al., 2015). The same has been observed for an increased activity in the prefrontal cortex, in particular Broca’s area (Lee et al., 2007; Suh et al., 2015). Our results are consistent with these previous findings by showing that CI users with higher performance have increased resting-state perfusion particularly in left-sided (inferior) frontal (BA 11, 47, 25) areas.
Overall, our results show that CI users with lower and higher speech comprehension recruit distinct brain networks not only during speech processing but also during rest. This points to different compensation strategies for the processing of the degraded CI speech signal and suggests different adaptation of the brain in response to the individual auditory experience.
Cognitive Abilities and Their Relationship With Brain Activation During Speech Processing
The performance of CI users has been shown to be influenced by cognitive factors, like verbal fluency and working memory capacity (Rönnberg et al., 2013; Finke et al., 2016a). However, it is currently widely unknown how individual differences in cognitive abilities relate to cortical response patterns during speech processing. The present study suggests that enhanced activation in (predominantly) temporo-frontal areas during speech processing is positively associated with higher word recognition scores and higher verbal intelligence. Temporo-frontal regions encompass areas related to auditory, executive, and memory functions that are used for speech comprehension in the context of the semantic-anomaly paradigm (Lau et al., 2008; Kutas and Federmeier, 2011). Consistent with our observations, a coupling between speech recognition scores and activation of auditory areas, in particular the Heschl’s gyrus (BA 41), the superior temporal, and the angular gyrus, has been observed previously (Łukaszewicz-Moszyńska et al., 2014). In sum, these results underpin that cognitive abilities and speech comprehension ability in CI users relate to specific speech-evoked cortical activation in temporo-frontal regions. Thus, enhanced activation in these regions seems to allow better verbal abilities and higher word comprehension ability with a CI.
Additionally, we observed better verbal abilities and enhanced working memory capacity being associated with increased activations in the superior and middle temporal gyrus. This indicates that better cognition results in enriched activations of typical auditory areas, which might be related to better speech comprehension with CI as well. As all cognitive tests recall phonetic, vocabulary, and memory abilities simultaneously, this might indicate the use of similar resources during the performance of these cognitive tests and the semantic speech task.
ERPs and Their Relationship With Brain Activation Detected With SPECT
In the current study, the CI users showed an N1-P2 ERP in response to the onset of the sentence, indicating processing of speech at the level of the auditory cortex in CI users (Naatanen and Picton, 1987; Friesen and Picton, 2010). Furthermore, the difference waves (ERPcritViol - ERPcritCorr), time locked to the onset of the final word of the sentence, revealed a more negative amplitude to sentences with a semantic violation compared to correct sentences, referred to as an N400 effect (Kutas and Hillyard, 1980c; Kutas and Hillyard, 1984). The N400 is considered as an index of neural effort of automatic word-into-context integration (Strauss et al., 2013). In other words, it is assumed to reflect the level of difficulty with which a word is integrated in the respective context (Van Petten et al., 1999). Interestingly, the observed latency of the N400 effect in CI users (∼700–800 ms) was delayed when compared with the N400 latency of NH listeners reported in the literature (∼400 ms; e.g., Lau et al., 2008). Nevertheless, it was comparable to a previous study, reporting a delayed N400 response in CI users when compared with NH listeners (Hahne et al., 2012). These results suggest that adverse listening conditions, as experienced by CI users with the degraded auditory input from the implant, lead to more effortful and thus delayed semantic integration processes (Finke et al., 2016a).
The present study showed negative correlations between the N400 response and the temporal (in particular BA 37), as well as activation in the visual cortex (SPECT difference image: speech condition - rest condition). Specifically, more negative N400 amplitudes in the present study were found to be associated with higher perfusion in a broad network, including temporal and occipital regions. The N400 has been suggested to be primarily generated in the left middle temporal gyrus (Lau et al., 2008; Friederici, 2012). BA 37 has also been suggested previously to be part of the semantic processing network (Lau et al., 2008; Ardila et al., 2015). The observed correlation with the temporal region suggests this region to be as well involved in the generation of the N400 in CI users. However, results of the present study also showed a strong correlation with visual areas, strongly suggesting an additional cross-modal recruitment of visual areas during semantic processing in CI users. The engagement of occipital areas might be related to the fact that although sentences are presented purely auditorily, they might be internally visualized. Furthermore, it has been previously reported that CI users show an enhanced audiovisual coupling (Schierholz et al., 2015, 2017; Strelnikov et al., 2015b) and that activation of the visual cortex by auditory stimulation is positively related to the CI performance (Giraud et al., 2001b, c; Strelnikov et al., 2013; Chen et al., 2016), indicating that cross-modal reorganization in the visual cortex and enhanced audiovisual coupling support speech processing in CI users.
The difference waves revealed that the N400 was followed by a positive deflection at around 900 ms after the final word onset. This late component has been referred to the P600 (Osterhout and Holcomb, 1992), which typically peaks between 300 and 800 ms (Friederici et al., 2000; Friederici, 2006). Similar to the N400 response, our results suggest a delayed P600 response in our CI users, which can be attributed to the degraded input from the implant, leading to delayed higher-level speech processing. Traditionally, the P600 has been related to syntactic processing effort in general and it is, for example, observed in the context of syntactical repair, syntactical complexity, and difficulties with syntactic integration (Osterhout and Holcomb, 1992; Hagoort et al., 1993; Kaan et al., 2000; Friederici et al., 2002). Nevertheless, the P600 has been recently discussed in a semantic context as well (see, e.g., Bornkessel-Schlesewsky and Schlesewsky, 2008). The studies by Kolk et al. (2003) and van Herten et al. (2005), for example, observed a P600 effect elicited by semantic anomalies, challenging the merely syntactic account of the P600. This view has been supported by the study by Shen et al. (2016), suggesting that the P600 reflects a general mechanism of semantic reinterpretation and conflict monitoring that leads to the retrieval of word knowledge from long-term memory. A systematic review on the effects of semantic incongruency by Van Petten and Luka (2012) has identified 21 out of 64 studies that exhibited a biphasic N400/P600 effect to incongruent sentences, confirming that indeed the P600 is elicited by semantic anomalies in sentences. Regarding the generators of the P600, the bilateral medial/posterior temporal cortex has been identified in NH listeners (Service et al., 2007). The current results showed that stronger P600 amplitudes were associated with higher perfusion in parietal and occipital areas. The parietal correlation involved BA 39 (angular gyrus), which has been shown to be a part of the semantic processing network (Lau et al., 2008). Additionally, stronger P600 responses were associated with higher perfusion in occipital areas, suggesting once again that CI users recruit additional visual regions during semantic processing.
Our results revealed that both the N400 and the P600 ERPs correlate with the activation of a broad but distinct network. Importantly, both components correlate with activation in occipital areas (N400: BAs 17, 19; P600: BAs 18, 19). This extends previous research by showing that CI users strongly rely on visual cortex activation during semantic speech processing. All in all, our findings show a close connection between ERP effects and cortical activation in CI users, demonstrating that the combination of SPECT and EEG measurements provides unique and valuable insights into the cognitive processes underlying speech comprehension in CI users.
Our results also extend previous studies by indicating that not only sensory but also cognitive ERPs, in particular the N400 and the P600 response, can distinguish – although in the present study not with statistical significance, but at least on the descriptive level – between CI users who have higher versus lower speech comprehension. Thus, our results might point to potential different abilities of lower and higher performing CI users in detecting and integrating semantic violations in sentences. We speculate that increasing the sample size would have resulted in less variance in the data and statistical group differences for the N400 and the P600 amplitudes, respectively. Interestingly, on the descriptive level, our results showed that the amplitudes of the N1, P2, and N400 are reduced in the lower compared to the higher CI performers, while we observed on a descriptive level the opposite pattern, that is an enhanced amplitude, for the P600 in the lower compared to the higher performers. These descriptive observations support the significant group differences revealed in the SPECT data and point together to different strategies for speech comprehension in lower and higher performers. While lower performers invest less neural resources in automatic word-into-context integration (reflected by the reduced N400 response, here observed on a descriptive level), they use additional explicit processing resources for semantic reinterpretation and retrieval of word knowledge from the long-term memory (reflected by the enhanced P600 response, here observed on a descriptive level). Furthermore, this is in line with our observation of increased frontal activation in the lower compared to the higher CI performers and the Ease of Language Understanding (ELU) model (Rönnberg et al., 2013), according to which additional cognitive resources are required for speech comprehension in demanding listening situations, which particularly applies to CI users with lower speech comprehension.
The present study showed that higher and lower CI outcome is associated with different brain activation patterns. Furthermore, our results revealed the meaningful applicability of a combined EEG and SPECT multimodal diagnostic approach for examination of speech processing in CI users. Our results revealed that based on a sentence discrimination task, activation of a temporo-frontal network can be detected in both diagnostic modalities correspondingly to previous observations with PET in CI users. Furthermore, the present results revealed significantly different activation patterns between lower and higher CI performers. The results point to the use of different compensational strategies for the degraded auditory input and different adaptations of the brain in response to the individual auditory experience for groups of CI users with higher and lower performance. Moreover, differences between these groups of CI users, at least on the descriptive level, were observed for the EEG data. Here, the lower performers showed reduced amplitudes for the sensory ERPs (auditory N1 and P2) and the later cognitive N400 ERP, whereas the opposite pattern, that is enhanced amplitudes, were observed for the P600 ERP in the lower compared to the higher performers. These findings point to more pronounced deficits/limitations in CI users with lower performance due to a particularly degraded auditory input and compensatory strategies to overcome these limitations by a stronger recruitment of higher-cognitive resources, involving frontal regions.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
The studies involving human participants were reviewed and approved by the local ethics committee at Hannover Medical School (vote no. 6678) and the German Federal Office for Radiation Protection (reference number Z 5-22461/2-2014-012). The patients/participants provided their written informed consent to participate in this study.
MK designed the study, recruited the study sample, collected and analyzed EEG and SPECT data, wrote the manuscript, and prepared the tables and figures. IS designed the study, recruited the study sample, collected and analyzed EEG data, contributed to the writing of the manuscript, and prepared the figures. MM contributed to the design of the study, assisted during the data collection, provided technical support, and revised the manuscript. FW provided technical support for the SPECT measurement, contributed to the SPECT data analysis, and revised the manuscript. AH contributed to the experimental setup, advised on the EEG data analysis, and revised the manuscript. AB contributed to the setup of the EEG facility. LG provided expertise for the SPECT acquisition protocol. FB allocated personnel resources enabling SPECT acquisitions. PS conceived and designed the study, supervised the EEG data collection, contributed to the EEG data analysis, and contributed to the writing of the manuscript. GB conceived and designed the study, supervised the SPECT data acquisition, contributed to the SPECT data analysis, and contributed to the writing of the manuscript. All authors contributed to the article and approved the submitted version.
This work was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy – EXC 2177/1 – Project ID 390895286. This work was supported by the DFG Cluster of Excellence EXC 1077/1 “Hearing4all.”
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
We would like to thank Prof. Dr. Stefan Debener for his methodological support and Prof. Dr. Bruno Kopp for his expertise with regard to the neurocognitive testing. We furthermore would like to thank Dr. Angelika Illg for support in patient recruitment and Dr. Mareike Finke for her contribution to the initial study design. Moreover, we thank Maria Nierada and Claudia Diekmann for their support during the SPECT scans. Last but not the least, we would like to thank all CI patients who participated in the present study.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2020.00787/full#supplementary-material
- ^ https://www.memoryclinic.ch/de/main-navigation/neuropsychologen/cerad-plus/
- ^ http://www.billauer.co.il/peakdet.html
Allen, A., Barnes, A., Singh, R. S., Patterson, J., Hadley, D. M., and Wyper, D. (2004). Perfusion SPECT in cochlear implantation and promontory stimulation. Nucl. Med. Commun. 25, 521–525. doi: 10.1097/00006231-200405000-00015
Apostolova, I., Lindenau, M., Fiehler, J., Heese, O., Wilke, F., Clausen, M., et al. (2008). Detection of a possible epilepsy focus in a preoperated patient by perfusion SPECT and computer-aided subtraction analysis. Nuklearmedizin 47, N65–N68.
Ardila, A., Bernal, B., and Rosselli, M. (2015). Language and visual perception associations: meta-analytic connectivity modeling of Brodmann area 37. Behav. Neurol. 2015:565871. doi: 10.1155/2015/565871
Benjamini, Y., and Hochberg, Y. (1995). Controlling the false discovery rate - a practical and powerful approach to multiple testing. J. R. Statist. Soc. Ser. B 57, 289–300. doi: 10.1111/j.2517-6161.1995.tb02031.x
Berding, G., Wilke, F., Rode, T., Haense, C., Joseph, G., Meyer, G. J., et al. (2015). Positron emission tomography imaging reveals auditory and frontal cortical regions involved with speech perception and loudness adaptation. PLoS One 10:e0128743. doi: 10.1371/journal.pone.0128743
Blackwood, D. H., Ebmeier, K. P., Muir, W. J., Sharp, C. W., Glabus, M., Walker, M., et al. (1994). Correlation of regional cerebral blood flow equivalents measured by single photon emission computerized tomography with P300 latency and eye movement abnormality in schizophrenia. Acta Psychiatr. Scand. 90, 157–166. doi: 10.1111/j.1600-0447.1994.tb01572.x
Blamey, P., Artieres, F., Baskent, D., Bergeron, F., Beynon, A., Burke, E., et al. (2013). Factors affecting auditory performance of postlinguistically deaf adults using cochlear implants: an update with 2251 patients. Audiol. Neurootol. 18, 36–47. doi: 10.1159/000343189
Bornkessel-Schlesewsky, I., and Schlesewsky, M. (2008). An alternative perspective on “semantic P600” effects in language comprehension. Brain Res. Rev. 59, 55–73. doi: 10.1016/j.brainresrev.2008.05.003
Buchert, R., Berding, G., Wilke, F., Martin, B., von Borczyskowski, D., Mester, J., et al. (2006). IBZM tool: a fully automated expert system for the evaluation of IBZM SPECT studies. Eur. J. Nucl. Med. Mol. Imaging 33, 1073–1083. doi: 10.1007/s00259-006-0067-9
Chen, L. C., Sandmann, P., Thorne, J. D., Bleichner, M. G., and Debener, S. (2016). Cross-modal functional reorganization of visual and auditory cortex in adult cochlear implant users identified with fNIRS. Neural Plast. 2016, 1–13. doi: 10.1155/2016/4382656
Coez, A., Zilbovicius, M., Ferrary, E., Bouccara, D., Isabelle, M., Ambert-Dahan, E., et al. (2008). Cochlear Implant benefits in deafness rehabilitation: PET study of temporal voice activations. J. Nucl. Med. 49, 60–67. doi: 10.2967/jnumed.107.044545
Coez, A., Zilbovicius, M., Ferrary, E., Bouccara, D., Mosnier, I., Ambert-Dahan, E., et al. (2009). Processing of voices in deafness rehabilitation by auditory brainstem implant. Neuroimage 47, 1792–1796. doi: 10.1016/j.neuroimage.2009.05.053
Coez, A., Zilbovicius, M., Ferrary, E., Bouccara, D., Mosnier, I., Ambert-Dahan, E., et al. (2014). Brain voice processing with bilateral cochlear implants: a positron emission tomography study. Eur. Arch. Otorhinolaryngol. 271, 3187–3193. doi: 10.1007/s00405-013-2810-8
Debener, S., Hine, J., Bleeck, S., and Eyles, J. (2008). Source localization of auditory evoked potentials after cochlear implantation. Psychophysiology 45, 20–24. doi: 10.1111/j.1469-8986.2007.00610.x
Delorme, A., and Makeig, S. (2004). EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21. doi: 10.1016/j.jneumeth.2003.10.009
Di Nardo, W., Giannantonio, S., Di Giuda, D., De Corso, E., Schinaia, L., and Paludetti, G. (2013). Role of auditory brain function assessment by SPECT in cochlear implant side selection. Acta Otorhinolaryngol. Ital. 33, 23–28.
Drennan, W. R., and Rubinstein, J. T. (2008). Music perception in cochlear implant users and its relationship with psychophysical capabilities. J. Rehabil. Res. Dev. 45, 779–789. doi: 10.1682/jrrd.2007.08.0118
Finke, M., Büchner, A., Ruigendijk, E., Meyer, M., and Sandmann, P. (2016a). On the relationship between auditory cognition and speech intelligibility in cochlear implant users: an ERP study. Neuropsychologia 87, 169–181. doi: 10.1016/j.neuropsychologia.2016.05.019
Finke, M., Sandmann, P., Bonitz, H., Kral, A., and Buchner, A. (2016b). Consequences of stimulus type on higher-order processing in single-sided deaf cochlear implant users. Audiol. Neurootol. 21, 305–315. doi: 10.1159/000452123
Finke, M., Sandmann, P., Kopp, B., Lenarz, T., and Buchner, A. (2015). Auditory distraction transmitted by a cochlear implant alters allocation of attentional resources. Front. Neurosci. 9:68. doi: 10.3389/fnins.2015.00068
Friederici, A. D., Bahlmann, J., Heim, S., Schubotz, R. I., and Anwander, A. (2006). The brain differentiates human and non-human grammars: functional localization and structural connectivity. Proc. Natl. Acad. Sci. U.S.A. 103, 2458–2463. doi: 10.1073/pnas.0509389103
Friederici, A. D., Hahne, A., and Saddy, J. (2002). Distinct neurophysiological patterns reflecting aspects of syntactic complexity and syntactic repair. J. Psycholinguist. Res. 31, 45–63. doi: 10.1023/a:1014376204525
Friederici, A. D., Meyer, M., and von Cramon, D. Y. (2000). Auditory language comprehension: an event-related fMRI study on the processing of syntactic and lexical information. Brain Lang. 74, 289–300. doi: 10.1006/brln.2000.2313
Gaylor, J. M., Raman, G., Chung, M., Lee, J., Rao, M., Lau, J., et al. (2013). Cochlear implantation in adults: a systematic review and meta-analysis. JAMA Otolaryngol. Head Neck Surg. 139, 265–272. doi: 10.1001/jamaoto.2013.1744
Giraud, A. L., Price, C. J., Graham, J. M., and Frackowiak, R. S. (2001a). Functional plasticity of language-related brain areas after cochlear implantation. Brain 124, 1307–1316. doi: 10.1093/brain/124.7.1307
Giraud, A. L., Price, C. J., Graham, J. M., Truy, E., and Frackowiak, R. S. (2001b). Cross-modal plasticity underpins language recovery after cochlear implantation. Neuron 30, 657–663. doi: 10.1016/s0896-6273(01)00318-x
Giraud, A. L., and Truy, E. (2002). The contribution of visual areas to speech comprehension: a PET study in cochlear implants patients and normal-hearing subjects. Neuropsychologia 40, 1562–1569. doi: 10.1016/s0028-3932(02)00023-4
Giraud, A.-L., Truy, E., Frackowiak, R. S. J., Grégoire, M.-C., Pujol, J.-F., and Collet, L. (2000). Differential recruitment of the speech processing system in healthy subjects and rehabilitated cochlear implant patients. Brain 123, 1391–1402. doi: 10.1093/brain/123.7.1391
Green, K. M. J., Julyan, P. J., Hastings, D. L., and Ramsden, R. T. (2005). Auditory cortical activation and speech perception in cochlear implant users: effects of implant experience and duration of deafness. Hear. Res. 205, 184–192. doi: 10.1016/j.heares.2005.03.016
Green, K. M. J., Julyan, P. J., Hastings, D. L., and Ramsden, R. T. (2011). Cortical activations in sequential bilateral cochlear implant users. Cochlear. Implants Int. 12, 3–9. doi: 10.1179/146701010X486507
Groenen, P. A., Beynon, A. J., Snik, A. F., and van den Broek, P. (2001). Speech-evoked cortical potentials and speech recognition in cochlear implant users. Scand. Audiol. 30, 31–40. doi: 10.1080/010503901750069554
Gungor, H. A., Yildiz, A., Aydin, F., Gungor, F., Boz, A., and Ozkaynak, S. (2005). Tc-99m HMPAO brain SPECT findings in mild and moderate Alzheimer’s disease: correlation with event related potentials. J. Neurol. Sci. 234, 47–53. doi: 10.1016/j.jns.2005.03.034
Hahne, A., Wolf, A., Müller, J., Mürbe, D., and Friederici, A. D. (2012). Sentence comprehension in proficient adult cochlear implant users: on the vulnerability of syntax AU - Hahne. A. Lang. Cogn. Process. 27, 1192–1204. doi: 10.1080/01690965.2011.653251
Henkin, Y., Yaar-Soffer, Y., Steinberg, M., and Muchnik, C. (2014). Neural correlates of auditory-cognitive processing in older adult cochlear implant recipients. Audiol. Neurootol. 19, (Suppl. 1), 21–26. doi: 10.1159/000371602
Heydebrand, G., Hale, S., Potts, L., Gotter, B., and Skinner, M. (2007). Cognitive predictors of improvements in adults’ spoken word recognition six months after cochlear implant activation. Audiol. Neurootol. 12, 254–264. doi: 10.1159/000101473
Hochmair-Desoyer, I., Schulz, E., Moser, L., and Schmidt, M. (1997). The HSM Sentence Test as a tool for evaluating the speech understanding in noise of cochlear implant users. Am. J. Otol. 18(Suppl.):S83.
Holden, L. K., Finley, C. C., Firszt, J. B., Holden, T. A., Brenner, C., Potts, L., et al. (2013). Factors affecting open-set word recognition in adults with cochlear implants. Ear. Hear. 34, 342–360. doi: 10.1097/AUD.0b013e3182741aa7
Huberfeld, G., Habert, M. O., Clemenceau, S., Maksud, P., Baulac, M., and Adam, C. (2006). Ictal brain hyperperfusion contralateral to seizure onset: the SPECT mirror image. Epilepsia 47, 123–133. doi: 10.1111/j.1528-1167.2006.00378.x
Jäncke, L., Wüstenberg, T., Schulze, K., and Heinze, H. J. (2002). Asymmetric hemodynamic responses of the human auditory cortex to monaural and binaural stimulation. Hear. Res. 170, 166–178. doi: 10.1016/s0378-5955(02)00488-4
Jung, T. P., Makeig, S., Humphries, C., Lee, T. W., McKeown, M. J., Iragui, V., et al. (2000a). Removing electroencephalographic artifacts by blind source separation. Psychophysiology 37, 163–178. doi: 10.1111/1469-8986.3720163
Jung, T.-P., Makeig, S., Westerfield, M., Townsend, J., Courchesne, E., and Sejnowski, T. J. (2000b). Removal of eye activity artifacts from visual event-related potentials in normal and clinical subjects. Clin. Neurophysiol. 111, 1745–1758. doi: 10.1016/s1388-2457(00)00386-2
Kallioinen, P., Olofsson, J., Nakeva von Mentzer, C., Lindgren, M., Ors, M., and Sahlen (2016). Semantic processing in deaf and hard-of-hearing children: large n400 mismatch effects in brain responses. Despite poor semantic ability. Front. Psychol. 7:1146. doi: 10.3389/fpsyg.2016.01146
Kolk, H. H., Chwilla, D. J., van Herten, M., and Oor, P. J. (2003). Structure and limited capacity in verbal working memory: a study with event-related potentials. Brain Lang. 85, 1–36. doi: 10.1016/s0093-934x(02)00548-5
Kollmeier, B., and Wesselkamp, M. (1997). Development and evaluation of a German sentence test for objective and subjective speech intelligibility assessment. J. Acoust. Soc. Am. 102, 2412–2421. doi: 10.1121/1.419624
Kutas, M., and Federmeier, K. D. (2011). Thirty years and counting: finding meaning in the N400 component of the event-related brain potential (ERP). Annu. Rev. Psychol. 62, 621–647. doi: 10.1146/annurev.psych.093008.131123
Lazard, D. S., Lee, H. J., Gaebler, M., Kell, C. A., Truy, E., and Giraud, A. L. (2010). Phonological processing in post-lingual deafness and cochlear implant outcome. Neuroimage 49, 3443–3451. doi: 10.1016/j.neuroimage.2009.11.013
Lazard, D. S., Vincent, C., Venail, F., Van de Heyning, P., Truy, E., Sterkers, O., et al. (2012). Pre-, Per- and postoperative factors affecting performance of postlinguistically deaf adults using cochlear implants: a new conceptual model over time. PLoS One 7:e48739. doi: 10.1371/journal.pone.0048739
Lee, H. J., Giraud, A. L., Kang, E., Oh, S. H., Kang, H., Kim, C. S., et al. (2007). Cortical activity at rest predicts cochlear implantation outcome. Cereb. Cortex 17, 909–917. doi: 10.1093/cercor/bhl001
Łukaszewicz-Moszyńska, Z., Lachowska, M., and Niemczyk, K. (2014). Auditory cortical activation and plasticity after cochlear implantation measured by PET using fluorodeoxyglucose. Funct. Neurol. 29, 121–125.
Mamach, M., Wilke, F., Durisin, M., Beger, F. A., Finke, M., Büchner, A., et al. (2018). Feasibility of 15O-water PET studies of auditory system activation during general anesthesia in children. EJNMMI Res. 8:11.
Medved, V., Petrovic, R., Isgum, V., Szirovicza, L., and Hotujac, L. (2001). Similarities in the pattern of regional brain dysfunction in negative schizophrenia and unipolar depression: a single photon emission-computed tomography and auditory evoked potentials study. Prog. Neuropsychopharmacol. Biol. Psychiatry 25, 993–1009. doi: 10.1016/s0278-5846(01)00172-5
Molina, V., Montz, R., Perez-Castejon, M. J., Martin-Loeches, M., Carreras, J. L., Calcedo, A., et al. (1995). Cerebral perfusion, electrical activity and effects of serotonergic treatment in obsessive-compulsive disorder. A preliminary study. Neuropsychobiology 32, 139–148. doi: 10.1159/000119227
Mortensen, M. V., Mirz, F., and Gjedde, A. (2006). Restored speech comprehension linked to activity in left inferior prefrontal and right temporal cortices in postlingual deafness. Neuroimage 31, 842–852. doi: 10.1016/j.neuroimage.2005.12.020
Naatanen, R., and Picton, T. (1987). The N1 wave of the human electric and magnetic response to sound: a review and an analysis of the component structure. Psychophysiology 24, 375–425. doi: 10.1111/j.1469-8986.1987.tb00311.x
Nadol, J. B. Jr. (1997). Patterns of neural degeneration in the human cochlea and auditory nerve: implications for cochlear implantation. Otolaryngol. Head Neck Surg. 117, 220–228. doi: 10.1016/s0194-5998(97)70178-5
O’Mahony, D., Coffey, J., Murphy, J., O’Hare, N., Hamilton, D., Rowan, M., et al. (1996). Event-related potential prolongation in Alzheimer’s disease signifies frontal lobe impairment: evidence from SPECT imaging. J. Gerontol. A Biol. Sci. Med. Sci. 51, M102–M107. doi: 10.1093/gerona/51a.3.m102
Rönnberg, J., Lunner, T., Zekveld, A., Sörqvist, P., Danielsson, H., Lyxell, B., et al. (2013). The ease of language understanding (ELU) model: theoretical, empirical, and clinical advances. Front. Syst. Neurosci. 7:31. doi: 10.3389/fnsys.2013.00031
Sandmann, P., Eichele, T., Buechler, M., Debener, S., Jäncke, L., Dillier, N., et al. (2009). Evaluation of evoked potentials to dyadic tones after cochlear implantation. Brain 132, 1967–1979. doi: 10.1093/brain/awp034
Sandmann, P., Kegel, A., Eichele, T., Dillier, N., Lai, W., Bendixen, A., et al. (2010). Neurophysiological evidence of impaired musical sound perception in cochlear-implant users. Clin. Neurophysiol. 121, 2070–2082. doi: 10.1016/j.clinph.2010.04.032
Sandmann, P., Plotz, K., Hauthal, N., de Vos, M., Schönfeld, R., and Debener, S. (2015). Rapid bilateral improvement in auditory cortex activity in postlingually deafened adults following cochlear implantation. Clin. Neurophysiol. 126, 594–607. doi: 10.1016/j.clinph.2014.06.029
Saur, D., Kreher, B. W., Schnell, S., Kummerer, D., Kellmeyer, P., Vry, M. S., et al. (2008). Ventral and dorsal pathways for language. Proc. Natl. Acad. Sci. U.S.A. 105, 18035–18040. doi: 10.1073/pnas.0805234105
Saur, D., Schelter, B., Schnell, S., Kratochvil, D., Küpper, H., Kellmeyer, P., et al. (2010). Combining functional and anatomical connectivity reveals brain networks for auditory language comprehension. Neuroimage 49, 3187–3197. doi: 10.1016/j.neuroimage.2009.11.009
Schierholz, I., Finke, M., Kral, A., Buchner, A., Rach, S., Lenarz, T., et al. (2017). Auditory and audio-visual processing in patients with cochlear, auditory brainstem, and auditory midbrain implants: an EEG study. Hum. Brain Mapp. 38, 2206–2225. doi: 10.1002/hbm.23515
Schierholz, I., Finke, M., Schulte, S., Hauthal, N., Kantzke, C., Rach, S., et al. (2015). Enhanced audio–visual interactions in the auditory cortex of elderly cochlear-implant users. Hear. Res. 328, 133–147. doi: 10.1016/j.heares.2015.08.009
Service, E., Helenius, P., Maury, S., and Salmelin, R. (2007). Localization of syntactic and semantic brain responses using magnetoencephalography. J. Cogn. Neurosci. 19, 1193–1205. doi: 10.1162/jocn.2007.19.7.1193
Shajahan, P. M., Glabus, M. F., Blackwood, D. H., and Ebmeier, K. P. (1997). Brain activation during an auditory ‘oddball task’ in schizophrenia measured by single photon emission tomography. Psychol. Med. 27, 587–594. doi: 10.1017/s0033291797004728
Shen, W., Fiori-Duharcourt, N., and Isel, F. (2016). Functional significance of the semantic P600: evidence from the event-related brain potential source localization. Neuroreport 27, 548–558. doi: 10.1097/WNR.0000000000000583
Signorini, M., Paulesu, E., Friston, K., Perani, D., Colleluori, A., Lucignani, G., et al. (1999). Rapid assessment of regional cerebral metabolic abnormalities in single subjects with quantitative and nonquantitative [18F]FDG PET: a clinical validation of statistical parametric mapping. Neuroimage 9, 63–80. doi: 10.1006/nimg.1998.0381
Sorqvist, P., Ljungberg, J. K., and Ljung, R. (2010). A sub-process view of working memory capacity: evidence from effects of speech on prose memory. Memory 18, 310–326. doi: 10.1080/09658211003601530
Soshi, T., Hisanaga, S., Kodama, N., Kanekama, Y., Samejima, Y., Yumoto, E., et al. (2014). Event-related potentials for better speech perception in noise by cochlear implant users. Hear. Res. 316, 110–121. doi: 10.1016/j.heares.2014.08.001
Stefanatos, G. A., Joe, W. Q., Aguirre, G. K., Detre, J. A., and Wetmore, G. (2008). Activation of human auditory cortex during speech perception: effects of monaural, binaural, and dichotic presentation. Neuropsychologia 46, 301–315. doi: 10.1016/j.neuropsychologia.2007.07.008
Strelnikov, K., Marx, M., Lagleyre, S., Fraysse, B., Deguine, O., and Barone, P. (2015a). PET-imaging of brain plasticity after cochlear implantation. Hear. Res. 322, 180–187. doi: 10.1016/j.heares.2014.10.001
Strelnikov, K., Rouger, J., Lagleyre, S., Fraysse, B., Démonet, J. F., Déguine, O., et al. (2015b). Increased audiovisual integration in cochlear-implanted deaf patients: independent components analysis of longitudinal positron emission tomography data. Eur. J. Neurosci. 41, 677–685. doi: 10.1111/ejn.12827
Strelnikov, K., Rouger, J., Demonet, J. F., Lagleyre, S., Fraysse, B., Deguine, O., et al. (2013). Visual activity predicts auditory recovery from deafness after adult cochlear implantation. Brain 136, 3682–3695. doi: 10.1093/brain/awt274
Suh, M. W., Park, K. T., Lee, H. J., Lee, J. H., Chang, S. O., and Oh, S. H. (2015). Factors contributing to speech performance in elderly cochlear implanted patients: an FDG-PET study: a preliminary study. J. Int. Adv. Otol. 11, 98–103. doi: 10.5152/iao.2015.424
Tobey, E. A., Devous, M. D. Sr., Buckley, K., Cooper, W. B., Harris, T. S., Ringe, W., et al. (2004). Functional brain imaging as an objective measure of speech perception performance in adult cochlear implant users. Int. J. Audiol. 43, (Suppl. 1), S52–S56.
Turgeon, C., Lazzouni, L., Lepore, F., and Ellemberg, D. (2014). An objective auditory measure toassess speech recognition in adult cochlear implant users. Clin. Neurophysiol. 125, 827–835. doi: 10.1016/j.clinph.2013.09.035
Tzourio-Mazoyer, N., Landeau, B., Papathanassiou, D., Crivello, F., Etard, O., Delcroix, N., et al. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage 15, 273–289. doi: 10.1006/nimg.2001.0978
Van Petten, C., Coulson, S., Rubin, S., Plante, E., and Parks, M. (1999). Time course of word identification and semantic integration in spoken language. J. Exp. Psychol. Learn. Mem. Cogn. 25, 394–417. doi: 10.1037//0278-73188.8.131.524
Vavatzanidis, N. K., Mürbe, D., Friederici, A. D., and Hahne, A. (2018). Establishing a mental Lexicon with cochlear implants: an ERP study with young children. Sci. Rep. 8:910. doi: 10.1038/s41598-017-18852-3
Wong, D., Pisoni, D. B., Learn, J., Gandour, J. T., Miyamoto, R. T., and Hutchins, G. D. (2002). PET imaging of differential cortical activation by monaural speech and nonspeech stimuli. Hear. Res. 166, 9–23. doi: 10.1016/s0378-5955(02)00311-8
Zekveld, A. A., Rudner, M., Johnsrude, I. S., Heslenfeld, D. J., and Rönnberg, J. (2012). Behavioral and fMRI evidence that cognitive ability modulates the effect of semantic context on speech intelligibility. Brain Lang. 122, 103–113. doi: 10.1016/j.bandl.2012.05.006
Keywords: single-photon emission computed tomography, electroencephalography, cochlear-implant, memory, N400, speech processing
Citation: Kessler M, Schierholz I, Mamach M, Wilke F, Hahne A, Büchner A, Geworski L, Bengel FM, Sandmann P and Berding G (2020) Combined Brain-Perfusion SPECT and EEG Measurements Suggest Distinct Strategies for Speech Comprehension in CI Users With Higher and Lower Performance. Front. Neurosci. 14:787. doi: 10.3389/fnins.2020.00787
Received: 10 April 2020; Accepted: 06 July 2020;
Published: 11 August 2020.
Edited by:Claude Alain, Rotman Research Institute (RRI), Canada
Reviewed by:Andrew Dimitrijevic, Sunnybrook Health Sciences Centre, Canada
Iiro P. Jääskeläinen, Aalto University, Finland
Copyright © 2020 Kessler, Schierholz, Mamach, Wilke, Hahne, Büchner, Geworski, Bengel, Sandmann and Berding. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Georg Berding, firstname.lastname@example.org
†These authors have contributed equally to this work