Skip to main content

BRIEF RESEARCH REPORT article

Front. Hum. Neurosci., 14 September 2023
Sec. Speech and Language
Volume 17 - 2023 | https://doi.org/10.3389/fnhum.2023.1122480

Examining cortical tracking of the speech envelope in post-stroke aphasia

  • 1Center for Education in Health Sciences, Northwestern University Feinberg School of Medicine, Chicago, IL, United States
  • 2Department of Communication Sciences and Disorders, University of Wisconsin-Madison, Madison, WI, United States
  • 3VA Pittsburgh Healthcare System, Pittsburgh, PA, United States
  • 4Department of Communication Sciences and Disorders, University of Pittsburgh, Pittsburgh, PA, United States
  • 5Roxelyn and Richard Pepper Department of Communication Science and Disorders, School of Communication. Northwestern University, Evanston, IL, United States

Introduction: People with aphasia have been shown to benefit from rhythmic elements for language production during aphasia rehabilitation. However, it is unknown whether rhythmic processing is associated with such benefits. Cortical tracking of the speech envelope (CTenv) may provide a measure of encoding of speech rhythmic properties and serve as a predictor of candidacy for rhythm-based aphasia interventions.

Methods: Electroencephalography was used to capture electrophysiological responses while Spanish speakers with aphasia (n = 9) listened to a continuous speech narrative (audiobook). The Temporal Response Function was used to estimate CTenv in the delta (associated with word- and phrase-level properties), theta (syllable-level properties), and alpha bands (attention-related properties). CTenv estimates were used to predict aphasia severity, performance in rhythmic perception and production tasks, and treatment response in a sentence-level rhythm-based intervention.

Results: CTenv in delta and theta, but not alpha, predicted aphasia severity. Neither CTenv in delta, alpha, or theta bands predicted performance in rhythmic perception or production tasks. Some evidence supported that CTenv in theta could predict sentence-level learning in aphasia, but alpha and delta did not.

Conclusion: CTenv of the syllable-level properties was relatively preserved in individuals with less language impairment. In contrast, higher encoding of word- and phrase-level properties was relatively impaired and was predictive of more severe language impairments. CTenv and treatment response to sentence-level rhythm-based interventions need to be further investigated.

Introduction

Aphasia is a communication disorder resulting from damage to brain regions that support language. Due to their communication impairments, people with aphasia (PWA) experience social isolation, loss of autonomy, and other consequences that affect their quality of life (Cruice et al., 2006; Hilari, 2011). Intervention approaches leveraging rhythmic aspects of speech and music have been shown to be beneficial in improving language production in PWA (e.g., Stahl et al., 2011, 2013). However, little is known about PWA’s encoding of speech rhythmic properties and their connection to language impairments, rhythmic perception/production, and treatment response to rhythm-based interventions.

The amplitude envelope conveys rhythmic cues that are important for speech perception (Rosen, 1992; Shannon et al., 1995). The neural processing of these cues in continuous speech can be assessed by examining cortical tracking using electroencephalography (EEG). Specifically, cortical tracking of the speech envelope (CTenv) can be used as a proxy to assess the neural processing of rhythmic properties in continuous speech, but has not been previously examined in post-stroke aphasia. CTenv refers specifically to the tracking of the acoustic/linguistic properties of the speech signal by the low-frequency cortical activity in the M/EEG (Ding and Simon, 2014). The cortical activity in the M/EEG reflects activity of populations of neurons (Lakatos et al., 2019; Gnanateja et al., 2022) – essential for processing rhythm cues in speech and music (Peelle et al., 2013; Doelling et al., 2019; Poeppel and Assaneo, 2020; Gnanateja et al., 2022). CTenv could thus open a window into PWA’s encoding of speech rhythmic properties, which could shed light on treatment response to rhythm-based interventions.

CTenv has been examined in the delta, theta, and alpha bands of neural oscillations. The delta band (~1–4 Hz) aligns with the word and phrase boundaries in connected speech and prosodic cues, and has been associated with higher-order language processing (Ghitza, 2017; Teoh et al., 2019). The theta band (∼4–8 Hz) broadly corresponds to the syllabic rate (58 Hz) found in multiple languages (Ding et al., 2017) and is key in segmenting auditory information into syllabic chunks (Ghitza, 2013; Kösem and van Wassenhove, 2016; Poeppel and Assaneo, 2020). Activity in this band has been shown to be stronger for intelligible speech (Peelle et al., 2013) than for unintelligible speech (Pefkou et al., 2017); or equal in both types of stimuli (Howard and Poeppel, 2010). The alpha band (~8–12 Hz) is associated with attentional processes (Obleser and Weisz, 2012; Wöstmann et al., 2017) and effortful listening (Paul et al., 2021).

CTenv in these bands has been used to investigate neurophysiological mechanisms underlying speech, language, and music processing in neurotypical populations (e.g., Doelling et al., 2014), including in Spanish speakers (e.g., Assaneo et al., 2019; Pérez et al., 2019; Lizarazu et al., 2021). These neurophysiological mechanisms could inform our understanding of how the encoding of rhythmic properties of speech are affected by aphasia and how they may influence treatment response to rhythm-based interventions. Therefore, the current study examines the association between CTenv and aphasia severity, rhythmic perception/production abilities, and treatment response to a sentence-level rhythm-based intervention in post-stroke aphasia.

Recent evidence from primary progressive aphasia (PPA; Dial et al., 2021) demonstrated an increased CTenv in theta but not delta compared to neurotypicals, implying that people with PPA may rely more on acoustic properties during continuous speech processing. To date, this is one of the few studies that investigate CTenv in adult neurogenic language disorders (see De Clercq et al., 2023; Kries et al., 2023). Thus, the association between CTenv and language impairments as measured by aphasia severity is, to our knowledge, yet to be examined.

CTenv may also be associated with rhythmic perception and production measures in individuals with aphasia. Some PWA show difficulties in rhythmic perception tasks (Zipse et al., 2014; Sihvonen et al., 2016), such as identifying whether two excerpts with rhythmic variations are the same or different. Some others show more difficulties than controls in rhythmic production tasks, such as continuing a rhythmic pattern after its offset – with some PWA showing preserved abilities in synchronizing to a given rhythm (Zipse et al., 2014). Importantly, a better understanding of rhythmic perception and production in aphasia can guide language treatment candidacy, especially for rhythm-based interventions. For example, Curtis et al. (2020) reported that PWA with better rhythmic abilities were more likely to benefit from an additional rhythmic cue (hand tapping) in response to melodic intonation therapy. Similarly, Stefaniak et al. (2021) found a correlation between the perception of rhythmic patterns and picture description fluency in PWA. The association between rhythmic measures and CTenv could help to understand rhythmic performance in aphasia in ways that can support candidacy and treatment response.

Rhythmic elements have been successfully employed in aphasia rehabilitation. For example, Stahl et al. (2011) examined how PWA learned lyrics using three conditions: melodic intoning, rhythmic speech, and a control condition, and reported a comparable degree of benefit between the melodic and rhythmic conditions. A follow-up study by Stahl et al. (2013) showed improved sentence learning using singing and rhythmic therapy. Also, simultaneous production of audiovisual speech, which depends on the synchronization of speech rhythmic properties between a patient and clinician, has resulted in improved sentence learning in PWA (Fridriksson et al., 2012, 2015), which is consistent with previous work on neurotypicals showing that audiovisual speech enhances CTenv and improves speech perception (Chandrasekaran et al., 2009; Crosse et al., 2015). However, it is currently unknown whether the neural encoding of rhythmic properties in connected speech (as reflected in CTenv) is predictive of treatment response in this type of rhythm-based interventions. The current study, therefore, provides a preliminary examination of the relationship between CTenv and treatment response in a rhythm-based treatment described by Quique et al. (2022).

In short, the purpose of this study was threefold. First, we examined CTenv in Spanish-speaking PWA and assessed the extent to which CTenv in the delta, theta, and alpha bands was associated with aphasia severity. Second, we examined if CTenv in those bands was related to behavioral measures of rhythmic perception and production in PWA. Third, we assessed if CTenv in those bands predicted treatment response to the sentence-level rhythm-based intervention previously reported by Quique et al. (2022).

Methods

Participants

Nine Spanish-speaking PWA were recruited from an outpatient neurology institute in Medellín, Colombia. These participants were a subset of the Quique et al. (2022) participants, selected because they lived in Medellin, where the necessary EEG instrumentation was located. Inclusionary criteria consisted of aphasia after a single left hemisphere cerebrovascular accident confirmed by neurology referral, a Western Aphasia Battery aphasia quotient between 20 and 85 (WAB-AQ, score range: [0–100]; Kertesz, 1982; Kertesz and Pascual-Leone, 2000), an aphasia post-onset greater than 4 months, and Spanish as a native language. Exclusion criteria consisted of other neurological conditions that could potentially affect speech or cognition, significant premorbid psychiatric history, and severe apraxia of speech. Nine participants were enrolled in the study (Table 1). All participants completed a rhythm-based intervention to learn scripted sentences after their EEG assessment (Quique et al., 2022). Thus, CTenv was assessed at one timepoint, before the rhythm-based intervention. All participants were able to hear pure tones at 40 dB HL at 0.5, 1, 2, and 4 kHz on an audiometric screening test. This study was approved by the institutional review board at the University of Pittsburgh and by the local partner institution in Colombia (Neuromedica).

TABLE 1
www.frontiersin.org

Table 1. Participants’ demographic information.

Sentence-level rhythm-based intervention

Quique et al. (2022) investigated the extent to which unison production of speech could support PWA in learning scripted sentences (i.e., formulaic sentences used in everyday communication). As unison production of speech depends on rhythmic properties, it was hypothesized that highlighting these properties could facilitate learning of scripted sentences. Specifically, Quique and colleagues used rhythmic cues during unison production of speech in which: (a) digitally-added beats were aligned with word stress (stress-aligned condition); (b) digitally-added beats were equally spaced across the sentence (metronomic condition); (c) sentences were presented without added beats (control condition). PWA learned a set of 30 scripted sentences over five training sessions distributed over 2 weeks. As shown in Table 1, participants were between 1 and 3.5 years post-onset of aphasia. PWA demonstrated learning of scripted sentences over time across conditions. Furthermore, both rhythm-enhanced conditions engendered larger treatment response (i.e., improved sentence-level learning) than the control condition.

Stimuli

A 9-min excerpt of a Spanish adaptation of the audiobook “Who was Albert Einstein” (Brallier, 2002) was used to assess CTenv. The excerpt was recorded by a human male native speaker of Spanish. The excerpt was divided into approximately 1-min segments, each beginning and ending with complete sentences to preserve the storyline. To ensure attention, participants answered a comprehension question after each segment, as previously used in PPA (Dial et al., 2021). See Supplementary material for an example of a story segment. The envelope spectrum of the excerpt was 6.3 Hz, which is closely related to the syllable rate (Obleser and Weisz, 2012; Ghitza, 2013; Kösem and van Wassenhove, 2016). See modulation spectrum in Supplementary Figure S1. Also, to validate the recording of auditory electrophysiology responses and ensure that auditory stimuli reached the auditory cortex, we recorded Auditory Long Latency Responses (ALLRs; Swink and Stuart, 2012; Reetzke et al., 2021; See Figure 1). For this purpose, we used a sinusoidally modulated (6 Hz) tone of carrier frequency of 1,000 Hz and another unmodulated pure tone of frequency of 1,000 Hz. Both tones were 1 second in duration and were repeated 100 times, with an inter-stimulus interval of 1 second.

FIGURE 1
www.frontiersin.org

Figure 1. Auditory Long Latency Responses (ALLRs, top) and scalp topography (bottom) of the first negative wave (N1) in each participant.

Instrumentation

A USB audio interface (Behringer U-PhoriaUMC404HD) was used to present the auditory stimuli. Electroencephalography (EEG) was performed using a NihonKohden 32channel clinical EEG system. Nineteen scalp electrodes were placed on the scalp following the international 1020 system (Klem et al., 1999), along with two mastoid electrodes. The online reference was placed at Cz. Electrode impedance was under 5 kΩ for most electrodes, and the maximum impedance did not exceed 20 kΩ. The sampling rate for EEG recording was 5,000 Hz. The data were re-referenced offline to the average of both mastoids. The analog electrical audio signal from one of the audio interface outputs was directly plugged into one of the leads on the EEG amplifier (with appropriate attenuation) to align the EEG to the stimulus onset. Further, as the EEG system was used for clinical purposes, ALLRs were recorded to ensure that the instrumentation was optimal. Specifically, if the instrument was sensitive to record reliable ALLRs in the latency range of 50 to 250 ms that were higher than the pre-stimulus baseline and had a morphology and topography consistent with ALLRs.

Procedure

Participants were first presented with the ALLRs stimuli and then listened to the audiobook. Stimuli were presented binaurally with insert earphones (ER-3A; Etymotic Research) at 84 dB SPL (Becker and Reinvang, 2007; Robson et al., 2014) using PsychoPy (Peirce et al., 2019). During ALLRs, participants were instructed to listen to the tones without any additional task. During the audiobook, participants were instructed to listen to the story and answer the multiple-choice questions. They were also asked to avoid extraneous movements and look at a fixation cross on the computer screen.

EEG processing

Preprocessing was performed using EEGLAB (Delorme and Makeig, 2004) in MATLAB (MathWorks Inc., 2019). To reduce data size and improve computational efficiency, we first down-sampled the data to 128 Hz. Then, we applied minimum-phase non-causal Sinc FIR filters with a high pass cut-off frequency of 1 Hz, filter order of 846; and a low-pass cut-off frequency of 15 Hz, filter order of 212. Filtered data were then re-referenced to the two mastoid channels (Di Liberto et al., 2015; O’Sullivan et al., 2017). Large movement artifacts were suppressed using Artifact Subspace Reconstruction (ASR; Mullen et al., 2013; Plechawska-Wojcik et al., 2019). After ASR, the data were separated into epochs from −5 s to 70 s (i.e., the onset of each 1-min segment), resulting in 9 epochs. Independent Component Analysis was performed on the epoched data to remove eye movement and muscle artifacts (Jung et al., 2001). The independent components were then inspected visually for time course, topography, and spectrum and were removed if they corresponded to ocular or muscular activity.

To extract the theta, and alpha bands from the EEG, the data was further filtered using third-order butterworth filters; theta (lf = 4 Hz; hf = 8 Hz), and alpha (lf = 8 Hz; hf = 15 Hz). For the delta however, a different preprocessing procedure was followed, where the high pass filter was changed. The raw EEG data was filtered using a windowed sinc filters (lf = 0.75, order 846; hf = 15, order 212). This data was subjected to the same preprocessing steps until ICA. The ICA weights that were derived from the 1–15 hz filtered data were copied over to 0.75–15 Hz filtered data (Barbieri et al., 2021). The ICA components with artifacts were then rejected. This was done to ensure good ICA decomposition, as highpass filtering with 1 Hz provides better stationarity in the EEG (Winkler et al., 2015). Finally, to extract the delta band, this data was low pass filtered at 4 Hz with a third-order butterworth filter which essentially resulted in a 0.75–4 Hz filter band. The magnitude response of the filters and filter characteristics used to derive the three different EEG bands are shown in Supplementary Figure S2.

EEG analysis

Preprocessed EEG was analyzed using custom routines in Matlab and the multivariate Temporal Response Function toolbox (mTRF; Crosse et al., 2016). We used mTRFs with forward modeling to estimate how well the temporal envelope predicted EEG activity; this provided an individual measure of CTenv (McHaney et al., 2021). First, the multiband Hilbert transform was used to estimate the temporal envelopes of each 1-min speech segment. Second, using the preprocessed EEG data and the temporal envelopes, the mTRFs were estimated via regularized linear ridge regression (see Figure 2 and Supplementary Figure S3). mTRFs were estimated for lags ranging from −100 ms to 650 ms. To reduce overfitting and obtain the best model fit across electrodes and participants, a regularization parameter optimized from 20 to 220 using k-fold (leave-one-out) cross-validation was applied (Crosse et al., 2016). Model fit was assessed using Pearson’s correlation between the predicted EEG (stimulus envelope convolved with mTRFs) and the observed EEG data, which was considered the CTenv estimate. To test the robustness of the obtained CTenv estimates, we compared them against random CTenv estimates generated from one hundred mismatched permutations of EEG responses and the audiobook envelopes. Comparing the observed CTenv against the random CTenv provides evidence regarding the association between cortical activity due to specific auditory input in the speech envelope rather than a chance response. As CTenv can be affected by the cortical lesions and related volume conducted electrical field at the electrodes, only the five electrodes that had the highest CTenv per participant were averaged to obtain one metric of cortical tracking.

FIGURE 2
www.frontiersin.org

Figure 2. Schematic representation of the Temporal Response Function (TRF), used to estimate cortical tracking of the speech envelope (CTenv; mapping of speech envelope to the neural data). A linear kernel to map the speech envelope of the audiobook onto the EEG is estimated using ridge regression. The cross-validated model fit is estimated using Pearson’s correlation between the EEG-predicted data and the observed EEG, which serves as a metric of CT.

Behavioral measures

To characterize language performance, participants were given the Spanish adaptation of the Western Aphasia Battery (WAB; Kertesz, 1982; Kertesz and Pascual-Leone, 2000). To characterize rhythmic perception and production, participants were given the computerized version of the Beat Alignment Test (CA-BAT; Harrison and Müllensiefen, 2018) and the Synchronization with the Beat of Music Task from the Beat Alignment Test (BAT-sync; Iversen and Patel, 2008). In the CA-BAT, participants listen to musical excerpts presented twice: once with superimposed beats arranged in synchrony with the excerpt and once with beats either too fast or too slow in relation to the excerpt. Participants are asked to indicate in which of the two presentations the beats were “on the beat” with the excerpt. In the BAT-sync, participants are asked to listen and tap with their left hand to the beat of 12 musical excerpts with various musical styles. The BAT-sync provides a measure of synchronization (ranging between 0 and 1) based on the correlation between intertap means (average tapping intervals for each of the 12 excerpts) and tactus levels of the underlying beat of each excerpt (most likely to occur tapping tempos).

Variables and statistical analysis

Statistical analyzes were performed in Rstudio (RStudio Team, 2016; R Core Team, 2020). We used a robust bayesian estimation to assess if the observed CTenv differed from an empirical chance distribution (see Kruschke, 2013). To examine the three study aims, we implemented bayesian generalized linear regression models using Stan (Carpenter et al., 2017; Stan Development Team, 2020) with the R brms package (Bürkner, 2018). This modeling approach is advantageous for the exploratory nature of the current study as it characterizes uncertainty in the estimates of interest through credible intervals (Kruschke and Liddell, 2018), rather than p-values, and quantifies whether the data support hypothesized effects or null findings.

The association between CTenv and aphasia severity was assessed using WAB-AQ as the dependent variable and CTenv estimates in the alpha, delta, and theta bands as predictors. The association between CTenv and rhythmic perception/production measures was assessed using two models, one for each rhythmic measure (CA-BAT and BAT-sync), with CTenv estimates in the alpha, delta, and theta bands as predictors. The association between CTenv and treatment response to the rhythm-based intervention from Quique et al. (2022) was assessed using mixed-effects models to take advantage of the available trial-level treatment data. Specifically, trial-level word accuracy (i.e., the proportion of correct words out of the total number of words in each sentence, at each session). was used as a dependent variable. Fixed effects were session and CTenv estimates. Random effects included random intercepts for participants and items, and a random slope of session for participants. Models were run separately for the rhythm-enhanced and control conditions. Details about each model can be found in Supplementary materials.

Results

Nine PWA were enrolled and completed the study. All participants showed ALLR peaks, indicating that: (1) the EEG setup was appropriate to record neural responses evoked by auditory signals, and (2) the auditory neural centers in all the participants received afferent input regarding sound onset information. See Figure 1 for a depiction of ALLRs per participant. Thus, we assessed CTenv in all the participants. The strength of CTenv estimates was consistent with previous reports (Di Liberto et al., 2015). Based on 90% credible intervals, the observed CTenv estimates were reliably higher than the randomly generated CTenv estimates across bands and participants (ß = −0.03, CI: −0.03, −0.02), indicating an association between cortical processing and a specific auditory input (i.e., the speech envelope). See Figure 3 for a depiction of observed versus random CTenv measures per participant and band.

FIGURE 3
www.frontiersin.org

Figure 3. Observed (averaged across the five best electrodes) CTenv (blue) and CTenv to null/random model (red; based on 100 random pairings of the speech envelope and EEG) in each participant for the delta, theta, and alpha band, respectively. Error bars represent 95% confidence intervals of chance distribution. The scalp topography of CTenv estimates (Pearson’s correlation) in the delta, theta, and alpha bands, are shown at the bottom of each panel for each participant.

Association between CTenv and aphasia severity

We assessed whether CTenv measures could be used as a metric to predict aphasia severity. CTenv in the theta band was a reliable predictor of aphasia severity (ß = 9.07, CI: 1.94, 15.29), suggesting that increased tracking in this band is associated with an increase in WAB-AQ (less impairment). The credible interval for CTenv in the delta band included zero (ß = −8.06, CI: −16.27, 1.48), and so did not meet the criteria we defined as a robust effect, but 93% of the posterior distribution for this effect was below zero. This provides weak but positive evidence that increased tracking in the delta band could be associated with a decrease in WAB-AQ (more impairment). CTenv in the alpha band was not a reliable predictor of aphasia severity (ß = 0.07, SE = CI: −8.26, 9.25). See Figure 4. See also Supplementary Figure S4 in our Supplementary materials.

FIGURE 4
www.frontiersin.org

Figure 4. (A) Shows scatterplots of the association between CTenv and aphasia severity. (B) Shows Scatterplots of the association between CTenv and CA-BAT. (C) Shows Scatterplots of the association between CTenv and BAT-sync. Black dots = raw data. Blue line = estimated slope from models reported in Supplementary Tables S1–S3. Gray area = 90% model uncertainty intervals. (D,E) Show the posterior distribution and 90% highest density intervals (HDIs) of the interaction effects between the theta band and session for the rhythm enhanced and control conditions. Dashed lines mark the 90% highest density intervals (HDIs) for the posterior distribution.

Association between cortical tracking and measures of rhythmic perception and production

CTenv did not reliably predict CA-BAT performance (delta: ß = 0.39, CI: −0.84, 1.63; theta: ß = −0.01, CI: −0.91, 0.87; alpha: ß = −0.29, CI: −1.53, 0.99). Similarly, CTenv did not reliably predict BAT-sync performance (delta: ß = −0.07, CI: −0.20, 0.06; theta: ß = 0.04, CI: −0.04, 0.13; alpha: ß = −0.05, CI: −0.18, 0.08). These results indicate little association between rhythm perception/production (as measured by CA-BAT and BAT-sync) with CTenv in PWA. See Figure 4. See also Supplementary Figures S5, S6 in our Supplementary materials.

Association between CTenv and a rhythm-based intervention

CTenv in the theta band showed a reliable main effect for the rhythm-enhanced conditions (ß = 1.06, CI: 0.48, 1.66), and a similar effect for the control condition (ß = 0.80, CI: 0.01, 1.59). Thus, these results indicate that the odds of producing a correct word, at the midpoint of treatment, were greater in people with higher CTenv in the theta-band. Consistent with Quique et al. (2022), within this sub-group of participants from Quique et al. (2022), there was also a reliable main effect of session for all models across conditions and bands (all estimates were positive, and the CI excluded zero), indicating a positive treatment response over time. Although the interaction effect between CTenv in theta and session for the rhythm-enhanced (ß = −0.12, CI: −0.27, 0.02) and control conditions (ß = −0.06, CI: −0.18, 0.05) included zero, examining the posterior distributions showed indications of a positive effect. Specifically, 93% of the posterior distribution was below zero for the rhythm-enhanced, and 83% for the control conditions (See Figure 4). This result provides weak but positive evidence that a higher CTenv in theta may be associated with a reduced treatment response to the sentence-level rhythm-based intervention. There were no other main effects or interaction effects for CTenv in alpha or delta bands (CI included zero with no substantial portions of their posterior distribution suggesting meaningful effects). See Supplementary materials for complete model output for all models (Supplementary Figures S7–S16).

Discussion

The current study examined CTenv in Spanish speakers with aphasia and its association with language impairments, behavioral measures of rhythmic perception/production, and treatment response to a rhythm-based intervention. CTenv in the theta band was associated with language impairments in our cohort of Spanish speakers with stroke-induced aphasia. We also found some evidence that CTenv in the delta band may be associated with language impairments. These preliminary findings may suggest that cortical activity in the theta and delta bands, which have been shown to be essential for speech processing in neurotypical individuals (Poeppel and Assaneo, 2020), could be affected by acquired brain lesions.

The direction of the relationship between language impairments and CTenv in the delta and theta bands had intriguing patterns. Higher CTenv in the delta band was likely associated with more-severe language impairments, whereas higher CTenv in theta band was associated with less-severe language impairments. Although this opposite directionality could indicate an inverse relationship between delta and theta rhythms to compensate for language impairments in aphasia, our results do not provide conclusive evidence. The association between higher CTenv in the delta band and more severe language impairments could reflect a compensatory mechanism to support language comprehension. Such increases in CTenv and decreased speech comprehension have been reported in older adults, suggesting that increased tracking could indicate altered speech processing abilities (e.g., Decruy et al., 2019). McHaney et al. (2021) reported an association between increased delta tracking and speech-in-noise comprehension in older adults (for speech in noise compared to quiet). However, other studies have shown that increased CTenv may reflect, instead, an imbalance between excitatory and inhibitory brain processes or the recruitment of additional brain regions for processing information (e.g., Brodbeck et al., 2018; Zan et al., 2020).

The association between CTenv in the theta band and language impairments could suggest that the higher CTenv, the more appropriately PWA can segment syllabic information (Teng et al., 2018). However, increased tracking in the theta band has been observed in individuals with logopenic PPA, who typically show phonological processing deficits, compared to neurotypical older adults (Dial et al., 2021). Dial et al. (2021) suggested that increased tracking in PPA could result from a shift of function from atrophic temporoparietal cortex to spared brain regions in both hemispheres; it could be further affected by increased functional connectivity in the temporal cortices and hypersynchrony in the frontal cortex (Ranasinghe et al., 2017). Such a shift of function could also be seen in brain damage due to stroke; however, the pathology differs from PPA. Further, Etard and Reichenbach (2019) reported that increased tracking in the theta band was associated with perceived speech clarity in young adults. Given the lack of evidence in CTenv and aphasia, or other acquired neurological disorders, these interpretations need to be investigated in future studies. The exact mechanism driving the relationship between cortical tracking and aphasia severity needs further large-scale intervention across a wide range of aphasia severities using spatially and temporally precise neuroimaging approaches.

None of the CTenv bands were reliable predictors of behavioral measures of rhythm perception (i.e., CA-BAT, BAT-sync); this lack of association is inconsistent with previous findings. For example, Nozaradan et al. (2011) found that in neurotypical participants, beat perception was associated with a sustained periodic response in the EEG spectrum. Similarly, Nozaradan et al. (2013) showed links between neural responses, listening to rhythmic sound patterns, and tapping to beats. Further, Colling et al. (2017) showed that CTenv and beat synchronization to a metronome were affected in individuals with dyslexia. Although we did not simultaneously use EEG with behavioral measures of rhythm perception/production, given previous evidence, an association would have been expected. Thus, more research is needed to understand associations between CTenv and measures of rhythm perception/production in acquired neurological disorders.

The preliminary interaction effect between CTenv in theta and session indicates that there may be an association between CTenv and treatment response to a sentence-level rhythm-based intervention, but this conclusion would require additional examination and replication. Rhythm-based interventions have been recommended in language disorders such as dyslexia (e.g., Goswami, 2011; Thomson et al., 2013), which has been linked to concomitant rhythmic difficulties (e.g., Wolff, 2002; Flaugnacco et al., 2014; Ladányi et al., 2020). Similarly, individuals with aphasia have shown rhythm processing deficits (Zipse et al., 2014; Sihvonen et al., 2016), but have also benefited from rhythm-based interventions (e.g., Stahl et al., 2011, 2013; Quique et al., 2022). Further research can continue to examine whether encoding of speech rhythmic properties could provide neurophysiologically informed decision-making regarding the choice of intervention in individuals with aphasia.

Limitations and future work

The current study had some limitations. First, it had a limited sample size of PWA; thus, reproducing our findings in future studies with larger sample sizes would be needed to provide robust evidence. Second, there were limited observations within each variable available for modeling (e.g., for aphasia severity and rhythm perception/production, we had one observation per participant). Thus, replicating these findings in larger samples with multiple observations per participant is necessary. Third, because of the cortical lesions in our participants with aphasia, we did not select electrodes of interest a priori; instead, we used the averaged CTenv of the five electrodes with the highest values. However, future studies should acquire anatomical data and perform source-level analysis to avoid confounds of volume conduction artifacts introduced by the brain lesion. Similarly, we believe that anatomical changes in the cortex after stroke may have an influence on CTenv, but we did not have lesion data for these participants; this information is needed to investigate the role of lesion size and location on CTenv in aphasia. Fourth, our results could not provide a clear mechanistic explanation for the directionality of the relationship between CTenv and aphasia severity. Future studies are necessary to understand the neurophysiological mechanisms underlying such findings. Fifth, CTenv was assessed only at one timepoint (before the rhythm-based intervention). Future studies using such neural metrics before and after intervention could provide objective neurophysiological evidence that can guide interventions, predict outcomes, and measure the mechanisms of treatment-related improvements. Overall, CTenv in the theta band was associated with language impairments. CTenv in theta may be associated with treatment response to sentence-level rhythm-based interventions, but further research is needed to provide robust evidence.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were reviewed and approved by the Institutional Review Board at the University of Pittsburgh and by Neuromedica, the local partner institution in Colombia. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

Data collection, analysis, and background literature review were conducted by YQ, GNG, WE, and MD. Data analyses and interpretation were conducted by YQ and GNG. Manuscript preparation was conducted by YQ, GNG, WE, MD, and BC. All authors contributed to the article and approved the submitted version.

Funding

This research project was funded by the Council of Academic Programs in Communication Sciences and Disorders (CAPCSD) Ph.D. Scholarship, the ASHFoundation New Century Scholars Doctoral Scholarship, and the Doctoral Student Award from the University of Pittsburgh School of Health and Rehabilitation Sciences to YQ. Funding to BC was by National Institute for Communication Sciences and Disorders award R01DC013315 and Northwestern University. Funding to GNG was provided by the Vice-Chancellors Graduate Research Education funds.

Acknowledgments

Sincere thanks to Sara Gonzalez for her work on data collection. Special acknowledgment to Sergio Cabrera, Vanessa Benjumea, and Neuromedica’s team. We are also profoundly grateful to our participants and their families in Medellin.

Conflict of interest

The authors declare that the research was conducted without any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2023.1122480/full#supplementary-material

References

Assaneo, M. F., Rimmele, J. M., Orpella, J., Ripollés, P., de Diego-Balaguer, R., and Poeppel, D. (2019). The lateralization of speech-brain coupling is differentially modulated by intrinsic auditory and top-down mechanisms. Front. Integr. Neurosci. 13, 1–11. doi: 10.3389/fnint.2019.00028

CrossRef Full Text | Google Scholar

Barbieri, E., Litcofsky, K. A., Walenski, M., Chiappetta, B., Mesulam, M.-M., and Thompson, C. K. (2021). Online sentence processing impairments in agrammatic and logopenic primary progressive aphasia: evidence from ERP. Neuropsychologia 151:107728. doi: 10.1016/j.neuropsychologia.2020.107728

CrossRef Full Text | Google Scholar

Becker, F., and Reinvang, I. (2007). Successful syllable detection in aphasia despite processing impairments as revealed by event-related potentials. Behav. Brain Funct. 3:6. doi: 10.1186/1744-9081-3-6

CrossRef Full Text | Google Scholar

Brallier, J. (2002). Who was Albert Einstein? New York: Penguin Workshop.

Google Scholar

Brodbeck, C., Presacco, A., Anderson, S., and Simon, J. Z. (2018). Over-representation of speech in older adults originates from early response in higher order auditory cortex. Acta Acustica united with Acustica 104, 774–777. doi: 10.3813/aaa.919221

CrossRef Full Text | Google Scholar

Bürkner, P.-C. (2018). Advanced Bayesian multilevel modeling with the R package brms. R J. 10, 395–411. doi: 10.32614/rj-2018-017

CrossRef Full Text | Google Scholar

Carpenter, B., Gelman, A., Hoffman, M. D., Lee, D., Goodrich, B., Betancourt, M., et al. (2017). Stan: a probabilistic programming language. J. Stat. Softw. 76, 1–34. doi: 10.18637/jss.v076.i01

CrossRef Full Text | Google Scholar

Chandrasekaran, C., Trubanova, A., Stillittano, S., Caplier, A., and Ghazanfar, A. A. (2009). The natural statistics of audiovisual speech. PLoS Comput. Biol. 5:e1000436. doi: 10.1371/journal.pcbi.1000436

CrossRef Full Text | Google Scholar

Colling, L. J., Noble, H. L., and Goswami, U. (2017). Neural entrainment and sensorimotor synchronization to the beat in children with developmental dyslexia: an EEG study. Front. Neurosci. 11:360. doi: 10.3389/fnins.2017.00360

CrossRef Full Text | Google Scholar

Crosse, M. J., Butler, J. S., and Lalor, E. C. (2015). Congruent visual speech enhances cortical entrainment to continuous auditory speech in noise-free conditions. J. Neurosci. 35, 14195–14204. doi: 10.1523/jneurosci.1829-15.2015

CrossRef Full Text | Google Scholar

Crosse, M. J., Di Liberto, G. M., Bednar, A., and Lalor, E. C. (2016). The multivariate temporal response function (mTRF) toolbox: a MATLAB toolbox for relating neural signals to continuous stimuli. Front. Hum. Neurosci. 10:604. doi: 10.3389/fnhum.2016.00604

CrossRef Full Text | Google Scholar

Cruice, M., Worrall, L., and Hickson, L. (2006). Perspectives of quality of life by people with aphasia and their family: suggestions for successful living. Top. Stroke Rehabil. 13, 14–24. doi: 10.1310/4jw5-7vg8-g6x3-1qvj

CrossRef Full Text | Google Scholar

Curtis, S., Nicholas, M. L., Pittmann, R., and Zipse, L. (2020). Tap your hand if you feel the beat: differential effects of tapping in melodic intonation therapy. Aphasiology 34, 580–602. doi: 10.1080/02687038.2019.1621983

CrossRef Full Text | Google Scholar

De Clercq, P., Kries, J., Mehraram, R., Vanthornhout, J., Francart, T., and Vandermosten, M. (2023). Detecting post-stroke aphasia using EEG-based neural envelope tracking of natural speech. medRxiv [Preprint]. doi: 10.1101/2023.03.14.23287194

CrossRef Full Text | Google Scholar

Decruy, L., Vanthornhout, J., and Francart, T. (2019). Evidence for enhanced neural tracking of the speech envelope underlying age-related speech-in-noise difficulties. J. Neurophysiol. 122, 601–615. doi: 10.1152/jn.00687.2018

CrossRef Full Text | Google Scholar

Delorme, A., and Makeig, S. (2004). EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21. doi: 10.1016/j.jneumeth.2003.10.009

CrossRef Full Text | Google Scholar

Di Liberto, G. M., O'Sullivan, J. A., and Lalor, E. C. (2015). Low-frequency cortical entrainment to speech reflects phoneme-level processing. Curr. Biol. 25, 2457–2465. doi: 10.1016/j.cub.2015.08.030

CrossRef Full Text | Google Scholar

Dial, H. R., Gnanateja, G. N., Tessmer, R. S., Gorno-Tempini, M. L., Chandrasekaran, B., and Henry, M. L. (2021). Cortical tracking of the speech envelope in Logopenic variant primary progressive aphasia. Front. Hum. Neurosci. 14:597694. doi: 10.3389/fnhum.2020.597694

CrossRef Full Text | Google Scholar

Ding, N., Patel, A. D., Chen, L., Butler, H., Luo, C., and Poeppel, D. (2017). Temporal modulations in speech and music. Neurosci. Biobehav. Rev. 81, 181–187. doi: 10.1016/j.neubiorev.2017.02.011

CrossRef Full Text | Google Scholar

Ding, N., and Simon, J. Z. (2014). Cortical entrainment to continuous speech: functional roles and interpretations. Front. Hum. Neurosci. 8:311. doi: 10.3389/fnhum.2014.00311

CrossRef Full Text | Google Scholar

Doelling, K. B., Arnal, L. H., Ghitza, O., and Poeppel, D. (2014). Acoustic landmarks drive delta-theta oscillations to enable speech comprehension by facilitating perceptual parsing. NeuroImage 85, 761–768. doi: 10.1016/j.neuroimage.2013.06.035

CrossRef Full Text | Google Scholar

Doelling, K. B., Assaneo, M. F., Bevilacqua, D., Pesaran, B., and Poeppel, D. (2019). An oscillator model better predicts cortical entrainment to music. Proc. Natl. Acad. Sci. 116:10113. doi: 10.1073/pnas.1816414116

CrossRef Full Text | Google Scholar

Etard, O., and Reichenbach, T. (2019). Neural speech tracking in the Theta and in the Delta frequency band differentially encode clarity and comprehension of speech in noise. J. Neurosci. 39:5750. doi: 10.1523/JNEUROSCI.1828-18.2019

CrossRef Full Text | Google Scholar

Flaugnacco, E., Lopez, L., Terribili, C., Zoia, S., Buda, S., Tilli, S., et al. (2014). Rhythm perception and production predict reading abilities in developmental dyslexia. Front. Hum. Neurosci. 8:392. doi: 10.3389/fnhum.2014.00392

CrossRef Full Text | Google Scholar

Fridriksson, J., Basilakos, A., Hickok, G., Bonilha, L., and Rorden, C. (2015). Speech entrainment compensates for Broca's area damage. Cortex 69, 68–75. doi: 10.1016/j.cortex.2015.04.013

CrossRef Full Text | Google Scholar

Fridriksson, J., Hubbard, H. I., Hudspeth, S. G., Holland, A. L., Bonilha, L., Fromm, D., et al. (2012). Speech entrainment enables patients with Broca's aphasia to produce fluent speech. Brain 135, 3815–3829. doi: 10.1093/brain/aws301

CrossRef Full Text | Google Scholar

Ghitza, O. (2013). The theta-syllable: a unit of speech information defined by cortical function. Front. Psychol. 4:138. doi: 10.3389/fpsyg.2013.00138

CrossRef Full Text | Google Scholar

Ghitza, O. (2017). Acoustic-driven delta rhythms as prosodic markers. Lang. Cogn. Neurosci. 32, 545–561. doi: 10.1080/23273798.2016.1232419

CrossRef Full Text | Google Scholar

Gnanateja, G. N., Devaraju, D. S., Heyne, M., Quique, Y. M., Sitek, K. R., Tardif, M. C., et al. (2022). On the role of neural oscillations across timescales in speech and music processing. Front. Comput. Neurosci. 16:872093. doi: 10.3389/fncom.2022.872093

CrossRef Full Text | Google Scholar

Goswami, U. (2011). A temporal sampling framework for developmental dyslexia. Trends Cogn. Sci. 15, 3–10. doi: 10.1016/j.tics.2010.10.001

CrossRef Full Text | Google Scholar

Harrison, P. M. C., and Müllensiefen, D. (2018). Development and validation of the computerised adaptive beat alignment test (CA-BAT). Sci. Rep. 8:12395. doi: 10.1038/s41598-018-30318-8

CrossRef Full Text | Google Scholar

Hilari, K. (2011). The impact of stroke: are people with aphasia different to those without? Disabil. Rehabil. 33, 211–218. doi: 10.3109/09638288.2010.508829

CrossRef Full Text | Google Scholar

Howard, M. F., and Poeppel, D. (2010). Discrimination of speech stimuli based on neuronal response phase patterns depends on acoustics but not comprehension. J. Neurophysiol. 104, 2500–2511. doi: 10.1152/jn.00251.2010

CrossRef Full Text | Google Scholar

Iversen, J., and Patel, A. (2008). “The beat alignment test (BAT): surveying beat processing abilities in the general population”, in: 10th international conference on music perception, and cognition (ICMPC 10), eds. K. I. Miyazaki, Y. Hiraga, M. Adachi, Y. Nakajima, and M. Tsuzaki).

Google Scholar

Jung, T. P., Makeig, S., McKeown, M. J., Bell, A. J., Lee, T. W., and Sejnowski, T. J. (2001). Imaging brain dynamics using independent component analysis. Proc. IEEE 89, 1107–1122. doi: 10.1109/5.939827

CrossRef Full Text | Google Scholar

Kertesz, A. (1982). Western aphasia battery test manual. San Antonio, TX: The Psychological Corporation.

Google Scholar

Kertesz, A., and Pascual-Leone, Á. (2000). Batería de afasías Western (The western aphasia battery en versión y adaptación castellana). Spain: Nau llibres.

Google Scholar

Klem, G., Lüders, H., Jasper, H., and Elger, C. (1999). The ten-twenty electrode system of the international federation. The International Federation of Clinical Neurophysiology. Electroencephalogr. Clin. Neurophysiol. Suppl. 52, 3–6.

Google Scholar

Kösem, A., and van Wassenhove, V. (2016). Distinct contributions of low- and high-frequency neural oscillations to speech comprehension. Lang. Cogn. Neurosci. 32, 536–544. doi: 10.1080/23273798.2016.1238495

CrossRef Full Text | Google Scholar

Kries, J., De Clercq, P., Gillis, M., Vanthornhout, J., Lemmens, R., Francart, T., et al. (2023). Exploring neural tracking of acoustic and linguistic speech representations in individuals with post-stroke aphasia. bioRxiv [Preprint]. doi: 10.1101/2023.03.01.530707

CrossRef Full Text | Google Scholar

Kruschke, J. K. (2013). Bayesian estimation supersedes the t test. J. Exp. Psychol. Gen. 142, 573–603. doi: 10.1037/a0029146

PubMed Abstract | CrossRef Full Text | Google Scholar

Kruschke, J. K., and Liddell, T. M. (2018). The Bayesian new statistics: hypothesis testing, estimation, meta-analysis, and power analysis from a Bayesian perspective. Psychon. Bull. Rev. 25, 178–206. doi: 10.3758/s13423-016-1221-4

CrossRef Full Text | Google Scholar

Ladányi, E., Persici, V., Fiveash, A., Tillmann, B., and Gordon, R. L. (2020). Is atypical rhythm a risk factor for developmental speech and language disorders? WIREs Cogn. Sci. 11:e1528. doi: 10.1002/wcs.1528

CrossRef Full Text | Google Scholar

Lakatos, P., Gross, J., and Thut, G. (2019). A new unifying account of the roles of neuronal entrainment. Curr. Biol. 29, R890–R905. doi: 10.1016/j.cub.2019.07.075

CrossRef Full Text | Google Scholar

Lizarazu, M., Carreiras, M., Bourguignon, M., Zarraga, A., and Molinaro, N. (2021). Language proficiency entails tuning cortical activity to second language speech. Cereb. Cortex 31, 3820–3831. doi: 10.1093/cercor/bhab051

CrossRef Full Text | Google Scholar

MathWorks Inc. (2019). MATLAB (R2019 a). (Natick, Massachusetts: The MathWorks Inc) Availabkle at: https://www.mathworks.com

Google Scholar

McHaney, J. R., Gnanateja, G. N., Smayda, K. E., Zinszer, B. D., and Chandrasekaran, B. (2021). Cortical tracking of speech in Delta band relates to individual differences in speech in noise comprehension in older adults. Ear Hear. 42, 343–354. doi: 10.1097/AUD.0000000000000923

CrossRef Full Text | Google Scholar

Mullen, T., Kothe, C., Chi, Y. M., Ojeda, A., Kerth, T., Makeig, S., et al. (2013). “Real-time modeling and 3D visualization of source dynamics and connectivity using wearable EEG”, in: 2013 35th annual international conference of the IEEE engineering in medicine and biology society (EMBC), 2184–2187.

Google Scholar

Nozaradan, S., Peretz, I., Missal, M., and Mouraux, A. (2011). Tagging the neuronal entrainment to beat and meter. J. Neurosci. 31:10234. doi: 10.1523/JNEUROSCI.0411-11.2011

CrossRef Full Text | Google Scholar

Nozaradan, S., Zerouali, Y., Peretz, I., and Mouraux, A. (2013). Capturing with EEG the neural entrainment and coupling underlying sensorimotor synchronization to the beat. Cereb. Cortex 25, 736–747. doi: 10.1093/cercor/bht261

CrossRef Full Text | Google Scholar

O’Sullivan, J., Chen, Z., Herrero, J., McKhann, G. M., Sheth, S. A., Mehta, A. D., et al. (2017). Neural decoding of attentional selection in multi-speaker environments without access to clean sources. J. Neural Eng. 14:056001. doi: 10.1088/1741-2552/aa7ab4

CrossRef Full Text | Google Scholar

Obleser, J., and Weisz, N. (2012). Suppressed alpha oscillations predict intelligibility of speech and its acoustic details. Cereb. Cortex 22, 2466–2477. doi: 10.1093/cercor/bhr325

CrossRef Full Text | Google Scholar

Paul, B. T., Chen, J., Le, T., Lin, V., and Dimitrijevic, A. (2021). Cortical alpha oscillations in cochlear implant users reflect subjective listening effort during speech-in-noise perception. PLoS One 16:e0254162. doi: 10.1371/journal.pone.0254162

CrossRef Full Text | Google Scholar

Peelle, J. E., Gross, J., and Davis, M. H. (2013). Phase-locked responses to speech in human auditory cortex are enhanced during comprehension. Cereb. Cortex 23, 1378–1387. doi: 10.1093/cercor/bhs118

CrossRef Full Text | Google Scholar

Pefkou, M., Arnal, L. H., Fontolan, L., and Giraud, A. L. (2017). theta-band and beta-band neural activity reflects independent syllable tracking and comprehension of time-compressed speech. J. Neurosci. 37, 7930–7938. doi: 10.1523/JNEUROSCI.2882-16.2017

CrossRef Full Text | Google Scholar

Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., et al. (2019). PsychoPy2: experiments in behavior made easy. Behav. Res. Methods 51, 195–203. doi: 10.3758/s13428-018-01193-y

CrossRef Full Text | Google Scholar

Pérez, A., Dumas, G., Karadag, M., and Duñabeitia, J. A. (2019). Differential brain-to-brain entrainment while speaking and listening in native and foreign languages. Cortex 111, 303–315. doi: 10.1016/j.cortex.2018.11.026

CrossRef Full Text | Google Scholar

Plechawska-Wojcik, M., Kaczorowska, M., and Zapala, D. (2019). “The artifact subspace reconstruction (ASR) for EEG signal correction. A comparative study’, in: Information systems architecture and technology: Proceedings of 39th international conference on information systems architecture and technology – ISAT 2018, eds. J. Świątek, L. Borzemski, and Z. Wilimowska, United States: Springer International Publishing), 125–135.

Google Scholar

Poeppel, D., and Assaneo, M. F. (2020). Speech rhythms and their neural foundations. Nat. Rev. Neurosci. 21, 322–334. doi: 10.1038/s41583-020-0304-4

CrossRef Full Text | Google Scholar

Quique, Y. M., Evans, W. S., Ortega-Llebaría, M., Zipse, L., and Dickey, M. W. (2022). Get in sync: active ingredients and patient profiles in scripted-sentence learning in Spanish speakers with aphasia. J. Speech Lang. Hear. Res. 65, 1478–1493. doi: 10.1044/2021_jslhr-21-00060

CrossRef Full Text | Google Scholar

R Core Team (2020). R: A language and environment for statistical computing. 3.6.3 Edn. (Vienna, Austria: R Foundation for Statistical Computing).

Google Scholar

Ranasinghe, K. G., Hinkley, L. B., Beagle, A. J., Mizuiri, D., Honma, S. M., Welch, A. E., et al. (2017). Distinct spatiotemporal patterns of neuronal functional connectivity in primary progressive aphasia variants. Brain 140, 2737–2751. doi: 10.1093/brain/awx217

CrossRef Full Text | Google Scholar

Reetzke, R., Gnanateja, G. N., and Chandrasekaran, B. (2021). Neural tracking of the speech envelope is differentially modulated by attention and language experience. Brain Lang. 213:104891. doi: 10.1016/j.bandl.2020.104891

CrossRef Full Text | Google Scholar

Robson, H., Cloutman, L., Keidel, J. L., Sage, K., Drakesmith, M., and Welbourne, S. (2014). Mismatch negativity (MMN) reveals inefficient auditory ventral stream function in chronic auditory comprehension impairments. Cortex 59, 113–125. doi: 10.1016/j.cortex.2014.07.009

CrossRef Full Text | Google Scholar

Rosen, S. (1992). Temporal information in speech: acoustic, auditory and linguistic aspects. Philos. Trans. R. Soc. Lond. 336, 367–373. doi: 10.1098/rstb.1992.0070

CrossRef Full Text | Google Scholar

RStudio Team (2016). “RStudio: Integrated development environment for R.” 1.1.456 ed (Boston, MA: RStudio, Inc.).

Google Scholar

Shannon, R. V., Zeng, F.-G., Kamath, V., Wygonski, J., and Ekelid, M. (1995). Speech recognition with primarily temporal cues. Science 270, 303–304. doi: 10.1126/science.270.5234.303

CrossRef Full Text | Google Scholar

Sihvonen, A. J., Ripollés, P., Leo, V., Rodríguez-Fornells, A., Soinila, S., and Särkämö, T. (2016). Neural basis of acquired Amusia and its recovery after stroke. J. Neurosci. 36, 8872–8881. doi: 10.1523/JNEUROSCI.0709-16.2016

CrossRef Full Text | Google Scholar

Stahl, B., Henseler, I., Turner, R., Geyer, S., and Kotz, S. A. (2013). How to engage the right brain hemisphere in aphasics without even singing: evidence for two paths of speech recovery. Front. Hum. Neurosci. 7:35. doi: 10.3389/fnhum.2013.00035

CrossRef Full Text | Google Scholar

Stahl, B., Kotz, S. A., Henseler, I., Turner, R., and Geyer, S. (2011). Rhythm in disguise: why singing may not hold the key to recovery from aphasia. Brain 134, 3083–3093. doi: 10.1093/brain/awr240

CrossRef Full Text | Google Scholar

Stan Development Team (2020). ‘RStan: the R interface to Stan". R package version 2.21.2).

Google Scholar

Stefaniak, J. D., Lambon Ralph, M. A., De Dios Perez, B., Griffiths, T. D., and Grube, M. (2021). Auditory beat perception is related to speech output fluency in post-stroke aphasia. Sci. Rep. 11:3168. doi: 10.1038/s41598-021-82809-w

CrossRef Full Text | Google Scholar

Swink, S., and Stuart, A. (2012). Auditory long latency responses to tonal and speech stimuli. J. Speech Lang. Hear. Res. 55, 447–459. doi: 10.1044/1092-4388(2011/10-0364)

CrossRef Full Text | Google Scholar

Teng, X., Tian, X., Doelling, K., and Poeppel, D. (2018). Theta band oscillations reflect more than entrainment: behavioral and neural evidence demonstrates an active chunking process. Eur. J. Neurosci. 48, 2770–2782. doi: 10.1111/ejn.13742

CrossRef Full Text | Google Scholar

Teoh, E. S., Cappelloni, M. S., and Lalor, E. C. (2019). Prosodic pitch processing is represented in delta-band EEG and is dissociable from the cortical tracking of other acoustic and phonetic features. Eur. J. Neurosci. 50, 3831–3842. doi: 10.1111/ejn.14510

CrossRef Full Text | Google Scholar

Thomson, J. M., Leong, V., and Goswami, U. (2013). Auditory processing interventions and developmental dyslexia: a comparison of phonemic and rhythmic approaches. Read. Writ. 26, 139–161. doi: 10.1007/s11145-012-9359-6

CrossRef Full Text | Google Scholar

Winkler, I., Debener, S., Müller, K. R., and Tangermann, M. (2015). “On the influence of high-pass filtering on ICA-based artifact reduction in EEG-ERP,” in: 2015 37th annual international conference of the IEEE engineering in medicine and biology society (EMBC), 4101–4105.

Google Scholar

Wolff, P. H. (2002). Timing precision and rhythm in developmental dyslexia. Read. Writ. 15, 179–206. doi: 10.1023/A:1013880723925

CrossRef Full Text | Google Scholar

Wöstmann, M., Lim, S.-J., and Obleser, J. (2017). The human neural alpha response to speech is a proxy of attentional control. Cereb. Cortex 27, 3307–3317. doi: 10.1093/cercor/bhx074

CrossRef Full Text | Google Scholar

Zan, P., Presacco, A., Anderson, S., and Simon, J. Z. (2020). Exaggerated cortical representation of speech in older listeners: mutual information analysis. J. Neurophysiol. 124, 1152–1164. doi: 10.1152/jn.00002.2020

CrossRef Full Text | Google Scholar

Zipse, L., Worek, A., Guarino, A. J., and Shattuck-Hufnagel, S. (2014). Tapped out: do people with aphasia have rhythm processing deficits? J. Speech Lang. Hear. Res. 57, 2234–2245. doi: 10.1044/2014_JSLHR-L-13-0309

CrossRef Full Text | Google Scholar

Keywords: aphasia (language), rhythm, cortical tracking of speech, EEG, Spanish

Citation: Quique YM, Gnanateja GN, Dickey MW, Evans WS and Chandrasekaran B (2023) Examining cortical tracking of the speech envelope in post-stroke aphasia. Front. Hum. Neurosci. 17:1122480. doi: 10.3389/fnhum.2023.1122480

Received: 13 December 2022; Accepted: 28 August 2023;
Published: 14 September 2023.

Edited by:

Matthew Walenski, East Carolina University, United States

Reviewed by:

Nai Ding, Zhejiang University, China
Jill Kries, KU Leuven, Belgium

Copyright © 2023 Quique, Gnanateja, Dickey, Evans and Chandrasekaran. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: G. Nike Gnanateja, nikegnanateja@gmail.com; Bharath Chandrasekaran, bchandra@northwestern.edu

Download