Skip to main content

ORIGINAL RESEARCH article

Front. Behav. Neurosci., 03 March 2022
Sec. Learning and Memory
Volume 16 - 2022 | https://doi.org/10.3389/fnbeh.2022.806520

Evaluation of a Simple Clinical Language Paradigm With Respect to Sensory Independency, Functional Asymmetry, and Effective Connectivity

Erik Rødland1,2 Kathrine Midgaard Melleby1,3 Karsten Specht1,4,5*
  • 1Department of Biological and Medical Psychology, University of Bergen, Bergen, Norway
  • 2Division of Psychiatry, Department of Child and Adolescent, Haukeland University Hospital, Bergen, Norway
  • 3Adult Habilitation Section, Telemark Hospital Skien, Skien, Norway
  • 4Mohn Medical Imaging and Visualization Centre, Haukeland University Hospital, Bergen, Norway
  • 5Department of Education, UiT The Arctic University of Norway, Tromsø, Norway

The present study replicates a known visual language paradigm, and extends it to a paradigm that is independent from the sensory modality of the stimuli and, hence, could be administered either visually or aurally, such that both patients with limited sight or hearing could be examined. The stimuli were simple sentences, but required the subject not only to understand the content of the sentence but also to formulate a response that had a semantic relation to the content of the presented sentence. Thereby, this paradigm does not only test perception of the stimuli, but also to some extend sentence and semantic processing, and covert speech production within one task. When the sensory base-line condition was subtracted, both the auditory and visual version of the paradigm demonstrated a broadly overlapping and asymmetric network, comprising distinct areas of the left posterior temporal lobe, left inferior frontal areas, left precentral gyrus, and supplementary motor area. The consistency of activations and their asymmetry was evaluated with a conjunction analysis, probability maps, and intraclass correlation coefficients (ICC). This underlying network was further analyzed with dynamic causal modeling (DCM) to explore whether not only the same brain areas were involved, but also the network structure and information flow were the same between the sensory modalities. In conclusion, the paradigm reliably activated the most central parts of the speech and language network with a great consistency across subjects, and independently of whether the stimuli were administered aurally or visually. However, there was individual variability in the degree of functional asymmetry between the two sensory conditions.

Introduction

Presurgical Mapping of Language

Presurgical mapping of areas responsible for the production and comprehension of language is essential to minimize the risk of postoperative deficits in patients who receive surgery in the language dominant hemisphere. Over many decades, the presurgical mapping of language has been done using the Wada test (Benbadis et al., 1998; Wada and Rasmussen, 2007; Niskanen et al., 2012), but today the Wada test has mostly been replaced by more non-invasive methods, particularly neuroimaging. These methods offer much higher spatial resolutions than the near dichotomous descriptions resulting from the Wada test, and allow clinicians and researchers to shed light not only on which hemisphere is speech dominant, but also on the localization of specific speech and language functions and the connectivity between different neuroanatomical areas. Among the available non-invasive alternatives, functional magnetic resonance imaging (fMRI) is most dominantly used in those clinical examinations (Deppe et al., 2000; Jansen et al., 2006; Binder, 2011; Silva et al., 2018). However, it has also been reported that clinical fMRI might be better in identifying motor areas than language areas (Orringer et al., 2012).

Today, there is a plethora of different fMRI paradigms varying in design, method and efficacy described in the literature (Fernández et al., 2003; Benke et al., 2006; Tie et al., 2014; Urbach et al., 2015; Benjamin et al., 2017), including resting-state fMRI (Tie et al., 2014; Sair et al., 2016). Some of the more commonly used paradigms in presurgical language mapping involve simple word generation and reading tasks. These result in robust activation but do not fully cover the complex multimodal nature of language processing. For this reason, comparative studies recommend using several paradigms with different tasks and both visual and auditory stimuli to ensure robust activation and asymmetry measures of all the relevant areas (Engström et al., 2004; Niskanen et al., 2012). There is, however, no common consensus on the optimal combination of tasks for clinical presurgical mapping of language and hemispheric dominance. Using several paradigms also prolong the scanning procedure, which has its own list of drawbacks; ranging from the unpleasantness and impracticality of laying still inside a noisy MR scanner for long periods of time, reduced reliability due to varying degree of attention, to the inevitable higher economic cost connected to longer scan times. More effective language paradigms, which can create robust activations of relevant areas involved in both production and comprehension of language, and which are independent of whether visual or auditory stimuli have been used, is therefore to be desired. As the need of fMRI in clinical practice continues to expand (Specht, 2020), there is also a growing need for both aurally and visually mediated paradigms to accommodate for impairments of sight and hearing in the clinical population.

The Neuroanatomy of Language

In order to develop a paradigm that activates the same speech and language areas with the same degree of hemispheric asymmetry, independent whether stimuli are presented visually or aurally, one has to operationalize current speech models into a multimodal paradigm. The classical Wernicke–Lichtheim model of speech processing, envisaged a simple network with a strong leftward asymmetry comprised of Broca’s area, responsible for the production of language; Wernicke’s area, responsible for language comprehension; and a third anatomically less specified area for the processing of concepts, as well as the connections between them (Shaw, 1874). While this early model is still taught, it is also well established that the Wernicke–Lichtheim model of speech processing is too simplistic and far from effective in describing the complexity of speech and their anatomical localization (Hickok and Poeppel, 2007; Specht, 2013, 2014). The traditional emphasis on the areas described by Paul Broca and Carl Wernicke has in later years given way to a more complex understanding of the language system. Current models recognize that both production and comprehension of language involve a widely distributed cortical network, and assume a more hierarchical structure with dynamic and context-dependent interactions between the different regions (Price, 2010, 2012).

One of the most influential models on the functional anatomy of language is the dual-stream-model (DSM), which postulates a ventral stream that is involved in speech comprehension and a dorsal stream that is involved in speech production. The ventral stream, which is partly bilaterally organized, provides important functions for basic speech perception, such as phonetic decoding, phonological and sub-lexical processing, and higher-order speech comprehension, like lexical, combinatorial and semantic processing, and appears to be organized along a gradient from posterior to anterior (Ueno et al., 2011; Specht, 2013). The dorsal stream, which is assumed to left lateralized, translates acoustic speech signals into articulatory representations and supports sensorimotor integration (Hickok and Poeppel, 2004, 2007; Poeppel and Hickok, 2004; McGettigan and Scott, 2012; Poeppel et al., 2012; Specht, 2014). There is also thought to exist a mostly left-hemispheric conceptual network, which is assumed to be widely distributed throughout the cortex. The two streams are considered to be hierarchically organized such that input to each processing step depends on the computational output of the previous step (Poeppel and Hickok, 2004; Hickok and Poeppel, 2007; Specht, 2013, 2014).

Anatomically, the early stages of speech processing are thought to occur bilaterally in the auditory regions of the dorsal superior temporal gyrus and superior temporal sulcus (Hickok and Poeppel, 2007; Specht, 2013, 2014). The two streams then emerge from the middle and posterior superior temporal sulcus, with the ventral stream spreading across structures in the superior and middle portions of the temporal lobe, and the dorsal stream comprising the premotor areas and the articulatory network, such as Broca’s area and the anterior insular (aIns) cortex (Specht, 2013). The ventral stream appears to be bilaterally organized, particularly at the phoneme-level of speech recognition, before forming a lateralization gradient from the posterior superior temporal lobe toward the anterior temporal lobe as the processing complexity increases. The dorsal stream does, however, appear to have a stronger left-sided lateralization (Hickok and Poeppel, 2000, 2004, 2007; Scott, 2000; McGettigan and Scott, 2012; Specht, 2013, 2014).

The dual stream model does not make any predictions when it comes to reading. On the other hand, it is known that, after the initial processing in the primary visual cortex, written letters and words are decoded by an area at the inferior border of the left occipital and left temporal lobe. This area is called the “visual word form area” and has strong connections to the left angular gyrus and into the speech and language network of the left temporal lobe (McCandliss et al., 2003; Dehaene and Cohen, 2011).

There are some further brain areas, which are not explicitly covered by the model but which nevertheless are repeatedly detected in neuroimaging studies (Price, 2012; Specht, 2014). Among the areas frequently identified in the studies covered by Price (2012) are the left angular gyrus and the left supramarginal gyrus. Put very simple and in the context of speech and language processes, these areas are typically associated with the semantic network, which is a wide-spread network and comprises besides these parts of the inferior parietal lobe, also areas of the temporal and frontal cortex, mostly of the left hemisphere (Binder et al., 2009; Fedorenko et al., 2011; Price, 2012; Graessner et al., 2021), and is broadly independent of the stimulus modality (Deniz et al., 2019). Similarly, the processing of sentences and syntax are assumed to take part in a network including the left posterior temporal lobe, the temporal pole and inferior frontal gyrus (IFG), and their degree of involvement in the processing might depend on syntactic complexity or predictability (Price, 2012; Matchin et al., 2017; Zaccarella and Friederici, 2017). Besides, outside of the classical language areas, also subcortical areas such as the basal ganglia, aIns, and cerebellum are repeatedly reported in neuroimaging studies (Price, 2012; Specht, 2014).

Aim

This study aimed to develop further a paradigm for clinical use that reliably activates the language network in terms of brain activations, brain asymmetry, and effective connectivity, independent of whether the stimuli were delivered visually or aurally. The resulting auditory paradigm was based on a pre-existing visual paradigm that had already shown good potential in clinical practice (Berntsen et al., 2006). This paradigm presented the participants with tasks based on the popular television show “Jeopardy!”. In such a task, participants are presented with an answer and are required to respond by generating a corresponding question. Therefore, this task requires several processing steps that are distributed over different areas of the ventral and dorsal stream. First, the content of the sentence has to be semantically decoded. Second, an appropriate target word has to be retrieved from the lexicon. Third, a corresponding answer has to be formulated as a grammatically correct question. Fourth, the answer has to be articulated. This requires both semantic and lexical processing that are functions of the ventral stream, as well as the production of a corresponding sentence as a response that is a function of the dorsal stream. Moreover, these processes are assumed to be broadly independent of the sensory modality of the original stimuli to be processed.

We hypothesized that, irrespective of the sensory modality, this task activates large parts of the left hemispheric speech and language network, comprising the ventral stream for processing the stimuli and identifying the semantic content, and the dorsal stream for generating the appropriate response. More specifically, we expected activations in the superior temporal gyrus and sulcus, inferior parietal lobe, temporal pole, inferior frontal gyrus/frontal operculum, and parts of the articulatory motor network, including subcortical structures. Further, it was expected that the individual asymmetry of the activations is independent of the stimulus modality. Finally, it was predicted that the measures of effective connectivity are comparable between the two sensory stimulations, indicating that they both activate not only the same brain areas but the same network configuration.

To test that the described processes are independent of the sensory modality of the original stimulus that had to be processed, an auditory and a visual version of the jeopardy paradigm was created. Accordingly, it was hypothesized that both versions would activate the dorsal and ventral stream of the speech and language network to the same extent, and that the effective connectivity within the speech and language network will be unchanged.

Materials and Methods

Participants

Twenty-one healthy Norwegian-speaking participants, consisting of 10 men and 11 women, were recruited for the study. The participants were all right-handed, aged 21–50 years, with a mean age of 25.3 years (SD 6.2). All subjects were informed of the aim of the study and the criterions for participation. The criterions excluded people with a clear left-handed preference, people with a history of psychiatric or neurological disorders, people with claustrophobic tendencies and people with metallic implants. To control for handedness, participants filled a modified Edinburgh handedness questionnaire with 15 items, which asked for hand-preferences for certain actions, with the answering options: left, both, or right hand). All participants selected “right hand” on 12 or more items (mean 14.5 items).

The study was conducted in accordance with the Helsinki declaration, and all subjects gave written informed consent before participation per the institutional guidelines. The study was approved by the Regional Committee for Medical Research Ethics (REK-Vest).

Paradigms

For this study, both a visual and an auditory paradigm were used. Aside from how the tasks were mediated, the conformation of the paradigms was mostly identical. Both paradigms were of a simple blocked design, with alternating experimental and control blocks. The active blocks of both paradigms contained simple tasks based on the television show “Jeopardy!” (Berntsen et al., 2006). The subjects were presented with an answer and were required to respond by generating the corresponding question. The participants were instructed to formulate their questions as a sentence, starting with the phrase “what is < target word >” A typical example used in one of the paradigms was “the color of the sky” to which the corresponding answer would be “what is blue?”. Participants were instructed not to formulate more complex sentences or sentences containing any other verbs. The participants were furthermore instructed to use covert responses in both paradigms to avoid head movements and magnetic field variations. For the control trials, participants were instructed to perceive them only attentively without any active processing. Each paradigm contained a total of 48 different tasks so that no task was repeated to the subjects at any point in the study. All stimuli were in Norwegian, and all participants were native speakers.

The visual paradigm consisted of eight active blocks, each containing six simple jeopardy-based tasks presented to the participants through MR compatible video goggles. The active blocks had a total duration of 30 s each, giving the subjects 5 s to covertly formulate an appropriate response to every task. The paradigm operated without any response data, which meant that the tasks were presented for the full 5 s before being replaced by a new task without a recorded input from the participants. The active blocks were interlaced with eight control blocks, intended to represent a baseline value without activation of the areas related to speech and language processing. These resting blocks alternated between six different rows of hashes resembling a sentence structure (e.g., #### # ## ###). These control stimuli were presented for 5 s each, as well. The combined duration of the active and the resting blocks was 8 min.

Correspondingly, the auditory paradigm was designed with eight active and eight resting blocks. The active blocks consisted of jeopardy tasks, which were similar, but not identical, to the tasks in the visual paradigm. In the auditory paradigm, the tasks were presented as pre-recorded audio files played through MR compatible headphones worn by the participants. The tasks were presented with the same 5 s frequency as in the visual paradigm. In the auditory paradigm, the resting blocks consisted of the same auditory stimuli being played backwards, rendering them unintelligible.

Both paradigms were precisely synchronized with the scanner using a synchronization box that forwarded the trigger signals from the scanner to the presentation software.

Neuroimaging

Before entering the MRI scanner, all participants were informed about the purpose of the study and explained how the study was structured. Inside the MRI scanner, the subject’s head was padded on both sides of the headphones with additional pads to restrict movements. The visual paradigm was presented using MR compatible video goggles. The auditory paradigm was mediated through MR compatible headphones.

The structural and functional scanning was performed using a 3 Tesla General Electric Medical Systems Signa HDxt scanner. The axial slices of the functional imaging were positioned parallel to the AC-PC line with reference to a high-resolution anatomical image of the entire brain volume, which was obtained using a T1-weighted gradient echo pulse sequence. During the functional imaging, both paradigms were presented to the participants in two separate runs, but the order of the presentation was alternated for each participant. In total, 320 (2 × 190) functional images were acquired, using an T2*-weighted echo-planar imaging (EPI) sequence with the following parameter: 30 axial slices (3.4 mm × 3.4 mm × 4.4 mm) with an interleaved slice order; matrix 128 × 128; TR 3000 ms; TE 30 ms; flip angle 90°. Diffusion tensor images were collected at the end of the procedure, but these data were not used in the here presented analysis.

Data Processing

The resulting EPI images were prepossessed and analyzed using SPM12 (v77711) running under MATLAB 2016a on a Linux Ubuntu 16.04 workstation. First, the images were realigned within and between the two conditions/runs to adjust for head movements during the scanning procedure, unwarped to correct for inhomogeneity issues and controlled for remaining movement artifacts. The images were then normalized according to the stereotactic references from Montreal Neurological Institute (MNI) and resampled with a voxel size of 2 × 2 × 2 mm. Finally, the images were smoothed using an eight millimeter Gaussian kernel to reduce noise and variation between the participants.

Head Movement

Prior to the statistical analysis, the realignment parameters of each participant were examined, with respect to total translation during the data acquisition and framewise displacement (Power et al., 2012). Only subject with less than 2 mm translation during the entire examination were considered in the further processing, and framewise displacements needed to be less than 0.5 mm.

General Linear Model

A fixed-effect statistical model, based on the general linear model (GLM), was used as a first-level analysis on the individual datasets. Data from the two different sensory conditions were modeled according to a design matrix using the hemodynamic response as the basis function. The following contrast was defined for each participant: (1) Jeopardy task > resting block for each sensory condition separately, i.e., auditory and visual paradigm, and (2) the differences between the two conditions, i.e., visual stimuli > auditory stimuli and auditory stimuli > visual stimuli. In the second-level analysis, a group level analysis with data from all the participants was performed, resting on a one-way ANOVA model, which utilized the contrast images from the first-level analysis. The variance estimation took into account that the data were not independent, and an equal variance across paradigms was assumed. First, brain activations were explored for the two sensory modalities separately. Results were examined by applying a family-wise error (FWE) corrected threshold at the voxel level of p < 0.05 and at least 10 voxels per cluster.

In order to identify the areas that were significantly activated during both conditions, a “real” conjunction analysis (Nichols et al., 2005) was performed and explored with the same corrected threshold as described above.

Differences between the two tasks were explored with a difference contrast, and the same threshold was applied. Finally, the MNI coordinates were used to anatomically locate the significantly activated brain areas by using the inbuilt brain atlas of SPM12 (Neuromorphometrics2), by using overlays on the anatomical atlas AAL3 (automated anatomical labeling atlas, version 3; Tzourio-Mazoyer et al., 2002; Rolls et al., 2020), integrated in MRIcron3, and overlays on a structural MNI template, which was explored by an expert (KS).

Intra-Class Correlation Coefficient

The main aim of the study was to explore whether results were not only comparable, as explored by the conjunction, but also reliable within individuals, across the sensory modality. Therefore, voxel-wise intra-class correlation coefficients (ICC) were estimated. An ICC is a measure of how the observed variance is split between within- and between-subject variance (Shrout and Fleiss, 1979; Fernández et al., 2003; Specht et al., 2003). Accordingly, an ICC between 0.7 and 1 is considered to indicate reasonable good reliability, since the between-subject variance is substantially higher than the within-subject variance.

Laterality

The lateralization of the activations has been evaluated with the LI-toolbox (Wilke and Lidzba, 2007). As areas of interest served the frontal, cingulate, temporal, parietal, occipital, central, and cerebellar areas. The analysis was performed using the integrated bootstrapping method with their default values, an exclusive mask of 11 mm around the midline, and clustering. The significances of the functional asymmetries were tested for both sensory conditions independently with simple t-tests, and were compared across sensory conditions using paired t-tests. Finally, to evaluate whether the two conditions resulted in the same degree of functional asymmetry within each participant, the laterality index for the auditory condition was correlated with the laterality index of the visual condition. This was done for each of the seven areas of interest separately. Correlations across different areas were not examined.

Between Subject Overlap

Probability maps of brain activations have been estimated. The individual t-maps were binarized at a threshold of t > 3.09 (p < 0.001) and summed up, using the ImCalc function in SPM12. The resulting probabilities maps were scaled to represent percent overlap of activations. These maps were explored at a threshold of at least 20% overlap (i.e., for the present study this corresponds to 4 or more subjects). This type of analysis allows identifying areas that have been consistently activated (p < 0.001). Second, this analysis also gives an impression to which degree areas of the right hemisphere have been activated. In other word, whether there was some variation in lateralization across subjects (Van der Haegen et al., 2011; Bishop, 2013; Hausmann et al., 2019).

Dynamic Causal Modeling

The primary aim of this study was to reliably activate the language network through two different conditions that differed in their sensory input by keeping the task constant. Therefore, it was expected that the underlying neuronal network that is the dorsal and ventral stream would be similarly activated independently from the sensory input.

To verify this, a dynamic causal modeling (DCM) analysis (Stephan et al., 2009; Friston, 2010; Osnes et al., 2011b; Penny, 2012; Morken et al., 2017; Friston et al., 2019) was conducted using seven areas that represented the dorsal and ventral stream and that were activated in both conditions. The procedure for extracting the time series followed in the main the guidelines as described by Zeidman et al. (2019). First, the coordinates of the ROIs were determined by identifying the seven most relevant areas from the described conjunction analysis. Two sets of models were constructed, one set for the visual and one for the auditory modality. The two sets used the same seven nodes for the speech and language network, and a modality-specific eighth node as sensory input area. All DCM models were restricted to the left hemisphere and resembled the dorsal and ventral stream. The following areas were included: the superior temporal gyrus [MNI: –56 –48 10], superior temporal sulcus [MNI: –54 –12 –16], middle temporal gyrus [MNI: –52 –40 –6], frontal operculum [MNI: –46 14 –2], IFG (pars opercularis) [MNI: –48 12 26], precentral gyrus (PreCG) [MNI: –52 –6 50] and supplementary motor area [MNI: –2 10 48]. In addition, a modality-specific area served as input area for the model. This eighth node of the model was identified by the difference contrast between the two conditions, which was the visual word form area [MNI: –28 –92 0] for the visual task and the primary auditory cortex [MNI: –44 –22 2] for the auditory task. To allow for individual variability in the precise localization of the activations, time courses were extracted from the individual local maximum of activation that was less than 8mm apart from the coordinates, described above (Zeidman et al., 2019).

For each sensory modality, 57 models were defined which varied both in their underlying functional connectivity (A-matrix, 3 models) and the influence of the sensory input on the network configuration (B-matrix, 19 models). The different models are displayed in Supplementary Figure 1. A Bayesian model selection was applied for the two sensory modalities independently for identifying the specific model that expresses the detected network best (Stephan et al., 2009; Penny et al., 2010). First, the most probable family of A-models was identified and, subsequently, the respective B-model was determined. The B-models were also grouped into three families, where the sensory input mostly influenced the dorsal stream, the ventral stream, or mainly areas for perception or production, while one B-model hypothesized that the sensory input does not influence at all. The estimated parameters from the favored model for the visual and for the auditory condition were compared using paired t-tests.

Similar to the voxel-wise neuroimaging results, ICCs have been estimated for all DCM parameters.

Results

Head Movement

The largest observed head movement across an entire timeseries was 1.22 mm (group mean: 0.52 ± 0.26 mm). Further, the maximal framewise displacement across the timeseries was estimated for each participant, and the averaged maximal framewise displacement across the entire group was 0.32 ± 0.19 mm. However, three participants showed a value slightly larger than 0.5 mm. An inspection of the framewise displacements across the timeseries showed that these values originated in all three cases from a single spike for a single volume, while the values for the rest of the timeseries were below the usual threshold of 0.5 mm (Power et al., 2012). Therefore, no participant was excluded from the subsequent analyses and no motion-censoring was performed.

Brain Activations/General Linear Model

When compared against the visual baseline, the visual paradigm showed increased BOLD responses mostly in the areas for reading, speech perception and production of the left hemisphere, corresponding to the dorsal and ventral stream. The activated areas comprised the IFG, PreCG, aIns, supplementary motor areas (SMA), basal ganglia (BG), thalamus, hippocampus, superior temporal gyrus (STG), superior temporal sulcus (STS), middle temporal gyrus (MTG), inferior temporal gyrus (ITG), the left fusiform gyrus, lingual gyrus, and cerebellum. In the right hemisphere, the cerebellum, anterior insula, precentral gyrus, hippocampus, and brainstem were involved (see Table 1A and Figure 1A).

TABLE 1
www.frontiersin.org

Table 1. The table (A–E) list all significant results [p(FWE) < 0.05 at voxel level, at least 10 voxels per cluster] for all estimated contrasts, and (f) the results from the analysis using an intraclass correlation coefficient (ICC > 0.6).

FIGURE 1
www.frontiersin.org

Figure 1. The figures display the results from the fMRI analysis with a section view and lateral render views of the left and right hemisphere. (A) Main activations for the visual (red) and auditory (green) variant of the paradigm after subtracting the corresponding control conditions [p(FWE) < 0.05 at voxel level, at least 10 voxels per cluster]. Activations for each condition are displayed with a voxel-wise threshold of p(FWE) < 0.05, at least 10 voxels per cluster. The areas that colored in yellow represent the additive overlap of both conditions. (B) Results from the real conjunction analysis across both paradigms [p(FWE) < 0.05, at least 10 voxels per cluster]. (C) Results from the cross-modal reliability analysis, using an intraclass correlation coefficient (ICC). Areas with an ICC > 0.6 are colored in red, with the conjunction analysis as background.

When compared against the auditory baseline, the auditory paradigm showed increased BOLD signals in very similar area of the left hemisphere as the visual paradigm, but with more activations within the temporal lobe and no activations within the fusiform and lingual gyrus (see Table 1B and Figure 1A).

Accordingly, the conjunction analysis showed for the central areas of the ventral and dorsal stream significantly increased brain activity for both paradigms, including the supplementary motor area, the left basal ganglia, the brainstem, the cerebellum, and the left and right anterior insula (see Table 1C and Figures 1B,C).

When exploring the differences in brain activations between the paradigms, only differential activations within areas related to the sensory processing were detected. When comparing the visual to the auditory paradigm, higher BOLD signals occurred bilaterally in the inferior occipital gyrus and lingual gyrus, and the left superior parietal lobe. Furthermore, the differential activations in the left hemisphere extended toward the area, aka the visual word form area. Interestingly, there was a strong differential effect within the left cerebellum (see Table 1D and Figure 2A).

FIGURE 2
www.frontiersin.org

Figure 2. The figures display the differential effects between the visual and auditory paradigm. Results are displayed at a FWE corrected threshold of p(FWE) < 0.05 and at least 10 voxels per cluster. (A) Stronger activation during the visual than auditory paradigm, (B) stronger activation during the auditory than visual paradigm.

When comparing the auditory to the visual paradigm, increased BOLD signals occurred in the left and right Heschl’s gyrus and posterior STG and STS, the left planum temporale, and the left caudate nucleus (see Table 1E and Figure 2B).

Intra-Class Correlation Coefficient

The consistency of the activations between the sensory modalities were evaluated by estimating voxel-wise the intraclass-correlation coefficients (ICC) (Specht et al., 2003). This confirmed the observation and demonstrated that the majority of the overlapping cross-modal activations were activated to the same extent by both paradigms (ICC > 0.5, more than 10 voxels per cluster) (see Table 1F and Figure 1C). Reliable activations were found in the MTG, ITG, IFG, PreCG, and SMA.

Laterality

All areas, except the cingulate and cerebellar area, showed a significant leftward asymmetry for both modalities, when tested separately and when a Bonferroni correction was applied (in total 21 t-test were performed). The cingulate area did not show any functional asymmetry and the cerebellar area demonstrated a leftward asymmetry for the visual and rightward asymmetry for the auditory paradigm. When comparing the two modalities, the results from the LI-toolbox indicated that there were significant differences only for the temporal lobe and the cerebellum, but only the difference for the cerebellum remains significant after Bonferroni correction. All other areas showed the same degree of asymmetry for both sensory modalities across the subjects (see Table 2). However, significant correlations between the condition-specific laterality indices were only found for the temporal lobe, the cingulate cortex, and central areas, but not frontal areas (see Table 2).

TABLE 2
www.frontiersin.org

Table 2. The table lists the laterality index (standard deviation) for the seven examined areas.

Between Subject Overlap

The analysis of between-subject overlap reflects a large consistency between the subject in their activations, although only a few areas showed a 100% overlap between subjects. This was only the case for the SMA and the premotor cortex. In more than 80% of the participants, the areas of the IFG were consistently activated with a significance value of t = 3.09 or above. In general, the areas that were seen in the conjunction analysis also showed a good between-subject overlap. It should also be mentioned that in more than 25% of the cases, activations were detected in the right temporal lobe for the auditory condition. In almost 50% of the cases, right frontal activations were detected during the visual condition.

Dynamic Causal Modeling

The DCM results were explored – separately for the two paradigms – in a hierarchical process. First, a Bayesian model selection (BMS) was applied to the three A-model families that varied in their underlying connectivity matrix (see Supplementary Figure 1a). This revealed the fully connected model for both modalities. Then the BMS was applied to the remaining 19 models that varied according to their B matrix (see Supplementary Figure 1b). This revealed no modulations of connection (the “null model”) in the visual paradigm, but increased connectivity between the IFG and the frontal operculum in the auditory paradigm. When exploring the estimated parameter with a one-sample t-test, however, this parameter appeared not to be significant (p = 0.07). Therefore, only the A-matrix was statistically compared between the two modalities. This was done by performing 64 paired t-tests (i.e., an 8 × 8 A-matrix for both modalities) but after applying a Bonferroni correction, there were no significant differences between the two A-matrices. Therefore, it was concluded that there were no systematic differences in the A-matrices between the two modalities. Accordingly, the two A-matrices were averaged for the analysis of the involved neuronal network. The averaged A-matrix was re-examined with 64 one-sample t-tests, and, accordingly, a Bonferroni correction was applied. This revealed a network of significant connections that mostly resampled the dorsal and ventral stream (see Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. The figure summarizes the results from the dynamic causal modeling analysis. Since there were no significant differences between paradigms, parameters have been averaged, and the figure displays the network configuration after one-sample t-tests and Bonferroni correction. (A) The figure displays the network configuration. All displayed connections are excitatory, but self-inhibitory connections are not displayed. The blue lines indicate the connections from the sensory area into the speech and language network, the orange lines should illustrate the ventral stream, and the green lines should illustrate the dorsal stream. (B) The same result as connectivity matrix. The colors represent the T-values for the connections, as indicated by the color bar to the right. The node “Sensory” corresponds to the visual word form area for the visual paradigm and the primary auditory cortex for the auditory paradigm. SMA, Supplementary Motor Area; PreCG, Precentral Gyrus; IFG, Inferior Frontal Gyrus; IFGop, Inferior Frontal Gyrus-pars opercularis; STG, Superior Temporal Gyrus; STS, Superior Temporal Sulcus; MTG, Middle Temporal Gyrus.

Although there were no significant differences between the A-matrices of the two paradigms, all 64 elements were subjected to an ICC analysis. The results revealed that the ICC values were lower than for the voxel-wise analysis described above (all ICCs < 0.4).

Discussion

The overall aim of this study was to develop a short language paradigm for clinical applications that would reliably activate central parts of the speech and language network, and that would give identical results independent whether stimuli are presented visually or aurally. For the present study a “Jeopardy!” paradigm was selected since it was expected that this paradigm would involve several core language processes, such as lexical, semantic, and syntactic processing (Berntsen et al., 2006). These functions are mostly covered by both the dual-stream (Hickok and Poeppel, 2004, 2007) and the extended dual-stream model (Specht, 2014), and it was hypothesized that this paradigm would confirm these models both in terms of activations and connectivity.

Brain Activations

As expected, both paradigms activated (when the corresponding sensory control condition was subtracted) a widespread network of left temporal and frontal areas (see Figures 1A,B). The visual paradigm largely resembles the activations found by the original study (see Figure 1; Berntsen et al., 2006). In accordance with the extended dual-stream model (Specht, 2014), activations were also detected within the basal ganglia and anterior insular cortex. While there was an activation of the right anterior insular cortex for both paradigms, activation of the right superior temporal sulcus was only observed for the auditory paradigm (see also Figure 2).

This is also reflected by the results from the analysis of the functional asymmetry that indicated differences between the two paradigms for the temporal lobe, but not for the frontal lobe. Although both paradigms showed a significant leftward asymmetry, the additional activation of the right superior temporal sulcus gave a significantly weaker leftward asymmetry for the auditory paradigm (see Table 2). However, the effect becomes non-significant when a Bonferroni correction is applied.

Interestingly, the visual paradigm apparently activated the left cerebellum stronger than the auditory paradigm. This was also confirmed by the analysis of functional asymmetry that demonstrated differential asymmetry only for the cerebellum but not occipital lobe (see Table 2 and Figure 2A). The stronger involvement of the cerebellum supports the notion that the cerebellum might be more involved in reading due to a required translation into phonological codes, which is not necessary for the processing of auditory information, while the cerebellar functional asymmetry of language processes is still controversially discussed (Mariën et al., 2013).

In general, there was a considerable consistency between the two paradigms, not only when comparing activation patterns, but also when analyzing them with a conjunction analysis and an analysis of the reliability of activations across paradigms. Both paradigms involved most of the hypothesized areas, and the ICC analysis demonstrated that the posterior STG and STS, MTG, frontal operculum, IFG, precentral gyrus, and SMA showed high reliability of the activation strength. These are important core areas of the speech and language network (Price, 2012; Specht, 2014). However, against our a priori hypothesis, the inferior parietal lobe, in particular the angular gyrus, and the temporal pole were not activated by this task – even not at a reduced significance threshold. This may indicate that this task activates the central parts with the most basal functions of the speech and language network, but the semantic and syntactic system is not activated to the full extend. One reason for that might be that the requested response sentences are very simple and were of the type “what is < target word >” and did not contain any complex syntax or other verbs.

Further, a detailed analysis of the functional asymmetry revealed that, although the overall group results were comparable, there were different degrees of functional asymmetry between the sensory conditions. More importantly, the laterality indices for the functional asymmetry correlated only for the cingulate, central, and temporal regions but not for the frontal lobe. This indicates that, for the frontal lobe, the individual degree of functional asymmetry was different between the auditory and visual condition (see Table 2). This can also be seen in the analysis of overlapping activations (see Figure 4, left). Here, the right frontal lobe seems to be activated during the visual condition in some participants, as there is overlapping activations in about 25–30% of the participants, while the same area was inactive for the auditory condition (see Figure 4, right). Although this activation did not become significant on the group level, it might indicate that there exists some individual variability within the right frontal lobe, and, more interestingly, that this area might become active only during reading but not listening. Thereby, an individually estimated functional asymmetry index might give different results for a reading and a listening task, while there was more consistency for the temporal lobe.

FIGURE 4
www.frontiersin.org

Figure 4. The figure displays the percent-wise overlap of brain activations for the visual (left) and auditory (right) sensory condition. The estimation is based on the individual spmT-maps, which were binarized at a threshold of T > 3.09 (corresponding to p < 0.001).

The results demonstrate that the two sensory conditions equally activated the areas of the left temporal lobe. Central parts for the comprehension of a language message are the posterior parts of the STG, STS, and MTG. These are the areas that are mostly associated with word processing and semantic decoding (Price, 2012). The left STS plays a particular function in this network, as it is not only a multisensory area (Specht and Wigglesworth, 2018), but also an area that is very sensitive to phonetic information in an acoustic signal, as demonstrated by the “sound morphing” paradigms (Specht et al., 2005, 2009; Osnes et al., 2011a). Later, Sammler et al. (2015) could demonstrate with a similar paradigm a corresponding effect for prosody processing for the right STS. This study also reflects this differential responsiveness and division of labor between the left and right STS since both paradigms activate the left STS, but only the auditory variant activates the right STS. This is further confirmed by the difference contrast between the paradigms that mostly showed differences for the right STS (see Figure 2B).

The other areas that were detected by this study can mainly be associated to lexical processing and sentence generation (Heim et al., 2009; Grande et al., 2012) and the generation of the (covert) response (Eickhoff et al., 2009; Price, 2012). Referring to the conjunction analysis, these areas comprise the IFG, frontal operculum, PreCG, SMA, but also the subcortical areas aIns and BG. While the contribution of the cortical areas in speech processes have been demonstrated by many studies before, these findings supplements the ongoing discussion that models of speech and language functions also have to include subcortical areas (Specht, 2014). There is an increasing awareness of various speech, language, and speech motor function of the aIns (Dronkers, 1996; Eickhoff et al., 2009; Ackermann and Riecker, 2010; Nieuwenhuys, 2012; Rodriguez et al., 2018). The same is true for the basal ganglia with discussed involvements in both perception and production (Eickhoff et al., 2009; Kotz et al., 2009; Enard, 2011; Lim et al., 2014; Kotz and Schmidt-Kassow, 2015). However, since the current paradigm involves both the perception and production, it is not a suitable paradigm for disentangling the specific function of these subcortical areas or anatomical subdivision to certain processes.

However, the paradigm was not efficient enough to activate the semantic and syntactic processing network to its full extend, since the angular gyrus, supramarginal gyrus, and temporal pole were not detected – even not at an uncorrected threshold. As aforementioned, the reason for this might be that this task is semantically and syntactically too simple since only everyday knowledge is requested by the task and neither the stimuli nor the responses are syntactically complex and always of the same type. The lack of activation in the anterior temporal lobe could be explained through the observations in patients with primary progressive aphasia of the semantic type. These patients show focal atrophy of the anterior temporal lobe causing predominantly semantic deficits (Zahn et al., 2005; Wilson et al., 2014). More specific, Wilson and colleagues could demonstrate that this degeneration seems to affect more the higher-level processing rather than simple syntactic processes (Wilson et al., 2014). Further, the activations within the semantic network might vary dependent on the processed category, amount of lexical information, or the requested semantic relation (Patterson et al., 2007; Binder et al., 2009; Graessner et al., 2021). Although the angular gyrus and its subdivisions are often involved in semantic processing at various levels of complexity, the angular gyrus needs to be seen as a part of a broader network with strong fronto-parietal interactions (Graessner et al., 2021), where at least some of the other nodes, like the anterior IFG, got activated by both tasks.

Eventually, a clinical paradigm that aims to activate not only the core areas but the speech and language system to its almost full extend should seek for a balance between a task that is easy enough to be perform by impaired patients but with a higher degree of syntactical, lexical, semantic and combinatory complexity and well-controlled semantic categories and semantic relations. Further, one might consider a way of actively controlling the performance of the participants to overcome the lack of overt responses through random catch trials which require an additional response through button presses. Those performance control might also stabilize the activation patterns in the detected network and might also improve the signal in those areas that were not reliably detected.

Dynamic Causal Modeling

The discovered core areas served as input areas for the DCM analysis, which was initially independently analyzed for the two paradigms. Importantly, the subsequent analyses revealed that there were no significant differences in connectivity between the two paradigms. However, the reliability of the estimated parameter was much lower than for the voxel-wise ICC analysis, which corresponds to earlier observations (Frässle et al., 2015). This may indicate that voxel-wise analyses are currently more stable than DCM based analyses – at least for one of the two sensory modalities. This may be an important aspect and limitation for future clinical application of network-based analysis methods.

Since there were no systematic differences between the paradigms, the estimated DCM parameters were averaged and jointly analyzed for the two paradigms. The absence of significant differences may further indicate that the sensory modality of the original stimuli does not influence the information flow within the given network.

Importantly, the dual-stream model nicely emerged from this analysis. According to the DCM results, the posterior STG and the IFG serve as central nodes of this network, with several forwards and backwards connections. The sensory information enters this network at all nodes of the temporal lobe and, interestingly, the PreCG. This latter aspect might give some further evidence to the ongoing discussion of direct motor involvement during speech perception (Liberman and Mattingly, 1985; Galantucci et al., 2006; Devlin and Aydelott, 2009; Osnes et al., 2011b). The current study indicates that also reading may involve motor areas. However, an in-depth discussion of this controversy is out of the scope of the present report.

The most obvious result from the DCM analysis is that the network structure nicely represents the ventral and dorsal streams (see Figure 3A). The high degree of bidirectional connectivity between the nodes of the temporal lobe may resemble the ventral stream. This is contrasted by a mostly frontal connectivity pattern that cumulates in more forward connections to the SMA. This pattern might resemble the dorsal stream and is, therefore, crucial for the production of the speech sounds. The SMA is an important area for generating motor responses (Nachev et al., 2008), but has also been linked to auditory processing and auditory imagery (Lima et al., 2016). Since the participants performed covert responses, the seen activations might be a mixture of motor preparation and auditory imagery. As can be seen in Figure 1, the activation involves both pre-SMA and SMA (Nachev et al., 2008) with the most significant spot within SMA. Notably, the ICC analysis indicated reliable activations in both parts of the SMA (see Figure 1C).

The posterior STG appears to be the node with the highest degree of connectivity and as the connecting node between the two hypothesized streams since there are bidirectional connections of the STG, not only to the other nodes within the temporal lobe, but also to the IFG, and PreCG, and a forward connection to the frontal operculum (see Figure 3). This is in accordance with current models that postulate a similar connectivity pattern of the posterior temporal areas as an area that connects the two streams (Hickok and Poeppel, 2007; Rauschecker and Scott, 2009; McGettigan and Scott, 2012; Specht, 2013, 2014; Binder, 2016). However, the used paradigm is not suitable to disentangle the different functions of the various connections in the comprehension and production process. Given the detected pattern, it is nonetheless evident that the frontal areas are mostly dedicated to the production since the ultimate endpoint of the connections is the SMA, which is the central area for planning and execution of motor actions.

Reliability and Future Clinical Application

Besides the general activation and connectivity patterns that were elicited by the two paradigms, it was also examined whether these patterns were replicable between the two sensory modalities. This was achieved by supplementing the analysis with an ICC analysis that tested whether the detected strength of the activations and the parameter of the effective connectivity were replicable, independent of the sensory input modality. Since the paradigm aimed to identify the speech and language network equally well by both modalities, this is an important measure prior to future clinical applications.

Comparing the conjunction and ICC analysis, the conjunction was more extended. This is mostly caused by the fact that the conjunction analysis highlights all areas that were significantly activated by both paradigms (Nichols et al., 2005), while the ICC analysis answers a much stronger question, namely which areas showed the same level of activation in both paradigms (Specht et al., 2003). It is important to emphasize that, although the areas that showed reliable levels of activations are much smaller than those indicated by the conjunction, all important nodes of the dual-stream model are included (Poeppel and Hickok, 2004; Hickok and Poeppel, 2007; Specht, 2013, 2014). This is an inevitable prerequisite for a clinical paradigm that aims to examine the speech and language network with only one type of paradigm that is independent of the sensory stimuli of the to be processed stimuli.

The ICC analysis of the estimated parameter from the DCM analyses indicated that the estimation of the parameters were more variable between the two paradigms. Although they were not significantly different between the two paradigms, the distribution of the within- and between-subject variance was inferior to the voxel-wise analysis. Consequently, a DCM analysis may be a useful tool for examining underlying network structures per se, but voxel-wise analyses appear to be superior in a clinical context – at least for the given paradigm. However, an expanded DCM analysis, which covers both hemisphere might be an even more efficient approach, despite computational limitations.

Irrespective of that, before such a paradigm could be routinely used in a clinical setup, it needs further replication in different population and across scanner (and software) platforms. This is especially of relevance in the light of the “replication crisis” (Maxwell et al., 2015).

Similarly, an examination of the effects of motion on the results and the explored network structure goes beyond the scope of the current report. However, this is an important aspect in the context of clinical fMRI where head movements are a substantial source of noise, and which needs to be examined further.

Conclusion

The results demonstrate that, independent from the sensory modality, this paradigm reliably activated the same brain networks, namely the core areas of the dorsal and ventral stream for speech processing. Only the cerebellum demonstrated differential effects. Further, the ICC analysis revealed that there was high reliability of brain activation across sensory modalities. This was supported by the fact that the DCM analysis showed that the underlying network structure and connectivity was the same between sensory modalities, although the parameter of the effective connectivity appear to vary with the sensory modality. However, a closer inspection of the individual functional asymmetry indicated that the degree of functional asymmetry was not the same for the two conditions. In particular, the visual condition showed in more individuals a right frontal activation than the auditory condition. This effect needs further exploration.

In conclusion, the explored paradigm activated the most central parts of the speech and language network, mostly independently of whether the stimuli were administered aurally or visually. Further, the DCM analysis revealed that the underlying connectivity patterns of the left hemisphere were similar, if not identical, for both paradigms, although the reliability was lower than for the activation data. Taking both aspects together, this paradigm appears to be suitable as a clinical paradigm since both patients with visual or aural disabilities can be equally examined. However, to stimulate the speech and language system more complete, one might better use syntactically and semantically more complex stimuli since the current paradigm could reliably activate only the most central core areas. Further, before this paradigm is ready for a broader clinical application, one needs to replicate these findings in an independent sample, which were examined on a different MR scanner. Further, the degree of leftward asymmetry, and its inter-individual variability, needs further critical examination. Third, the applicability of the paradigm in congenital or acquired deaf or blind people should be evaluated. Nevertheless, the provided evidence let one assume that this type of paradigm is highly suitable for various clinical applications.

Data Availability Statement

The datasets presented in this article are not readily available because data sharing has been restricted through the evaluation of the ethical committee. Requests to access the datasets should be directed to KS, Karsten.specht@uib.no.

Ethics Statement

The studies involving human participants were reviewed and approved by Regional Ethical Committee – West, University of Bergen. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

ER, KM, and KS designed the study and analyzed the data, based on the GLM. ER and KM created and recorded the stimuli, collected the data. KS performed the DCM and asymmetry analyses. ER and KS wrote the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This study was supported by a grant to KS from the Bergen Research Foundation/Trond Mohn Foundation (When a sound becomes speech) and the Research Council of Norway (217932: It’s time for some music, and 276044: When default is not default: Solutions to the replication crisis and beyond).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We thank all participants for their participation and the staff of the radiological department of the Haukeland University Hospital for their help during data acquisition. The study was conducted in collaboration with the Department of Psychology and Behavioural Sciences, Aarhus University, Denmark.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnbeh.2022.806520/full#supplementary-material

Footnotes

  1. ^ http://fil.ion.ucl.ac.uk/spm
  2. ^ http://www.neuromorphometrics.com
  3. ^ http://nitrc.org/projects/mricron

References

Ackermann, H., and Riecker, A. (2010). The contribution(s) of the insula to speech production: a review of the clinical and functional imaging literature. Brain Struct. Funct. 214, 419–433. doi: 10.1007/s00429-010-0257-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Benbadis, S. R., Binder, J. R., Swanson, S. J., Fischer, M., Hammeke, T. A., Morris, G. L., et al. (1998). Is speech arrest during Wada testing a valid method for determining hemispheric representation of language? Brain Lang. 65, 441–446. doi: 10.1006/brln.1998.2018

PubMed Abstract | CrossRef Full Text | Google Scholar

Benjamin, C. F., Walshaw, P. D., Hale, K., Gaillard, W. D., Baxter, L. C., Berl, M. M., et al. (2017). Presurgical language fMRI: mapping of six critical regions. Hum. Brain Mapp. 38, 4239–4255. doi: 10.1002/hbm.23661

PubMed Abstract | CrossRef Full Text | Google Scholar

Benke, T., Köylü, B., Visani, P., Karner, E., Brenneis, C., Bartha, L., et al. (2006). Language lateralization in temporal lobe epilepsy: a comparison between fMRI and the Wada test. Epilepsia 47, 1308–1319. doi: 10.1111/j.1528-1167.2006.00549.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Berntsen, E. M., Rasmussen, I. A., Samuelsen, P., Xu, J., Haraldseth, O., Lagopoulos, J., et al. (2006). Putting the brain in Jeopardy: a novel comprehensive and expressive language task? Acta Neuropsychiatr. 18, 115–119. doi: 10.1111/j.1601-5215.2006.00134.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, J. R. (2011). Functional MRI is a valid noninvasive alternative to Wada testing. Epilepsy Behav. 20, 214–222. doi: 10.1016/j.yebeh.2010.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, J. R. (2016). “fMRI of language systems,” in Neuromethods, Vol. 119, eds S. H. Faro and F. B. Mohamed (New York, NY: Springer), 355–385. doi: 10.1007/978-1-4939-5611-1_12

CrossRef Full Text | Google Scholar

Binder, J. R., Desai, R. H., Graves, W. W., and Conant, L. L. (2009). Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cereb Cortex 19, 2767–2796.

Google Scholar

Bishop, D. V. M. (2013). Cerebral asymmetry and language development: cause, correlate, or consequence? Science 340:1230531. doi: 10.1126/science.1230531

PubMed Abstract | CrossRef Full Text | Google Scholar

Dehaene, S., and Cohen, L. (2011). The unique role of the visual word form area in reading. Trends Cogn. Sci. 15, 254–262. doi: 10.1016/j.tics.2011.04.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Deniz, F., Nunez-Elizalde, A. O., Huth, A. G., and Gallant, J. L. (2019). The representation of semantic information across human cerebral cortex during listening versus reading is invariant to stimulus modality. J. Neurosci. 39, 7722–7726. doi: 10.1523/JNEUROSCI.0675-19.2019

PubMed Abstract | CrossRef Full Text | Google Scholar

Deppe, M., Knecht, S., Papke, K., Lohmann, H., Fleischer, H., Heindel, W., et al. (2000). Assessment of hemispheric language lateralization: a comparison between fMRI and fTCD. J. Cereb. Blood Flow Metab. 20, 263–268. doi: 10.1097/00004647-200002000-00006

PubMed Abstract | CrossRef Full Text | Google Scholar

Devlin, J. T., and Aydelott, J. (2009). Speech perception: motoric contributions versus the motor theory. Curr. Biol. 19, R198–R200. doi: 10.1016/j.cub.2009.01.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Dronkers, N. F. (1996). A new brain region for coordinating speech articulation. Nature 384, 159–161.

Google Scholar

Eickhoff, S. B., Heim, S., Zilles, K., and Amunts, K. (2009). A systems perspective on the effective connectivity of overt speech production. Philos. Trans. R Soc. A Math. Phys. Eng. Sci. 367, 2399–2421. doi: 10.1098/rsta.2008.0287

PubMed Abstract | CrossRef Full Text | Google Scholar

Enard, W. (2011). FOXP2 and the role of cortico-basal ganglia circuits in speech and language evolution. Curr. Opin. Neurobiol. 21, 415–424. doi: 10.1016/j.conb.2011.04.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Engström, M., Ragnehed, M., Lundberg, P., and Söderfeldt, B. (2004). Paradigm design of sensory-motor and language tests in clinical fMRI. Neurophysiol. Clin. 34, 267–277. doi: 10.1016/j.neucli.2004.09.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Fedorenko, E., Behr, M. K., and Kanwisher, N. (2011). Functional specificity for high-level linguistic processing in the human brain. Proc. Natl. Acad. Sci. U. S. A. 108, 16428–16433. doi: 10.1073/pnas.1112937108

PubMed Abstract | CrossRef Full Text | Google Scholar

Fernández, G., Specht, K., Weis, S., Tendolkar, I., Reuber, M., Fell, J., et al. (2003). Intrasubject reproducibility of presurgical language lateralization and mapping using fMRI. Neurology 60, 969–975. doi: 10.1212/01.WNL.0000049934.34209.2E

PubMed Abstract | CrossRef Full Text | Google Scholar

Frässle, S., Stephan, K. E., Friston, K. J., Steup, M., Krach, S., Paulus, F. M., et al. (2015). Test-retest reliability of dynamic causal modeling for fMRI. Neuroimage 117, 56–66. doi: 10.1016/j.neuroimage.2015.05.040

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J. (2010). The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11, 127–138. doi: 10.1038/nrn2787

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J., Preller, K. H., Mathys, C., Cagnan, H., Heinzle, J., Razi, A., et al. (2019). Dynamic causal modelling revisited. Neuroimage 199, 730–744. doi: 10.1016/j.neuroimage.2017.02.045

PubMed Abstract | CrossRef Full Text | Google Scholar

Galantucci, B., Fowler, C. A., and Turvey, M. T. (2006). The motor theory of speech perception reviewed. Psychon. Bull. Rev. 13, 361–377. doi: 10.3758/bf03193857

PubMed Abstract | CrossRef Full Text | Google Scholar

Graessner, A., Zaccarella, E., and Hartwigsen, G. (2021). Differential contributions of left-hemispheric language regions to basic semantic composition. Brain Struct. Funct. 226, 501–518. doi: 10.1007/s00429-020-02196-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Grande, M., Meffert, E., Schoenberger, E., Jung, S., Frauenrath, T., Huber, W., et al. (2012). From a concept to a word in a syntactically complete sentence: an fMRI study on spontaneous language production in an overt picture description task. Neuroimage 61, 702–714. doi: 10.1016/j.neuroimage.2012.03.087

PubMed Abstract | CrossRef Full Text | Google Scholar

Hausmann, M., Brysbaert, M., van der Haegen, L., Lewald, J., Specht, K., Hirnstein, M., et al. (2019). Language lateralisation measured across linguistic and national boundaries. Cortex 111, 134–147. doi: 10.1016/j.cortex.2018.10.020

PubMed Abstract | CrossRef Full Text | Google Scholar

Heim, S., Eickhoff, S. B., Ischebeck, A. K., Friederici, A. D., Stephan, K. E., and Amunts, K. (2009). Effective connectivity of the left BA 44, BA 45, and inferior temporal gyrus during lexical and phonological decisions identified with DCM. Hum. Brain Mapp. 30, 392–402. doi: 10.1002/hbm.20512

PubMed Abstract | CrossRef Full Text | Google Scholar

Hickok, G. S., and Poeppel, D. (2000). Towards a functional neuroanatomy of speech perception. Trends Cogn. Sci. 4, 131–138. doi: 10.1016/s1364-6613(00)01463-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Hickok, G. S., and Poeppel, D. (2004). Dorsal and ventral streams: a framework for understanding aspects of the functional anatomy of language. Cognition 92, 67–99. doi: 10.1016/j.cognition.2003.10.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Hickok, G. S., and Poeppel, D. (2007). The cortical organization of speech processing. Nat. Rev. Neurosci. 8, 393–402.

Google Scholar

Jansen, A., Menke, R., Sommer, J., Förster, A. F., Bruchmann, S., Hempleman, J., et al. (2006). The assessment of hemispheric lateralization in functional MRI–robustness and reproducibility. Neuroimage 33, 204–217. doi: 10.1016/j.neuroimage.2006.06.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Kotz, S. A., and Schmidt-Kassow, M. (2015). Basal ganglia contribution to rule expectancy and temporal predictability in speech. Cortex 68, 48–60. doi: 10.1016/j.cortex.2015.02.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Kotz, S. A., Schwartze, M., and Schmidt-Kassow, M. (2009). Non-motor basal ganglia functions: a review and proposal for a model of sensory predictability in auditory language perception. Cortex 45, 982–990. doi: 10.1016/j.cortex.2009.02.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Liberman, A. M., and Mattingly, I. G. (1985). The motor theory of speech perception revised. Cognition 21, 1–36. doi: 10.1016/0010-0277(85)90021-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Lim, S.-J., Fiez, J. A., and Holt, L. L. (2014). How may the basal ganglia contribute to auditory categorization and speech perception? Front. Neurosci. 8:230 doi: 10.3389/fnins.2014.00230

PubMed Abstract | CrossRef Full Text | Google Scholar

Lima, C. F., Krishnan, S., and Scott, S. K. (2016). Roles of supplementary motor areas in auditory processing and auditory imagery. Trends Neurosci. 39, 527–542. doi: 10.1016/j.tins.2016.06.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Mariën, P., Ackermann, H., Adamaszek, M., Barwood, C. H. S., Beaton, A., Desmond, J., et al. (2013). Consensus paper: language and the cerebellum: an ongoing enigma. Cerebellum 13, 386–410. doi: 10.1007/s12311-013-0540-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Matchin, W., Hammerly, C., and Lau, E. (2017). The role of the IFG and pSTS in syntactic prediction: evidence from a parametric study of hierarchical structure in fMRI. Cortex 88, 106–123. doi: 10.1016/j.cortex.2016.12.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Maxwell, S. E., Lau, M. Y., and Howard, G. S. (2015). Is psychology suffering from a replication crisis? What does “failure to replicate” really mean? Am. Psychol. 70, 487–498. doi: 10.1037/a0039400

PubMed Abstract | CrossRef Full Text | Google Scholar

McCandliss, B. D., Cohen, L., and Dehaene, S. (2003). The visual word form area: expertise for reading in the fusiform gyrus. Trends Cogn. Sci. 7, 293–299. doi: 10.1016/s1364-6613(03)00134-7

PubMed Abstract | CrossRef Full Text | Google Scholar

McGettigan, C., and Scott, S. K. (2012). Cortical asymmetries in speech perception: what’s wrong, what’s right and what’s left? Trends Cogn. Sci. 16, 269–276. doi: 10.1016/j.tics.2012.04.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Morken, F., Helland, T., Hugdahl, K., and Specht, K. (2017). Reading in dyslexia across literacy development: a longitudinal study of effective connectivity. Neuroimage 144, 92–100. doi: 10.1016/j.neuroimage.2016.09.060

PubMed Abstract | CrossRef Full Text | Google Scholar

Nachev, P., Kennard, C., and Husain, M. (2008). Functional role of the supplementary and pre-supplementary motor areas. Nat. Rev. Neurosci. 9, 856–869. doi: 10.1038/nrn2478

PubMed Abstract | CrossRef Full Text | Google Scholar

Nichols, T. E., Brett, M., Andersson, J., Wager, T., and Poline, J.-B. (2005). Valid conjunction inference with the minimum statistic. Neuroimage 25, 653–660.

Google Scholar

Nieuwenhuys, R. (2012). The insular cortex. A review. Prog. Brain Res. 195, 123–163.

Google Scholar

Niskanen, E., Könönen, M., Villberg, V., Nissi, M., Ranta-Aho, P., Säisänen, L., et al. (2012). The effect of fMRI task combinations on determining the hemispheric dominance of language functions. Neuroradiology 54, 393–405. doi: 10.1007/s00234-011-0959-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Orringer, D. A., Vago, D. R., and Golby, A. J. (2012). Clinical applications and future directions of functional MRI. Semin. Neurol. 32, 466–475. doi: 10.1055/s-0032-1331816

PubMed Abstract | CrossRef Full Text | Google Scholar

Osnes, B., Hugdahl, K., Hjelmervik, H., and Specht, K. (2011a). Increased activation in superior temporal gyri as a function of increment in phonetic features. Brain Lang. 116, 97–101. doi: 10.1016/j.bandl.2010.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Osnes, B., Hugdahl, K., and Specht, K. (2011b). Effective connectivity analysis demonstrates involvement of premotor cortex during speech perception. Neuroimage 54, 2437–2445. doi: 10.1016/j.neuroimage.2010.09.078

PubMed Abstract | CrossRef Full Text | Google Scholar

Patterson, K., Nestor, P. J., and Rogers, T. T. (2007). Where do you know what you know? The representation of semantic knowledge in the human brain. Nat. Rev. Neurosci. 8, 976–987. doi: 10.1038/nrn2277

PubMed Abstract | CrossRef Full Text | Google Scholar

Penny, W. D. (2012). Bayesian models of brain and behaviour. ISRN Biomath. 2012, 1–19. doi: 10.5402/2012/785791

CrossRef Full Text | Google Scholar

Penny, W. D., Stephan, K. E., Daunizeau, J., Rosa, M. J., Friston, K. J., Schofield, T. M., et al. (2010). Comparing families of dynamic causal models. PLoS Comput. Biol. 6:e1000709. doi: 10.1371/journal.pcbi.1000709

PubMed Abstract | CrossRef Full Text | Google Scholar

Poeppel, D., Emmorey, K., Hickok, G. S., and Pylkkänen, L. (2012). Towards a new neurobiology of language. J. Neurosci. 32, 14125–14131. doi: 10.1523/JNEUROSCI.3244-12.2012

PubMed Abstract | CrossRef Full Text | Google Scholar

Poeppel, D., and Hickok, G. S. (2004). Towards a new functional anatomy of language. Cognition 92, 1–12. doi: 10.1016/j.cognition.2003.11.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Power, J. D., Barnes, K. A., Snyder, A. Z., Schlaggar, B. L., and Petersen, S. E. (2012). Spurious but systematic correlations in functional connectivity MRI networks arise from subject motion. Neuroimage 59, 2142–2154. doi: 10.1016/j.neuroimage.2011.10.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Price, C. J. (2010). The anatomy of language: a review of 100 fMRI studies published in 2009. Ann. N. Y. Acad. Sci. 1191, 62–88. doi: 10.1111/j.1749-6632.2010.05444.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Price, C. J. (2012). A review and synthesis of the first 20 years of PET and fMRI studies of heard speech, spoken language and reading. Neuroimage 62, 816–847. doi: 10.1016/j.neuroimage.2012.04.062

PubMed Abstract | CrossRef Full Text | Google Scholar

Rauschecker, J. P., and Scott, S. K. (2009). Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat. Neurosci. 12, 718–724. doi: 10.1038/nn.2331

PubMed Abstract | CrossRef Full Text | Google Scholar

Rodriguez, S. M., Archila-Suerte, P., Vaughn, K. A., Chiarello, C., and Hernandez, A. E. (2018). Anterior insular thickness predicts speech sound learning ability in bilinguals. Neuroimage 165, 278–284. doi: 10.1016/j.neuroimage.2017.10.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Rolls, E. T., Huang, C. C., Lin, C. P., Feng, J., and Joliot, M. (2020). Automated anatomical labelling atlas 3. Neuroimage 206:116189. doi: 10.1016/j.neuroimage.2019.116189

PubMed Abstract | CrossRef Full Text | Google Scholar

Sair, H. I., Yahyavi-Firouz-Abadi, N., Calhoun, V. D., Airan, R. D., Agarwal, S., Intrapiromkul, J., et al. (2016). Presurgical brain mapping of the language network in patients with brain tumors using resting-state fMRI: comparison with task fMRI. Hum. Brain Mapp. 37, 913–923. doi: 10.1002/hbm.23075

PubMed Abstract | CrossRef Full Text | Google Scholar

Sammler, D., Grosbras, M.-H. H., Anwander, A., Bestelmeyer, P. E. G., and Belin, P. (2015). Dorsal and ventral pathways for prosody. Curr. Biol. 25, 3079–3085.

Google Scholar

Scott, S. K. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain 123, 2400–2406.

Google Scholar

Shaw, T. C. (1874). On aphasia. Br. Med. J. 1, 574–575.

Google Scholar

Shrout, P. E., and Fleiss, J. L. (1979). Intraclass correlations: uses in assessing rater reliability. Psychol. Bull. 86, 420–428. doi: 10.1037//0033-2909.86.2.420

PubMed Abstract | CrossRef Full Text | Google Scholar

Silva, M. A., See, A. P., Essayed, W. I., Golby, A. J., and Tie, Y. (2018). Challenges and techniques for presurgical brain mapping with functional MRI. Neuroimage. Clin. 17, 794–803. doi: 10.1016/j.nicl.2017.12.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K. (2013). Mapping a lateralization gradient within the ventral stream for auditory speech perception. Front. Hum. Neurosci. 7:629. doi: 10.3389/fnhum.2013.00629

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K. (2014). Neuronal basis of speech comprehension. Hear. Res. 307, 121–135. doi: 10.1016/j.heares.2013.09.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K. (2020). Current challenges in translational and clinical fMRI and future directions. Front. Psychiatry 10:924. doi: 10.3389/fpsyt.2019.00924/full

CrossRef Full Text | Google Scholar

Specht, K., Osnes, B., and Hugdahl, K. (2009). Detection of differential speech-specific processes in the temporal lobe using fMRI and a dynamic “sound morphing” technique. Hum. Brain Mapp. 30, 3436–3444. doi: 10.1002/hbm.20768

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K., Rimol, L. M. L. M. L. M., Reul, J., and Hugdahl, K. (2005). “Soundmorphing”: a new approach to studying speech perception in humans. Neurosci. Lett. 384, 60–65. doi: 10.1016/j.neulet.2005.04.057

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K., and Wigglesworth, P. (2018). The functional and structural asymmetries of the superior temporal sulcus. Scand. J. Psychol. 59, 74–82. doi: 10.1111/sjop.12410

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K., Willmes, K., Shah, N. J., and Jäncke, L. (2003). Assessment of reliability in functional imaging studies. J. Magn. Reson. Imaging 17, 463–471. doi: 10.1002/jmri.10277

PubMed Abstract | CrossRef Full Text | Google Scholar

Stephan, K. E., Penny, W. D., Daunizeau, J., Moran, R. J., and Friston, K. J. (2009). Bayesian model selection for group studies. Neuroimage 46, 1004–1017.

Google Scholar

Tie, Y., Rigolo, L., Norton, I. H., Huang, R. Y., Wu, W., Orringer, D., et al. (2014). Defining language networks from resting-state fMRI for surgical planning- A feasibility study. Hum. Brain Mapp. doi: 10.1002/hbm.22231

PubMed Abstract | CrossRef Full Text | Google Scholar

Tzourio-Mazoyer, N., Landeau, B., Papathanassiou, D., Crivello, F., Etard, O., Delcroix, N., et al. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage 15, 273–289.

Google Scholar

Ueno, T., Saito, S., Rogers, T. T., and Lambon Ralph, M. A. (2011). Lichtheim 2: synthesizing aphasia and the neural basis of language in a neurocomputational model of the dual dorsal-ventral language pathways. Neuron 72, 385–396. doi: 10.1016/j.neuron.2011.09.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Urbach, H., Mast, H., Egger, K., and Mader, I. (2015). Presurgical MR Imaging in Epilepsy. Clin. Neuroradiol. 25, 151–155.

Google Scholar

Van der Haegen, L., Cai, Q., Seurinck, R., and Brysbaert, M. (2011). Further fMRI validation of the visual half field technique as an indicator of language laterality: a large-group analysis. Neuropsychologia 49, 2879–2888. doi: 10.1016/j.neuropsychologia.2011.06.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Wada, J., and Rasmussen, T. (2007). Intracarotid injection of sodium amytal for the lateralization of cerebral speech dominance. J. Neurosurg. 17, 266–282. doi: 10.3171/jns.2007.106.6.1117

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilke, M., and Lidzba, K. (2007). LI-tool: a new toolbox to assess lateralization in functional MR-data. J. Neurosci. Methods 163:9.

Google Scholar

Wilson, S. M., DeMarco, A. T., Henry, M. L., Gesierich, B., Babiak, M., Mandelli, M. L., et al. (2014). What role does the anterior temporal lobe play in sentence-level processing? Neural correlates of syntactic processing in semantic variant primary progressive aphasia. J. Cogn. Neurosci. 26, 970–985. doi: 10.1162/jocn_a_00550

PubMed Abstract | CrossRef Full Text | Google Scholar

Zaccarella, E., and Friederici, A. D. (2017). The neurobiological nature of syntactic hierarchies. Neurosci. Biobehav. Rev. 81, 205–212. doi: 10.1016/j.neubiorev.2016.07.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Zahn, R., Buechert, M., Overmans, J., Talazko, J., Specht, K., Ko, C.-W. W. C.-W. C.-W. W., et al. (2005). Mapping of temporal and parietal cortex in progressive nonfluent aphasia and Alzheimer’s disease using chemical shift imaging, voxel-based morphometry and positron emission tomography. Psychiatry Res. Neuroimaging 140, 115–131. doi: 10.1016/j.pscychresns.2005.08.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Zeidman, P., Jafarian, A., Corbin, N., Seghier, M. L., Razi, A., Price, C. J., et al. (2019). A guide to group effective connectivity analysis, part 1: first level analysis with DCM for fMRI. Neuroimage 200, 174–190. doi: 10.1016/j.neuroimage.2019.06.031

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: functional magnetic resonance imaging, fMRI, language network, asymmetry, speech perception, clinical fMRI, reliability, dynamic causal modeling

Citation: Rødland E, Melleby KM and Specht K (2022) Evaluation of a Simple Clinical Language Paradigm With Respect to Sensory Independency, Functional Asymmetry, and Effective Connectivity. Front. Behav. Neurosci. 16:806520. doi: 10.3389/fnbeh.2022.806520

Received: 31 October 2021; Accepted: 10 February 2022;
Published: 03 March 2022.

Edited by:

Alfredo Brancucci, Foro Italico University of Rome, Italy

Reviewed by:

Emiliano Zaccarella, Max Planck Institute for Human Cognitive and Brain Sciences, Germany
Maxim Kireev, N. P. Bechtereva Institute of the Human Brain (RAS), Russia

Copyright © 2022 Rødland, Melleby and Specht. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Karsten Specht, Karsten.specht@uib.no

Download