ORIGINAL RESEARCH article

Front. Hum. Neurosci., 04 October 2013

Sec. Speech and Language

Volume 7 - 2013 | https://doi.org/10.3389/fnhum.2013.00634

The role of accent imitation in sensorimotor integration during processing of intelligible speech

  • PA

    Patti Adank 1,2*

  • SR

    Shirley-Ann Rueschemeyer 2,3

  • HB

    Harold Bekkering 2

  • 1. Department of Speech, Hearing and Phonetic Sciences, Division of Psychology and Language Sciences, University College London London, UK

  • 2. Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen Nijmegen, Netherlands

  • 3. Department of Psychology, University of York York, UK

Abstract

Recent theories on how listeners maintain perceptual invariance despite variation in the speech signal allocate a prominent role to imitation mechanisms. Notably, these simulation accounts propose that motor mechanisms support perception of ambiguous or noisy signals. Indeed, imitation of ambiguous signals, e.g., accented speech, has been found to aid effective speech comprehension. Here, we explored the possibility that imitation in speech benefits perception by increasing activation in speech perception and production areas. Participants rated the intelligibility of sentences spoken in an unfamiliar accent of Dutch in a functional Magnetic Resonance Imaging experiment. Next, participants in one group repeated the sentences in their own accent, while a second group vocally imitated the accent. Finally, both groups rated the intelligibility of accented sentences in a post-test. The neuroimaging results showed an interaction between type of training and pre- and post-test sessions in left Inferior Frontal Gyrus, Supplementary Motor Area, and left Superior Temporal Sulcus. Although alternative explanations such as task engagement and fatigue need to be considered as well, the results suggest that imitation may aid effective speech comprehension by supporting sensorimotor integration.

Introduction

In everyday communication, humans have to deal with a variety of challenging listening situations, such as the presence of background noise, poor telephone connections, or unfamiliar accents. Recent studies have investigated cognitive and neural mechanisms underlying the ability to process speech effectively in such challenging listening situations (Davis and Johnsrude, 2003; Rodd et al., 2005, 2010; Obleser et al., 2007, 2011; Adank and Devlin, 2010; Obleser and Kotz, 2010; Peelle et al., 2010; Adank, 2012). The aforementioned studies have established that effective speech comprehension recruits areas involved in speech perception including left Superior Temporal Sulcus (STS), areas involved in linguistic and articulatory processes including left Inferior Frontal Gyrus (IFG), areas involved in speech production including Precentral Gyrus and Supplementary Motor Area (SMA).

There is growing consensus that cortical regions associated with speech production, such as ventral premotor cortex, IFG, SMA/pre-SMA, and primary motor cortex play an active and essential role in effective speech comprehension (Skipper et al., 2006; D'Ausilio et al., 2009; Londei et al., 2009; Callan et al., 2010; Tremblay and Small, 2011; Adank, 2012). Despite this emerging agreement (but see Venezia et al., 2012), much is still unclear about precisely how speech motor and premotor regions contribute to the comprehension process.

Recent theories suggests that mental simulation of perceived actions may aid listeners when predicting upcoming speech signals (Wilson and Knoblich, 2005; Pickering and Garrod, 2007). Simulation theories of action perception argue that observing actions results in the automatic generation of motor plans required to perform these actions. The motor plans are used to produce a forward model of the incoming action. These forward models use information about the movement properties of muscles to generate a simulated course of movement in parallel with, or even in anticipation of, the perceived movement (Grush, 2004). Any discrepancy between the simulated movement from the forward model and the real-world movement results in prediction errors and leads to corrective commands. This type of forward model serves to anticipate others' (speech) actions as if they were produced by the observer (Locatelli et al., 2012). Forward models thus, generate a series of predictions that may be used as disambiguating information in situations of when action perception is made more difficult due to noisy or ambiguous observing conditions (Wilson and Knoblich, 2005; Pickering and Garrod, 2007). For speech, these type of conditions involve listening to speech in the presence of background noise, or listening to someone speak with an unfamiliar regional accent (Adank et al., 2009).

Some propose that the prediction signal generated by the forward models is optimized for the type of actions the observers have had experience executing themselves (Knoblich and Sebanz, 2006; Grafton, 2009). Behavioral evidence for this prediction comes from a study on basketball players by Aglioti et al. (2008). Aglioti et al. compared elite basketball players (experts at observing the action and also at performing the action) with a control group (basketball referees; experts at observing the action but not playing themselves) on how effectively each group could judge whether a basketball would be thrown in the basket. They showed that being experienced in performing an action such as throwing a basketball allows one to efficiently predict the outcome of other players throwing a basketball.

For speech, we recently found that short-term experience with speaking in an unfamiliar regional accent helps comprehension of sentences spoken in that accent (Adank et al., 2010). Adank et al.'s participants listened to sentences spoken in an unfamiliar accent in background noise in a pre-test phase and verbally repeated key words. Next, participants were divided into six groups and either received no training, listened to sentences in the unfamiliar accent without speaking, repeated the accented sentences in their own accent, listened to and transcribed the accented sentences, listened to and imitated the accented sentences, or listened to and imitated the accented sentences without being able to hear their own vocalizations. Post-training measures showed that participants who imitated the speaker's accent could repeat key words in poorer signal-to-noise ratios (i.e., under more challenging listening conditions) than participants who had not imitated the accent. Adank et al.'s results demonstrate that having experience with speaking in a specific way (i.e., in an unfamiliar accent) can positively affect speech processing by optimizing comprehension of sentences spoken in a similar way.

The neural underpinnings of the optimizing effect of experience with performing an action on action perception have been investigated a several fMRI experiments (Calvo-Merino et al., 2005; Wilson and Iacoboni, 2006; Lahav et al., 2007). Calvo-Merino et al. (2005) studied professional dancers and compared neural activity when male and female dancers viewed dance moves that were performed by their own gender (i.e., moves they would perform when dancing themselves), compared to other moves performed by the other gender (i.e., moves they would always observe, but not perform, when dancing themselves). Calvo-Merino's design allowed them to separate brain responses related to dance motor representation from those related to visual knowledge about the observed moves. Left dorsal premotor cortex, left inferior parietal sulcus, and right cerebellum showed a higher Blood Oxygenated-Level Dependent (BOLD-) response for dance moves they would perform themselves than for observed moves.

Lahav et al. (2007) trained non-musicians in playing several simple melodies on the piano. Participants were scanned while listening to pieces they had learnt to play, plus pieces they had not learnt. An increased BOLD-response was found in bilateral IFG, the posterior middle premotor region and the inferior parietal lobule (IPL) (Supramarginal gyrus and Angular gyrus) bilaterally and left cerebellum.

Finally, Wilson and Iacoboni tested neural responses for 5 native and 25 non-native speech sounds varying in how easy they were to produce for their native American-English speaking participants. They report increased activation in bilateral superior temporal areas for foreign speech sounds that are more difficult to produce. No correlations were found between the ease with which the foreign sounds could be produced in speech motor areas. Nevertheless, in a region-of-interest analysis, they report that the BOLD-response in premotor cortex, a speech motor area, was elevated for listening to foreign speech sounds compared to native speech sounds. The results from Wilson and Iacoboni thus, contradict the expectation that having had more experience with a motor action (which is the case for native speech sounds) activates motor areas to a higher degree. However, Wilson did not explicitly evaluate how experience with producing the foreign speech sounds affected brain activation when listening to those speech sounds. Thus, it remains unclear how speech motor experience affects neural activation during speech comprehension.

The present study will examine how motor experience with imitating a novel accent affects the activation of cortical areas associated with speech perception and production when subsequently listening to the accented speech in background noise. Based on Adank et al.'s behavioral study, we hypothesized that speech imitation experience supports speech comprehension through the integrating of information from specific speech perception and production areas.

We mapped out the neural bases associated with increased robustness of speech comprehension after imitating an unfamiliar accent and after repeating the accented speech in one's own accent. Two groups of listeners were scanned using an adapted version of the staircase procedure described in Adank et al. (2010). In a pre-test, participants were examined on their comprehension of sentences spoken in an unfamiliar accent in background noise. Next, one group of participants repeated a series of accented sentences in their own accent, while the second group vocally imitated the sentences in the unfamiliar accent in a training session. Finally, both groups were tested again on their comprehension of accented speech in background noise. If imitation supports sensorimotor integration, we expected a different pattern of activation of cortical areas associated with either speech comprehension or speech production for listeners who have had experience with imitating the unfamiliar accent during the training session. We focused on three left-lateralized regions, namely posterior STS, IFG, and SMA, all of which have been associated with speech perception and speech production tasks (for STS: Scott et al., 2000; Blank et al., 2002; Narain et al., 2003; Crinion and Price, 2005; Tremblay and Gracco, 2009, for IFG: Davis and Johnsrude, 2003; Tremblay and Gracco, 2009; Adank and Devlin, 2010, for SMA: Alario et al., 2006; Wong et al., 2008; Fridriksson et al., 2009).

Materials and methods

Participants

We tested 36 participants (23F and 13M, mean 21.8 years, SD 2.9 years, range 18–29 years). All participants were right-handed, native speakers of Dutch, with no history of oral or written language impairment, or neurological or psychiatric disease. All gave written informed consent and were paid for their participation. The study was approved by the local ethics committee of the Radboud University Nijmegen. Four participants were excluded from further analysis: one participant (F) originally allocated to the imitation group was excluded as the second run was not collected due to technical difficulties, another participant allocated to the imitation group (M) was excluded due to an abnormality in his structural scan, a repeat group participant (M) was excluded as his scan was interrupted due to him feeling unwell, and a final participant (F) in the repeat group was excluded due to poor performance on the behavioral task; her results showed more than 50% missed responses. All analyses were run on the results from the remaining 32 participants. The repeat group consisted of 15 participants (9F and 6M, mean 21.6 years, SD 3.2 years, range 18–29 years) and the imitation group consisted of 17 participants (10F and 6M, mean 22.2 years, SD 2.8 years, range 18–28 years).

Stimulus material

The test stimulus set consisted of 96 sentences (see Table A1 in the Appendix) that were taken from the speech reception threshold (SRT) corpus (Plomp and Mimpen, 1979). This corpus has been widely used for assessing intelligibility of different types of stimuli, for example, for foreign-accented speech (Van Wijngaarden et al., 2002). Sentences in the SRT corpus are designed to resemble short samples of conversational speech. All consist of maximally eight or nine syllables and do not include words longer than three syllables. The 48 sentences used in the training (listed in Table A2) were taken from an earlier study on comprehension of accented speech (Adank et al., 2012) and produced by the same speaker as the 96 test sentences. Finally, 30 sentences were recorded in Standard Dutch and used in a hearing pre-test (cf. Table A3).

All sentences were spoken by a single female speaker of Standard Dutch. The test sentences were recorded in both Standard Dutch and an unfamiliar, novel accent. The training sentences were recorded in the novel accent only. The novel accent was created by instructing the speaker to read sentences with an adapted orthography (see also Adank et al., 2010). This accent was designed to sound different from Standard Dutch and not intended to replicate an existing Dutch accent. The orthography was systematically altered to achieve changes in all 15 Dutch vowels as listed in Table 1. Only vowels bearing primary or secondary stress were included in the orthography conversion. An example of a sentence in Standard Dutch and a converted version is given below, including a broad phonetic transcription using the International Phonetic Alphabet (IPA, 1999):

Table 1


            
              Intended vowel conversions for obtaining the novel accent.

Intended vowel conversions for obtaining the novel accent.

The left column shows the altered orthography in Standard Dutch, and the right column shows the intended change in pronunciation of the vowel in broad phonetic transcription, using the International Phonetic Alphabet (IPA, 1999).

  • Standard Dutch: “De bal vloog over de schutting”

  •                           [The ball flew over the fence]

  • After conversion: “De baal flog offer de schuuttieng”

The stimulus materials used in the scanner was created as follows. The sentences in background noise were created by adding continuous speech-shaped noise to the accented sentences in quiet at so that the final signal-to-noise ratio (SNR) was set to −2 dB, 0 dB, +2 dB, +4 dB, +6 dB, +8 dB, +10 dB, +12 dB, +14 dB, or +16 dB, thus, resulting in 10 versions for each sentence. Speech-shaped noise was added using Matlab (Mathworks Inc.). An acoustically matched—yet unintelligible—baseline version of each sentence was created by spectrally rotating (Blesser, 1972) and time-reversing the sentence, using Praat (Boersma and Weenink, 2003). No noise was added to the baseline sentences. Sentences in all conditions were subsequently saved at 70 dB SPL.

MRI data acquisition

Whole-brain imaging was performed at the Donders Centre for Brain, Cognition and Behavior, Centre for Cognitive Neuroimaging, at a 3T MR scanner (Magnetom Trio, Siemens Medical Systems, Erlangen, Germany), in Nijmegen, The Netherlands. The sentences were presented over sound-attenuating (~30 dB SPL) electrostatic headphones (MRConFon, Magdeburg, Germany) during continuous scanner acquisition (GE-EPI, echo time 35 ms; 32 axial slices; voxel size 3.5 × 3.5 × 3.5 mm, slice thickness 3 mm, inter-slice gap 0.5 mm; field of view 224 mm; flip angle 70). All functional images were acquired in two runs. Functional scans were obtained every 2 s in a continuous scanning paradigm (TR 2 s). Listeners watched a fixation cross that was presented on a screen and viewed through a mirror attached to the head coil. After the acquisition of functional images, a high-resolution structural scan was acquired (T1-weighted MP-RAGE, 192 slices, echo time 3.93 ms; field of view 256 mm, slice thickness 1 mm). Finally, a diffusion-weighted scan was acquired with 68 directions (not included in the present analysis). Total scanning time was around 60 min.

Procedure

The experiment had a mixed-subject design with two groups (Figure 1): Repeat or Imitate. Participants were randomly allocated to the Repeat or Imitation group. Imaging data was collected in two runs and participants performed a training task in between these runs. Before the start of the scanning session, participants completed a hearing test in a quiet room. This test established the SRT as described in Plomp and Mimpen (1979) for sentences spoken in Standard Dutch by the same female speakers as used for the test in the scanner. Using a similar procedure as in (Adank et al., 2010), participants responded by repeating back what they had heard and the experimenter scored key words (see Table A1 in the Appendix). In this “standard” version, of the SRT, participants were presented the first sentence at +10 dB SNR. If they correctly repeated more than two out of four keywords, the next sentence was played at +2 dB SNR, if they got more than two key words in this sentence correct, the next sentence was played at −6 dB. This procedure continued until the participant either got two keywords right, or if he or she got fewer than two keywords right. If they repeated two keywords correctly, the next sentence was played at the same SNR as the previous sentence, while the SNR was increased with +2 dB if fewer than two keywords were repeated. If the participant than got more than two keywords right, the SNR decreased in steps of −2 dB. This staircase procedure was repeated for 30 sentences (Table A3). The final SNR, or SRT, was calculated as the average across all instances for which a reversal occurred, following Adank et al. (2010) and Plomp and Mimpen (1979). A reversal occurs whenever the SNR changed direction (e.g., from 0 dB SNR to +2 db SNR, or to −2 dB SNR). The participants showed an average SRT of −3.69 dB (0.95 dB SD), indicating that their hearing was in the normal range for participants with normal hearing (Plomp and Mimpen, 1979).

Figure 1

While in the scanner, participants in both groups performed an intelligibility judgment in the first run. Following the procedure in Adank et al. (2010), participants in the repeat group were instructed to listen to each accented sentence and then to repeat it in their own accent, without imitating the accent. Participants were explicitly instructed not to imitate the speaker's accent. If participants imitated the accent, they were reminded by the experimenter to repeat the sentence in Standard Dutch. Participants repeated 48 sentences. In the imitation group, participants were instructed to imitate vocally the precise pronunciation of the accented sentence. If participants repeated the sentence in their own accent, the experimenter instructed to imitate the accent as they heard it spoken. Participants imitated 48 sentences during the training and training was not attested using a formal scoring device. The scanner was turned off during the training phase. Next, the second run commenced and participants performed the intelligibility judgment task from run 1 for another series of accented sentenced. A total of 96 test sentences was presented across the two runs, plus 48 training sentences in between runs. Per run, 24 acoustically-matched baseline stimuli and 48 test sentences were presented. The distribution of the test sentences and baseline stimuli (which were created based on the 96 test sentences) were counterbalanced across runs and participants.

A single trial in the intelligibility judgment task began with a tone signal of 250 Hz with a duration of 200 ms, followed by a pause of 200 ms, followed by the presentation of a sentence. After the sentence was presented, participants judged its intelligibility, using a button-box with four buttons (one per finger) that they were holding in their right hand. If they found the sentence completely unintelligible, they pressed with their index finger (score 1), if they found the sentence slightly more intelligible they pressed the button under their middle finger (score 2), if they found the sentence intelligible they pressed with their ring finger (score 3) and if they found the sentence very intelligible they pressed the button under their little finger (score 4). This procedure was the same across participants. The SNR of the following sentence depended on the score of the previous sentence. If the participant had rated a specific sentence as 1 or 2, the next sentence was played at an easier (higher) SNR, and if the participant had rated a specific sentence as 3 or 4, the next sentence was played at a less favorable SNR. For instance, if a participant heard a sentence with a SNR of +4 dB and rated the sentence as unintelligible (score 2), then SNR for the next sentence was increased with 2 dB and the next sentence was presented at +6 dB. Alternatively, if a participant heard a sentence with a SNR of +4 dB and rated the sentence as very intelligible (score 4), then SNR for the next sentence was decreased with 2 dB and the next sentence was presented at +2 dB. Each run started with a sentence presented at +10 dB. The SNR decreased in steps of 2 dB until reaching −2 dB SNR and increased in steps of 2 dB up to +16 dB SNR. If these limits were reached, the SNR stayed the same until the participant pressed 3 or 4 (for +16 dB) or 1 or 2 (for −2 dB). As in the behavioral test outside the scanner, SRT was calculated as the average across all instances for which a reversal occurred. This procedure was identical across both runs. Intensity of stimulus presentation was set at a comfortable level for each participant in a familiarization session in which six sentences in Standard Dutch in quiet spoken by the same speaker (not included in the both runs experiment) were presented while the scanner was running. Stimulus presentation was performed using Presentation (Neurobehavioral Systems, Albany, CA), running on a Pentium 4 with 2 GB RAM, and a 2.8 GHz processor. The two runs plus training lasted ~40 min.

Data processing

The neuroimaging data were pre-processed and analyzed using SPM8 (Wellcome Imaging Department, University College London, London, UK). The first four volumes of every functional run from each participant were excluded from the analysis to minimize T1-saturation effects. Next, the image time series were spatially realigned using a least-squares approach that estimates six rigid-body transformation parameters (Friston et al., 1995) by minimizing head movements between each image and the reference image, that is, the first image in the time series. Next, the time series for each voxel was temporally realigned to acquisition of the middle slice. Subsequently, images were normalized onto a custom Montreal Neurological Institute (MNI)-aligned EPI template (standard in SPM8) using both linear and non-linear transformations and resampled at an isotropic voxel size of 2 mm. All participants' functional images were smoothed using an 8 mm FWHM Gaussian filter. Each participants' functional image was processed using a unified segmentation procedure as implemented in SPM8. After segmentation of the T1 structural image (using the unified segmentation model) and co-registration to the mean functional image (Ashburner and Friston, 1997), functional images were spatially normalized into standard stereotaxic space of the MNI using segmentation parameters of the T1 structural image. A high-pass filter was applied with a 0.0078 Hz (128 s) cut-off to remove low-frequency components from the data, such as scanner drifts. The fMRI time series were analyzed within the context of the General Linear Model using an event-related approach, including an autoregressive AR (1) model during Classical (ReML) parameter estimation.

Three events of interest were identified and entered into a subject-specific General Linear Model: acoustically matched baseline sentences (Baseline), Unintelligible Speech (i.e., all sentences rated 1 or 2 by the participant), Intelligible Speech (i.e., all sentences rated 3 or 4 by the participant) for each run. Parameter estimates were calculated for each voxel, and contrast maps were constructed for each participant. The statistical model also considered six separate covariates describing the head-related movements (as estimated by the spatial realignment procedure). Linear-weighted contrasts were used to specify the main contrasts.

Intelligibility (having two levels: Intelligible Speech and Baseline), Run (Run 1 or Run 2), and Group (Repeat and Imitate) were analyzed in a 2 × 2 × 2 mixed ANOVA design, with Run as a within-subject and Group as a between-subject factor. We first identified areas showing a higher BOLD response for Intelligible Speech than for Baseline. Second, we identified areas that showed a higher BOLD-response as a result of having imitated during the training. We compared the activation across both runs for both groups. We used the following contrast to identify areas that showed a higher BOLD-response for the Imitate group than the Repeat group by directly comparing the second to the first run for each group: [Intelligible speech Imitation Group Run 2 > Intelligible speech Imitation Group Run 1] > [Intelligible speech Repeat Group Run 2 > Intelligible speech Repeat Group Run 1] (in short: [Imitation Run 2 > 1] > [Repeat Run 2 > 1]). Second, we identified areas showing a higher BOLD-response as a result of having repeated vs. imitated during the training using the contrast: [Intelligible speech Repeat Group Run 2 > Intelligible speech Repeat Group Run 1] > [Intelligible speech Imitation Group Run 2 > Intelligible speech Imitation Group Run 1] (in short: [Repeat Run 2 > 1]> [Imitation Run 2 > 1]). We also tested the contrasts [Imitate Run 2 > Run 1] and [Repeat Run 2 > Run 1] to get a general idea of areas that showed increases in BOLD-response across runs for both groups.

Significant activations in the whole-brain analysis were assessed with a Family Wise Error (FWE) correction for multiple comparisons (p < 0.05), in combination with a minimal cluster extent of 20 voxels. Finally, anatomical localization was performed using the Anatomy toolbox in SPM8 (Eickhoff et al., 2005) and verified with MRICRON using the ch2 template in MNI space (Rorden and Brett, 2000).

Results

Behavioral results

A repeated-measures ANOVA with Group (Repeat or Imitate) and Run (1 or 2) on the SRT value in dB per run (Figure 2) revealed no main effects, but did show an interaction between Run and Group [F(1, 30) = 6.21, p = 0.018, η2p = 0.17]. The absence of a main effect of Run indicated that both groups of listeners judged the sentence materials to be equally intelligible when played at similar SNRs. The presence of an interaction between Run and Group indicates that listeners in both groups were affected differently by the training. Two post-hoc analyses showed a marginally significant trend for the repeat group to require more favorable SNRs in the second run [t(14) = −2.033, p = 0.061], while no such trend was found for the imitate group [t(16) = 0.946, p = 0.358].

Figure 2

Neuroimaging results

Table 2 and Figure 2 show the network of areas involved in processing intelligible speech (the contrast Intelligibility included all stimuli that had been rated as 3 or 4 in the intelligibility judgment test) vs. the acoustically matched baseline. Intelligible speech was associated with higher BOLD values with peak coordinates in left and right anterior and posterior STS, left IFG, left insula, and SMA.

Table 2

Locationmm3tEquivalent ZMNI
[INTELLIGIBILITY > BASELINE]
Left STS96610.03Inf*−62 −32 4
Left STG6.465.97−58 −4 −8
Left MTG5.315.02−60 −16 −8
Left IFG (Triangularis)3066.706.17−50 18 24
Left IFG (Orbitalis)3746.616.09−46 28 −4
Left Insula6.005.60−36 22 −4
Right STG2505.945.5556 −20 −6
Right STG5.835.4660 2 −12
Left SMA715.325.03−2 10 60
[IMITATION RUN 2 > 1] > [REPEAT RUN 2 > 1]
Left STS7508.477.49−62 −32 2
Left MTG6.896.31−58 −20 −8
Left STS5.445.13−58 −8 −8
Left IFG (Triangularis)1646.425.94−52 18 26
Left SMA205.305.01−4 10 64
Left IFG (Orbitalis)265.284.99−46 26 −4
[IMITATE RUN 2 > RUN 1]
Left MTG613115.79Inf−58 −34 2
Left STG14.18Inf−58 −2 −8
Left anterior insula12.66Inf−30 24 −4
Left SMA144811.90Inf−4 20 46
R ACC6.636.1112 30 28
L ACC5.675.32−8 32 22
Right STG151211.42Inf60 2 −10
Left anterior insula10.89Inf32 24 −2
Right TP9.06Inf52 12 −18
White matter68010.49Inf−14 6 6
White matter3759.33Inf12 10 0
Right STG476.115.6946 −36 2
White matter455.625.28−4 −24 −18
White matter4.804.58−6 −20 −10
Left FFG235.515.20−38 −38 −20

Activation for peaks separated by at least 8 mm for the whole-brain analysis for the contrasts [Intelligibility > Baseline], [Imitation Run 2 > 1] > [Repeat Run 2 > 1], and [Run 1] > [Run 2].

Coordinates in MNI standard space. ACC, Anterior Cingulate Cortex; FFG, Fusiform Gyrus; IFG, Inferior Frontal Gyrus; MTG, Middle Temporal Gyrus; SMA, Supplementary Motor Area; STG, Superior Temporal Gyrus; STS, Superior Temporal Gyrus; TP, Temporal Pole.

*

Inf denotes Equivalent Z > 10.0.

We tested the contrasts [Imitation Run 2 > 1] > [Repeat Run 2 > 1] and [Repeat Run 2 > 1] > [Imitation Run 2 > 1] to probe for areas showing an effect of training in the second run for both groups (see Table 2 and Figure 3) in a whole-brain analysis. The areas showing an increased BOLD-response for the contrast [Imitation Run 2 > 1] > [Repeat Run 2 > 1] effect included left STS, an area in left IFG, and a cluster in left SMA. For the reverse contrast [Repeat Run 2 > 1] > [Imitation Run 2 > 1] no significant clusters were found. However, at a less stringent significance level of p < 0.001, we found two clusters in left and right Supramarginal Gyrus (SMG) at [−62 −38 36] (157 voxels) and [60 −42 22] (50 voxels). Finally we found no supra-threshold voxels for the contrast [Repeat Run 2 > Run 1], while for [Imitate Run 2 > Run 1] significant clusters were found in MTG bilaterally, left STG, SMA, bilateral Anterior Insula, Right Temporal Pole, and left Fusiform Gyrus.

Figure 3

Discussion

Main findings

We investigated the neural bases associated with increased robustness of speech comprehension after imitating an unfamiliar accent and after repeating the accented speech in one's own accent. The aim of the study was to establish the effect of short term-experience with imitating accented speech on the activation of cortical areas associated with speech perception and production when subsequently listening to the accent in background noise.

Based on previous studies on the role of motor experience, it was expected that motor experience with speaking in a novel unfamiliar accent action would result in a changed activation pattern brain areas associated with speaking and listening to speech, including left IFG, STS, and SMA. Previous studies on motor experience in general action processing (Calvo-Merino et al., 2005; Lahav et al., 2007) predicted that experience with performing a motor act (such as performing a dance move in Calvo-Merino et al.) increased activation in motor areas associated with performing that act while participants passively observed the act. However, a previous study on the effect of long-term speech production experience on speech sound perception (Wilson and Iacoboni, 2006) found that activation in premotor cortex was elevated for listening to foreign speech sounds.

Our results show that motor experience with accented speech leads to increased activation in speech perception and speech motor areas during subsequent comprehension of that accent. Areas in left STS, left IFG, and left SMA showed an interaction between run and group: they were more active during the second run for the imitation group compared to the difference between the first and second run for the repeat group. This effect was supported by a significant behavioral interaction between Run and Group.

Interestingly, these results are largely in line with previous studies that established the effect of perceptual experience with unfamiliar or distorted speech streams on neural activation during comprehension (Adank and Devlin, 2010; Eisner et al., 2010). Both studies monitored participants' neural activations while they perceptually adapted to distorted speech. Adank and Devlin presented participants with 64 sentences spoken by at a normal speed, before playing 64 sentences that had been artificially speeded up (time-compressed speech) to 45% of their original duration (Dupoux and Green, 1997). Participants performed a speeded sentence-verification task (i.e., deciding whether a sentence such as “Cars have four wheels” is true or false). When presented with the time-compressed sentences, performance on the sentence-verification task initially deteriorated sharply, but it quickly improved, signaling successful perceptual adaptation. Four cortical areas showed a pattern in the BOLD-activations in line with the behavioral results: one area in left ventral premotor cortex, two left lateralized areas in STG/STS, and one right-lateralized area in STG/STS. Eisner et al. studied the cortical changes in activity related to learning to perceptually adapt to noise-vocoded speech (Shannon et al., 1995). They included two types of distortion: one that was learnable and another was not Eisner et al. found that activity in left IFG correlated positively with individual variation in working memory scores, and that left IFG and left STS were sensitive to the extent to which the stimulus displayed learnable variation.

Nevertheless, a difference between the present study and Adank and Devlin's is that Adank and Devlin used an online learning design. They monitored neural responses as participants perceptually adapted. In the present study, we used an offline design: we did not monitor participants' neural responses during the training, to avoid problems with head motion during speaking. Consequently, it is possible that left IFG, STS, and SMA were even more active during the training phase for the imitation group. Adank and Devlin report that areas that showed an increase in their BOLD-response when participants were first confronted with the distortion (time-compressed sentences) in the speech signal also remained more active after participants had adapted. Thus, speech perception and speech production regions continued to be active when the listening conditions remained challenging.

However, others have suggested that other areas other than IFG, SMA, and STS are the focus of adaptation-related changes in speech perception. For instance, Rauschecker and Scott (2009) suggest hat the IPL provides an interface “where feed-forward signals from motor preparatory networks in the inferior frontal cortex and premotor cortex can be matched with feedback signals from sensory areas” (p. 722). Increased BOLD-activation was found—albeit at a lowered significance level—for the repeat group in the second run for the contrast [Repeat Run 2 > 1] > [Imitation Run 2 > 1]. However, no activation was found in IPL for the imitation group in the second run compared to the repeat group. It is possible that IPL was more active during the training phase for the imitation group and that the current design has not been able to record this activity as the scanner was turned off during training.

The question of which areas are active while participants are acquiring motor experience may be resolved in an online imitation design, for instance in an experiment in which participants are scanned while vocally imitating or repeating a novel speech stream. Such a design may allow for scrutiny of imitation-related activity in IPL, IFG, STS, and SMA.

Interaction between group and run

The absolute pattern of results for the interaction effect observed in the behavioral data as well as in the neuroimaging data was not completely in line with our predictions. Namely, while we hypothesized an interaction in participants' performance between Run × Group, we expected that interaction to be driven by a significant increase in performance in Run 2 for participants who had imitated in contrast to participants who repeated in their own accent. Instead, the interaction in the current data set is driven by a relative decrease in performance for participants in the Repeat as compared to the Imitate Group. Thus, while the overall and relative pattern of results is in line with our predictions, the pattern within the second alone warrants further discussion.

The first explanation of these results is in terms of the simulation theories discussed in the Introduction. Simulation models presume that imitative motor involvement helps perceivers anticipate other people's actions better by generating forward models (Grush, 2004). These models serves to anticipate others' actions as if they were our own, something which is proposed to be beneficial for ambiguous or noisy signals or actions (Wilson and Knoblich, 2005; Pickering and Garrod, 2007). Simulation models predict that imitation improves perception, and several studies have recently confirmed this prediction for manual actions (Locatelli et al., 2012) and for speech (Adank et al., 2010). Simulation models generally presume that imitation is predominantly covert (and that any overt imitative action is actively suppressed, cf. Baldissera et al., 2001), but it is unclear how the act of overt motor imitation can support perception.

Moreover, recent behavioral research on perceptual learning in speech indicates that listeners adapt to accented speech by updating their internal phoneme representations for the speech sounds in question (e.g., Norris et al., 2003; Evans and Iverson, 2004). Information from speech articulators may be used to inform feed forward models used in the simulation process. Two recent studies support the possibility of such a supporting mechanism. D'Ausilio et al. (2009) demonstrated that repetitive Transcranial Magnetic Stimulation (rTMS) to lip and tongue motor cortex shifted perception of speech toward speech sounds involving these articulators, while Ito et al. (2009) demonstrated that perception of vowel sounds is affected by feedback from facial muscles while speaking these vowels with an altered pronunciation. Information from muscles can therefore, be used to inform perception. This information could be used by the forward models to improve prediction of utterances in the unfamiliar accent. In contrast, in the repeat group no updated information was available from motor neurons in articulation muscles. In turn, this could have resulted in the trend toward poorer comprehension after training in the current experiment.

However, an alternative explanation of the interaction between group and run and the trend toward lower intelligibility judgments in repeat group could be that participants in the repeat group performed a less engaging task. It seems plausible that having to replicate the way in which a sentence is being said is a task that requires more attention and effort than a task in which participants had to pronounce what was being said in their own accent. This possible increased lower engagement in the Repeat group may have had the following two consequences.

First, participants in the Repeat group could have gotten more fatigued and bored than those in the Imitation group. Consequently, participants in the Repeat group may not have been able (or willing) to pay as much attention in the first run. There exists anecdotal and experimental support for the notion that it is harder to maintain attention in unchallenging, monotonous tasks than for cognitively demanding but interesting tasks (see Robertson and O'connell, 2010, for an overview). Generally, behavioral performance deteriorates for longer sustained tasks, and this deterioration is faster for unchallenging tasks (such as repeating compered to imitating). A recent meta-analysis of neuroimaging studies on vigilant attention (i.e., the cognitive function that enables prolonged engagement in intellectually unchallenging, uninteresting activities) identified a network of—mostly right-lateralized—brain regions subserving vigilant attention, including dorsomedial, mid- and ventrolateral prefrontal cortex, anterior insula, parietal areas (intraparietal sulcus, temporoparietal junction), and finally subcortical structures (cerebellar vermis, thalamus, putamen, and midbrain) (Langner and Eickhoff, 2013). Note that our network of areas displaying the interaction effect ([Imitation Run 2 > 1] > [Repeat Run 2 > 1]) was exclusively left-lateralized and does not appear to overlap with the network identified by Langner and Eickhoff. Furthermore, inspecting of the reverse contrast ([Repeat Run 2 > 1] > [Imitation Run 2 > 1]) at a less stringent significance level indicated involvement of left IPL, whereas Langner and Eickhoff's analysis showed no clusters in the left parietal lobe.

Second, participants in the Repeat group performed a task that may have not allowed them to engage with the speaker of the accent as much as participants in the Imitation group. This decreased engagement could then have made them less willing to put as much effort into understanding the accented sentences in the second run as in the first run. This notion is speculative, but is supported by an experiment on the effect of imitating a target's (such as a speaker) behavior on the perception of the social characteristics of that target. In a within-subject design, Adank et al. (2013) asked a group of participants to listen to sentences spoken by two speakers of a regional accent (Glaswegian) of British-English. Next, they repeated the sentences spoken by one of the Glaswegian speakers in their own accent, and subsequently participants imitated sentences spoken by the second Glaswegian speaker. The order of repeating and imitating was counterbalanced across participants. After each repeating or imitation session, participants completed a questionnaire probing the speakers' perceived power, competence, and social attractiveness. Imitating had a positive effect on the perceived social attractiveness of the speaker compared to repeating. The authors explained the positive results of imitating by stating that the act of imitating another's accent makes the speaker part of participants' social in-group in a way that repeating does not. Since people are more positively biased toward people in their in-group than those outside (Brewer, 1979), such in-group favoritism then made the speaker seem more subjectively pleasant. However, it is unclear whether a comparable effect of imitation on perception of the speaker would also occur for the constructed accent used in our study. Hearing accented speech automatically invokes attitudes associated with speakers of the accent (Giles, 1970; Bishop et al., 2005; Coupland and Bishop, 2007) and it seems unlikely that constructed accents are associated with specific language attitudes as is the case for existing accents.

The issue of differences in engagement with the task across groups issue may be addressed in future studies by explicitly matching training tasks for “level of interest.” One way to achieve this would be to create two different constructed accents that are matched for intelligibility and have participants repeat sentences for one of the accents and imitate for the other accent in a within-subject design. Subsequently, their intelligibility of both accents could be tested to verify that imitation had a positive effect on the intelligibility of the imitated constructed accent. However, we chose not to take this route in the present study, as we aimed to replicate the design in Adank et al. (2010) as closely as possible using an fMRI design. Yet, future studies could, for instance, pre-test a variety of tasks on how much they are perceived to be equally interesting or engaging by participants or use multiple constructed accents.

A final explanation for the pattern in the behavioral results could be that the behavioral task used in the experiment, perhaps in combination with the presence of continuous scanner noise, did not accurately reflect the extent to which participants' performance changed over the two runs. Our earlier behavioral study described in Adank et al. (2010) used the standard SRT procedure (which is the same as described for the hearing test in the Materials and Methods section), and it seems plausible that this test provides a more fine-grained measure of intelligibility processing than the test used in the scanner. Adank et al.'s SRT procedure involved overt vocal responses, which was not optimal given our fMRI design. Future studies should thus, consider using a more fine-grained behavioral task, possibly in combination with a sparse sampling design (Hall et al., 1999).

Conclusion

Previous studies (e.g., Davis and Johnsrude, 2003; Adank, 2012) have shown that processing of intelligible speech involves neural substrates associated with speech production and speech perception. This study adds to a growing body of literature showing that processing of intelligible speech involves sensorimotor processes. However, further research is required to establish in what manner overt imitation of speech contributes to these sensorimotor processes in challenging listening situations. Ideally, future studies involving imitation related-training should make a point of explicitly disentangling the possible contribution of factors such as participant (dis)engagement and listening environment. Only by addressing these issues can the contribution of imitative actions to speech perception be elucidated.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Statements

Acknowledgments

We wish to thank Pascal de Water and Paul Gaalman for technical assistance and Ivan Toni for useful comments on the design of the study. The study was supported by the Dutch Organization for Scientific Research (NWO) through a VENI grant 016094-053 to Shirley-Ann Rueschemeyer and a VICI grant 453–05–001 to Harold Bekkering.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    AdankP. (2012). The neural bases of difficult speech comprehension and speech production and their overlap: two Activation Likelihood Estimation (ALE) meta-analyses. Brain Lang. 122, 4254. 10.1016/j.bandl.2012.04.014

  • 2

    AdankP.DavisM.HagoortP. (2012). Neural dissociation in processing noise and accent in spoken language comprehension. Neuropsychologia50, 7784. 10.1016/j.neuropsychologia.2011.10.024

  • 3

    AdankP.DevlinJ. T. (2010). On-line plasticity in spoken sentence comprehension: adapting to time-compressed speech. Neuroimage49, 11241132. 10.1016/j.neuroimage.2009.07.032

  • 4

    AdankP.EvansB. G.Stuart-SmithJ.ScottS. K. (2009). Comprehension of familiar and unfamiliar native accents under adverse listening conditions. J. Exp. Psychol. Hum. Percept. Perform. 35, 520529. 10.1037/a0013552

  • 5

    AdankP.HagoortP.BekkeringH. (2010). Imitation improves language comprehension. Psychol. Sci. 21, 19031909. 10.1177/0956797610389192

  • 6

    AdankP.StewartA. J.ConnellL.WoodJ. (2013). Accent imitation positively affects accent attitudes. Front. Psychol. 4:280. 10.3389/fpsyg.2013.00280

  • 7

    AgliotiS. M.CesariP.RomaniM.UrgesiC. (2008). Action anticipation and motor resonance in elite basketball players. Nat. Neurosci. 11, 11091116. 10.1038/nn.2182

  • 8

    AlarioF. X.ChainayH.LehericyS.CohenL. (2006). The role of the supplementary motor area (SMA) in word production. Brain Res. 1076, 129143. 10.1016/j.brainres.2005.11.104

  • 9

    AshburnerJ.FristonK. J. (1997). Multimodal image coregistration and partitioning—a unified framework. Neuroimage6, 209217. 10.1006/nimg.1997.0290

  • 10

    BaldisseraF.CavallariP.CraigheroL.FadigaL. (2001). Modulation of spinal excitability during observation of hand actions in humans. Eur. J. Neurosci. 13, 190194. 10.1046/j.0953-816x.2000.01368.x

  • 11

    BishopH.CouplandN.GarrettP. (2005). Conceptual accent evaluation: thirty years of accent prejudice in the UK. Acta Linguist. Hafniensia. Int. J. Linguist. 37, 131154. 10.1080/03740463.2005.10416087

  • 12

    BlankS. C.ScottS. K.MurphyK.WarburtonE.WiseR. J. S. (2002). Speech production: wernicke, broca and beyond. Brain125, 18291838. 10.1093/brain/awf191

  • 13

    BlesserB. (1972). Speech perception under conditions of spectral transformation. I. Phonetic characteristics. J. Speech Hear. Res. 15, 541.

  • 14

    BoersmaP.WeeninkD. (2003). Praat: doing Phonetics by Computer. Available online at: http://www.praat.org

  • 15

    BrettM.AntonJ.-L.ValabregueR.PolineJ.-B. (2002). Region of interest analysis using an SPM toolbox. Neuroimage167, 479.

  • 16

    BrewerM. B. (1979). In-group bias in minimal intergroup situation: a cognitive-motivational analysis. Psychol. Bull. 86, 307324. 10.1037/0033-2909.86.2.307

  • 17

    CallanD. E.CallanA. M.GamezM.SatoM.KawatoM. (2010). Premotor cortex mediates perceptual performance. Neuroimage51, 844858. 10.1016/j.neuroimage.2010.02.027

  • 18

    Calvo-MerinoB.GlaserD. E.GrezesJ.PassinghamR. E.HaggardP. (2005). Action observation and acquired motor skills. Cereb. Cortex15, 12431249. 10.1093/cercor/bhi007

  • 19

    CouplandJ.BishopH. (2007). Ideologised values for British accents. J. Socioling. 11, 7493. 10.1111/j.1467-9841.2007.00311.x

  • 20

    CrinionJ. T.PriceC. J. (2005). Right anterior superior temporal activation predicts auditory sentence comprehension following aphasic stroke. Brain128, 28582871. 10.1093/brain/awh659

  • 21

    D'AusilioA.PülvermüllerF.SalmasP.BufalariI.BegliominiC.FadigaL. (2009). The motor somatotopy of speech perception. Curr. Biol. 19, 381385. 10.1016/j.cub.2009.01.017

  • 22

    DavisM. H.JohnsrudeI. S. (2003). Hierarchical processing in spoken language comprehension. J. Neurosci. 23, 34233431.

  • 23

    DupouxE.GreenK. (1997). Perceptual adjustment to highly compressed speech: effects of talker and rate changes. J. Exp. Psychol. Hum. Percept. Perform. 23, 914927. 10.1037/0096-1523.23.3.914

  • 24

    EickhoffS. B.StephanK. E.MohlbergH.GrefkesC.FinkG. R.AmuntsK.et al. (2005). A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. Neuroimage25, 13251335. 10.1016/j.neuroimage.2004.12.034

  • 25

    EisnerF.McGettiganC.FaulknerA.RosenS.ScottS. K. (2010). Inferior frontal gyrus activation predicts individual differences in perceptual learning of cochlear-implant simulations. J. Neurosci. 30, 71797186. 10.1523/JNEUROSCI.4040-09.2010

  • 26

    EvansB. G.IversonP. (2004). Vowel normalization for accent: an investigation of best exemplar locations in northern and southern British English sentences. J. Acoust. Soc. Am. 115, 352361. 10.1121/1.1635413

  • 27

    FridrikssonJ.MoserD.RyallsJ.BonillaL.RordenC.BaylisG. (2009). Modulation of frontal lobe speech areas associated with the production and perception of speech movements. J. Speech Lang. Hear. Res. 52, 812819. 10.1044/1092-4388(2008/06-0197)

  • 28

    FristonK. J.HolmesA. P.PolineJ. B.GrasbyP. J.WilliamsS. C.FrackowiakR. S.et al. (1995). Analysis of fMRI time-series revisited. Neuroimage2, 4553. 10.1006/nimg.1995.1007

  • 29

    GilesH. (1970). Evaluative reactions to accents. Educ. Rev. 22, 211227. 10.1080/0013191700220301

  • 30

    GraftonS. (2009). Embodied cognition and the simulation of action to understand others. Ann. N.Y. Acad. Sci. 1156, 97117. 10.1111/j.1749-6632.2009.04425.x

  • 31

    GrushR. (2004). The emulation theory of representation: motor control, imagery, and perception. Behav. Brain Sci. 27, 377435. 10.1017/S0140525X04000093

  • 32

    HallD. A.HaggardM. P.AkeroydM. A.PalmerA. R.SummerfieldA. Q.ElliotM. R.et al. (1999). “Sparse” temporal sampling in auditory fMRI. Hum. Brain Mapp. 7, 213223. 10.1002/(SICI)1097-0193(1999)7:3<213::AID-HBM5>3.0.CO;2-N

  • 33

    IPA. (1999). Handbook of the International Phonetic Association: A guide to the use of the International Phonetic Alphabet. Cambridge: Cambridge University Press.

  • 34

    ItoT.TiedeM.OstryD. J. (2009). Somatosensory function in speech perception. Proc. Natl. Acad. Sci. 106, 12451248. 10.1073/pnas.0810063106

  • 35

    KnoblichG.SebanzN. (2006). The social nature of percepton and action. Curr. Dir. Psychol. Sci. 15, 99104. 10.1111/j.0963-7214.2006.00415.x

  • 36

    LahavA.SaltzmanE.SchlaugG. (2007). Action representation of sound: audiomotor recognition network while listening to newly acquired actions. J. Neurosci. 27, 308314. 10.1523/JNEUROSCI.4822-06.2007

  • 37

    LangnerR.EickhoffS. B. (2013). Sustaining attention to simple tasks: a meta-analytic review of the neural mechanisms of vigilant attention. Psychol. Bull. 139, 870900. 10.1037/a0030694

  • 38

    LocatelliM.GattiR.TettamantiM. (2012). Training of manual actions improves language understanding of semantically related action sentences. Front. Psychol. 3:547. 10.3389/fpsyg.2012.00547

  • 39

    LondeiA.D'AusilioA.BassoD.SestieriC.GrattaC. D.RomaniG. L.et al. (2009). Sensory-motor brain network connectivity for speech comprehension. Hum. Brain Mapp. 31, 567580. 10.1002/hbm.20888

  • 40

    NarainC.ScottS. K.WiseR. J. S.RosenS.LeffA.IversenS. D.et al. (2003). Defining a left-lateralized response specific to intelligible speech using fMRI. Cereb. Cortex13, 13621368. 10.1093/cercor/bhg083

  • 41

    NorrisD.McQueenJ. M.CutlerA. (2003). Perceptual learning in speech. Cogn. Psychol. 47, 204238. 10.1016/S0010-0285(03)00006-9

  • 42

    ObleserJ.KotzS. A. (2010). Expectancy constraints in degraded speech modulate the language comprehension network. Cereb. Cortex20, 633640. 10.1093/cercor/bhp128

  • 43

    ObleserJ.MeyerL.FriedericiA. D. (2011). Dynamic assignment of neural resources in auditory comprehension of complex sentences. Neuroimage56, 23102320. 10.1016/j.neuroimage.2011.03.035

  • 44

    ObleserJ.WiseR. J. S.DresnerM. A.ScottS. K. (2007). Functional integration across brain regions improves speech perception under adverse listening conditions. J. Neurosci. 27, 22832289. 10.1523/JNEUROSCI.4663-06.2007

  • 45

    PeelleJ. E.TroianiV.WingfieldA.GrossmanM. (2010). Neural processing during older adults' comprehension of spoken sentences: age differences in resource allocation and connectivity. Cereb. Cortex20, 773782. 10.1093/cercor/bhp142

  • 46

    PickeringM. J.GarrodS. (2007). Do people use language production to make predictions during comprehension?Trends Cogn. Sci. 11, 105110. 10.1016/j.tics.2006.12.002

  • 47

    PlompR.MimpenA. M. (1979). Speech reception threshold for sentences as a function of age and noise level. J. Acoust. Soc. Am. 66, 13331342. 10.1121/1.383554

  • 48

    RauscheckerJ. P.ScottS. K. (2009). Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat. Neurosci. 12, 718724. 10.1038/nn.2331

  • 49

    RobertsonI. H.O'connellR. G. (2010). Vigilant attention, in Attention and Time, eds NobreA. C.CoullJ. T. (Oxford: Oxford University Press), 7988.

  • 50

    RoddJ. M.DavisM. H.JohnsrudeI. S. (2005). The neural mechanisms of speech comprehension: fMRI studies of semantic ambiguity. Cereb. Cortex15, 12611269. 10.1093/cercor/bhi009

  • 51

    RoddJ. M.LongeO. L.RandallB.TylerL. K. (2010). The functional organisation of the fronto-temporal language system: evidence from syntactic and semantic ambiguity. Neuropsychologia48, 13241335. 10.1016/j.neuropsychologia.2009.12.035

  • 52

    RordenC.BrettM. (2000). Stereotaxic display of brain lesions. Behav. Neurol. 12, 191200.

  • 53

    ScottS. K.BlankS. C.RosenS.WiseR. J. S. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain123, 24002406. 10.1093/brain/123.12.2400

  • 54

    ShannonR. V.ZengF. G.KamathV.WygonskiJ.EkelidM. (1995). Speech recognition with primarily temporal cues. Science270, 303304. 10.1126/science.270.5234.303

  • 55

    SkipperJ. I.NusbaumH. C.SmallS. L. (2006). Lending a helping hand to hearing: another motor theory of speech perception, in Action to Language Via the Mirror Neuron System, ed ArbibM. A. (Cambridge, MA: Cambridge University Press), 250285. 10.1017/CBO9780511541599.009

  • 56

    TremblayP.GraccoV. L. (2009). Contribution of the pre-SMA to the production of words and non-speech oral motor gestures, as revealed by repetitive transcranial magnetic stimulation (rTMS). Brain Res. 1268, 112124. 10.1016/j.brainres.2009.02.076

  • 57

    TremblayP.SmallS. L. (2011). On the context-dependent nature of the contribution of the ventral premotor cortex to speech perception. Neuroimage57, 15611571. 10.1016/j.neuroimage.2011.05.067

  • 58

    Van WijngaardenS. J.SteenekenH. J.HoutgastT. (2002). Quantifying the intelligibility of speech in noise for non-native talkers. J. Acoust. Soc. Am. 112, 30043013. 10.1121/1.1512289

  • 59

    VeneziaJ. H.SaberiK.ChubbC.HickokG. (2012). Response bias modulates the speech motor system during syllable discrimination. Front. Psychol. 3:157. 10.3389/fpsyg.2012.00157

  • 60

    WilsonM.KnoblichG. (2005). The case for motor involvement in perceiving conspecifics. Psychol. Bull. 131, 460473. 10.1037/0033-2909.131.3.460

  • 61

    WilsonS. M.IacoboniM. (2006). Neural responses to non-native phonemes varying in producibility: evidence for the sensorimotor nature of speech perception. Neuroimage33, 316325. 10.1016/j.neuroimage.2006.05.032

  • 62

    WongP. C. M.UppandaA. K.ParrishT. B.DharS. (2008). Cortical mechanisms of speech perception in noise. J. Speech Lang. Hear. Res. 51, 10261041. 10.1044/1092-4388(2008/075)

Appendix

Table A1

Nr.Standard DutchAccented Dutch
1De bal vloog over de schuttingDe baal vlog offer de schuuttieng
2Morgen wil ik maar één liter melkMoorgen wiel iek mar èn litter meelk
3Deze kerk moet gesloopt wordenDesse keerk mut geslopt woorden
4De spoortrein was al gauw kapotDe sportreen waas aal goew kaappoot
5De nieuwe fiets is gestolenDe niwwe fits ies gestollen
6Zijn manier van werken ligt mij nietZeen mannir vaan weerken liegt mee nit
7Het slot van de voordeur is kapotHet sloot vaan de vordur ies kaappoot
8Dat hotel heeft een slechte naamDaat hotteel heft ‘n sleechte nam
9De jongen werd stevig aangepaktDe joongen weerd steffig angepaakt
10Het natte hout sist in het vuurHet naatte hoet siest ien het vur
11Zijn fantasie kent geen grenzenZeen faantassih keent gèn greenzen
12De aardappels liggen in de schuurDe ardaappels liegen ien de schur
13Alle prijzen waren verhoogdAalle preezen warren verhogt
14Zijn leeftijd ligt boven de dertigZeen lèfteed liegt boffen de deertieg
15Het dak moet nodig hersteld wordenHet daak mut noddieg heersteeld woorden
16De kachel is nog steeds niet aanDe kaachel ies noog stèds nit an
17Van de viool is een snaar kapotVaan de vij-jol ies ‘n snar kaappoot
18De tuinman heeft het gras gemaaidDe tuunmaan heft het graas gemajt
19De appels aan de boom zijn rijpDe aappels an de bom zeen reep
20Voor het eerst was er nieuwe haringVor het erst waas eer niwwe harrieng
21Het loket bleef lang geslotenHet lokkeet blef laang geslotten
22Er werd een diepe kuil gegravenEer weerd 'n dippe koel gegraffen
23Zijn gezicht heeft een rode kleurZeen geziecht hèft 'n rodde klur
24Het begon vroeg donker te wordenHet beggoon vrug doonker te woorden
25Het gras was helemaal verdroogdHet graas waas hèllemal verdrogt
26Spoedig kwam er een einde aanSpuddieg kwaam eer ‘n eende an
27Ieder half uur komt hier een bus langsIdder haalf ur koomt hir ‘n buus laangs
28De bel van de voordeur is kapotDe beel vaan de vordur ies kaappoot
29De wind waait vandaag uit het westenDe wiend wajt vaandag uut het weesten
30De slang bewoog zich door het grasDe slaang bewog ziech dor het graas
31De kamer rook naar sigarenDe kammer rok nar siggarren
32De appel had een zure smaakDe aappel haad 'n zurre smak
33De trein kwam met een schok tot stilstandDe treen kwaam meet ‘n schook toot stielstaand
34De koeien werden juist gemolkenDe kujjen weerden juust gemoolken
35Het duurt niet langer dan een minuutHet durt nit laanger daan ‘n minnut
36De grijze lucht voorspelt regenDe greeze luucht vorspeelt règgen
37Hij kon de hamer nergens vindenHee koon de hammer neergens vienden
38Deze berg is nog niet beklommenDesse beerg ies noog nit bekloommen
39De bel van mijn fiets is kapotDe beel vaan meen fits ies kaappoot
40De auto heeft een lekke bandDe oetoh hèft ‘n leekke baand
41Het moeilijke werk bleef liggenHet muj-leekke weerk blef lieggen
42Het vliegtuig vertrekt over een uurHet vligtuug vertreekt offer ‘n ur
43De jongens vechten de hele dagDe joongens veechten de hèlle daag
44De schoenen moeten verzoold wordenDe schunnen mutten verzold woorden
45In de krant staat vandaag niet veel nieuwsIen de kraant stat vaandag nit vèl niws
46Door de neus ademen is beterDor de nus addemmen ies better
47Het kind was niet in staat te sprekenHet kiend waas nit ien stat te sprekken
48De witte zwaan dook onder waterDe wiette zwan dok oonder watter
49Hij nam het pak onder zijn armHee naam het paak oonder zeen aarm
50Gelukkig sloeg de motor niet afGeluukkieg slug de mottor nit aaf
51De leraar gaf hem een laag cijferDe lèrrar gaaf heem ‘n lag seeffer
52Het huis brandde tot de grond toe afHet huus braande toot de groond tuh aaf
53De foto is mooi ingelijstDe fotto ies moi iengeleest
54Mijn broer gaat elke dag fietsenMeen brur gat eelke daag fitsen
55Een kopje koffie zal goed smakenEen koopje kooffih zaal gud smakken
56De schrijver van dit boek is doodDe schreeffer vaan diet buk ies dot
57Zij heeft haar proefwerk slecht gemaaktZee heft har prufweerk sleecht gemakt
58De sigaar ligt in de asbakDe siggar liegt ien de aasbaak
59De appelboom stond in volle bloeiDe aappelbom stoond ien voolle bluj
60Er wordt in dit land geen rijst verbouwdEer woordt ien diet laand gèn reest verbuwd
61Hij kan er nu eenmaal niets aan doenHee kaan eer nuh ènmal nits an dun
62De kleren waren niet gewassenDe klerren warren nit gewaassen
63Het gedicht werd voorgelezenHet gediecht weerd vorgelèssen
64Haar gezicht was zwart van het vuilHar geziecht waas zwaart vaan het vuul
65De letters stonden op hun kopDe leetters stoonden oop huun koop
66De groene appels waren erg zuurDe grunne aappels warren eerg zur
67In het gebouw waren vier liftenIen het geboew warren vir lieften
68Lopen is gezonder dan fietsenLoppen ies gezoonder daan fitsen
69Het lawaai maakte hem wakkerHet lawwai makte heem waakker
70Mijn buurman heeft een auto gekochtMeen burmaan heft ‘n oetoh gekoocht
71Als het flink vriest kunnen we schaatsenAals het flienk frist kuunnen we schatsen
72De kast was een meter verschovenDe kaast waas 'n metter verschoffen
73Oude meubels zijn zeer in trekOede mubbels zeen zèr ien treek
74De portier ging met vakantieDe poortir gieng meet vaakkaantih
75De lantaarn gaf niet veel licht meerDe laantarn gaaf nit vèl liecht mer
76Door zijn snelheid vloog hij uit de bochtDoor zeen sneelheed vlog hee uut de boocht
77Het is hier nog steeds veel te koudHet ies hir noog steds vèl te koed
78De oude man was kaal gewordenDe oede maan waas kal gewoorden
79De bomen waren helemaal kaalDe bommen warren hèllemal llemal kal
80Rijden onder invloed is strafbaarReedden oonder ienvlud ies straafbar
81Onze bank geeft vijf procent renteOonze baank geft veef prosseent reente
82Het verslag in de krant is kortHet verslaag ien de kraant ies koort
83In de vijver zwemmen veel vissenIen de veeffer zweemmen vel viessen
84Honden mogen niet in het gebouwHoonden moggen nit ien het geboew
85Een flinke borrel zal mij goed doenEen flienke boorrel zaal mee gud dun
86Gisteren waaide het nog harderGiesteren wajde het noog haarder
87Het meisje stond lang te wachtenHet meesje stoond laang te waachten
88De volgende dag kwam hij ook nietDe voolgende daag kwaam hee ok nit
89Het geschreeuw is duidelijk hoorbaarHet geschrew ies duudeleek horbar
90Eindelijk kwam de trein op gangEendeleek kwaam de treen oop gaang
91De grote stad trok hem wel aanDe grotte staad trook heem weel an
92De bus is vandaag niet op tijdDe buus ies vaandag nit oop teed
93Onze dochter speelt goed blokfluitOonze doochter spèlt gud blookfluut
94Ook in de zomer is het hier koelOk ien de zommer ies het hir kul
95Zij moesten vier uur hard werkenZee musten vir ur haard weerken
96Niemand kan de Fransman verstaanNimmaand kaan de Fraansmaan verstan

Test sentences from the Speech Reception Threshold corpus (Plomp and Mimpen, 1979).

Only vowels bearing primary or secondary lexical stress were included in the conversion of the orthography.

Table A2

Nr.Standard DutchAccented Dutch
1Architecten hebben een beroepAarchitteecten heeben een berup
2Asperges zijn groentenAaspeerges zeen grunten
3Aardappels worden geschildArdaappels woorden geschield
4Bananen zijn fruitBannannen zen fruut
5Bijen vliegen rond op zoek naar voedselBeej-jen vliggen roond oop zuk nar vudsel
6Bevers bouwen dammen in de rivierBeffers buwwen damen ien de rivvir
7Beren hebben vier potenBerren heeben vir potten
8Bisschoppen dragen klerenBiesschopen draggen klerren
9Biefstukken worden verkocht door slagersBifstuukken woorden verkoocht dor slaggers
10Blikopeners kunnen in winkels gekocht wordenBliekoppeners kuunen ien wienkels gekoocht woorden
11Bromfietsen rijden op de wegBroomfitsen reeden oop de weeg
12Chirurgen moeten lang studerenChieruurgen mutten laang studderren
13Druiven zijn eetbaarDruuven zeen etbar
14Ezels dragen zware vrachtenEssels draggen zwarre vraachten
15Ezels kunnen koppig zijnEssels kuunen kooppieg zeen
16Forellen hebben schubbenForrelen heeben schuubben
17Ganzen kunnen ver vliegenGaanzen kuunen veer vliggen
18Haaien hebben sterke tandenHajjen heeben steerke taanden
19Heggenscharen worden in de tuin gebruiktHeegenscharren woorden ien de toen gebroekt
20Kapiteins voeren het bevel op schepenKaapitteins vurren het bevveel oop scheppen
21Kasten zijn van houtKaasten zeen vaan hoet
22Kroketten zijn gefrituurdKrokkeetten zeen gefritturd
23Lammetjes komen van schapenLametjes kommen vaan schappen
24Lepels worden gebruikt voor het eten van soepLeppels woorden gebroekt vor het etten vaan sup
25Leeuwen hebben manenLewwen heeben mannen
26Luipaarden hebben een vachtLuuparden heeben ‘n vaacht
27Makrelen ademen door kieuwenMaakrellen addeemen dor kiwwen
28Messen worden gebruikt als keukengereiMeessen woorden gebroekt aals kukkengeree
29Monniken wonen in een kloosterMoonnieken wonnen ien ‘n kloster
30Nachtegalen hebben verenNaachtegallen heeben verren
31Ooms zijn deel van de familieOms zeen del vaan de fammillih
32Otters kunnen goed zwemmenOoters kuunen gut zweemen
33Pinguïns eten veel visPiengiens etten vel vies
34Presidenten werken in de politiekPressideenten weerken ien de pollittik
35Ratelslangen kruipen op hun buikRattelslaangen kroepen oop huun boek
36Roodborstjes hebben een snavelRodboorstjes heeben een snaffel
37Schuurtjes worden gebruikt voor opslagSchurtjes woorden gebroekt vor oopslaag
38Slagers hebben winkelsSlaggers heeben wienkels
39Sloffen worden gemaakt in een fabriekSloofen woorden gemakt ien ‘n fabbrik
40Tantes zijn altijd vrouwenTaantes zeen aalteed vroewen
41Tijgers hebben een staartTeegers heeben ‘n start
42Tomaten groeien aan plantenTommatten grujjen an plaanten
43Vaders zijn oudersVadders zeen oeders
44Vlinders hebben voelsprietenVlienders heeben vulspritten
45Wandelschoenen zijn gefabriceerde goederenWaandelschunnen zeen gefaabrisserde gudderen
46Wijflessen hebben een kurkWeenfleessen heeben ‘n kuurk
47Wetenschappers moeten lang studerenWettenschaapers mutten laang studderren
48Wortels groeien in een moestuinWoortels grujjen ien ‘n mustuun

Training sentences.

Only vowels bearing primary or secondary lexical stress were included in the conversion of the orthography.

Table A3

Nr.Standard DutchKey words
1Het regent al de hele dagregent al hele dag
2De schaatsen zijn in het vet gezetschaatsen zijn vet gezet
3In juni zijn de dagen het langstjuni zijn dagen langst
4De bakkers bezorgen vandaag nietbakkers bezorgen vandaag niet
5Hij rookte zijn sigaret oprookte zijn sigaret op
6Het was heel stil in de duinenwas heel stil duinen
7Door zijn haast maakte hij veel foutenDoor haast veel fouten
8De kat likt het schoteltje leegkat likt schoteltje leeg
9De hond blafte de hele nachthond blafte hele nacht
10Het tempo lag voor hem veel te hoogtempo hem veel hoog
11Lawaai maakt je op den duur doofLawaai maakt duur doof
12Moeizaam klom de man naar bovenMoeizaam klom man boven
13Hij probeerde het nog een keerHij probeerde nog keer
14Toch lijkt me dat een goed voorstelToch lijkt goed voorstel
15Dat was voor hem een bittere pilDat hem bittere pil
16De nieuwe zaak was pas geopendnieuwe zaak pas geopend
17De rivier trad buiten haar oeversrivier trad buiten oevers
18De biefstuk was vandaag erg malsbiefstuk vandaag erg mals
19Op het gras mag men niet lopengras mag niet lopen
20De trein vertrekt over twee uurtrein vertrekt twee uur
21De kat van de buren is wegkat van buren weg
22De wagen reed snel de berg afwagen reed snel berg
23Iedereen genoot van het uitzichtIedereen genoot van uitzicht
24Steile trappen zijn gevaarlijkSteile trappen zijn gevaarlijk
25De zon gaat in het westen onderzon gaat westen onder
26De zak zat vol oude rommelzak vol oude rommel
27Zij werd misselijk van het rijdenZij werd misselijk rijden
28Het licht in de gang brandt nog steedslicht gang brandt steeds
29In de kerk werd mooi orgel gespeeldkerk mooi orgel gespeeld
30De jongens gingen er gauw vandoorjongens gingen gauw vandoor

Hearing test sentences, spoken in Standard Dutch.

Summary

Keywords

imitation, fMRI, speech, accent, sensorimotor

Citation

Adank P, Rueschemeyer S-A and Bekkering H (2013) The role of accent imitation in sensorimotor integration during processing of intelligible speech. Front. Hum. Neurosci. 7:634. doi: 10.3389/fnhum.2013.00634

Received

20 April 2013

Accepted

12 September 2013

Published

04 October 2013

Volume

7 - 2013

Edited by

Hans-Jochen Heinze, University of Magdeburg, Germany

Reviewed by

Shanqing Cai, Boston University, USA; Marc Sato, CNRS and Grenoble University, France

Copyright

*Correspondence: Patti Adank, Department of Speech, Hearing and Phonetic Sciences, Division of Psychology and Language Sciences, University College London, Room 323, Chandler House, 2 Wakefield St., London, WC1N 1PF, UK e-mail:

This article was submitted to the journal Frontiers in Human Neuroscience.

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics