ORIGINAL RESEARCH article

Front. Psychol., 17 January 2011

Sec. Auditory Cognitive Neuroscience

Volume 1 - 2010 | https://doi.org/10.3389/fpsyg.2010.00241

Stimulus Complexity and Categorical Effects in Human Auditory Cortex: An Activation Likelihood Estimation Meta-Analysis

  • FS

    Fabienne Samson 1,2*

  • TA

    Thomas A. Zeffiro 3

  • AT

    Alain Toussaint 1,2

  • PB

    Pascal Belin 4,5

  • 1. Centre d'Excellence en Troubles Envahissants du Développement de l'Université de Montréal Montréal, QC, Canada

  • 2. Centre de Recherche Fernand-Seguin, Department of Psychiatry, Université de Montréal Montréal, QC, Canada

  • 3. Neural Systems Group, Massachusetts General Hospital Boston, MA, USA

  • 4. Institute of Neuroscience and Psychology, University of Glasgow Glasgow, Scotland, UK

  • 5. International Laboratories for Brain, Music and Sound, McGill University and Université de Montréal Montréal, QC, Canada

Abstract

Investigations of the functional organization of human auditory cortex typically examine responses to different sound categories. An alternative approach is to characterize sounds with respect to their amount of variation in the time and frequency domains (i.e., spectral and temporal complexity). Although the vast majority of published studies examine contrasts between discrete sound categories, an alternative complexity-based taxonomy can be evaluated through meta-analysis. In a quantitative meta-analysis of 58 auditory neuroimaging studies, we examined the evidence supporting current models of functional specialization for auditory processing using grouping criteria based on either categories or spectro-temporal complexity. Consistent with current models, analyses based on typical sound categories revealed hierarchical auditory organization and left-lateralized responses to speech sounds, with high speech sensitivity in the left anterior superior temporal cortex. Classification of contrasts based on spectro-temporal complexity, on the other hand, revealed a striking within-hemisphere dissociation in which caudo-lateral temporal regions in auditory cortex showed greater sensitivity to spectral changes, while anterior superior temporal cortical areas were more sensitive to temporal variation, consistent with recent findings in animal models. The meta-analysis thus suggests that spectro-temporal acoustic complexity represents a useful alternative taxonomy to investigate the functional organization of human auditory cortex.

Introduction

Current accounts of the functional organization of auditory cortex, mostly based on response specificity to different sound categories, describe an organizational structure that is both hierarchical and hemispherically specialized (Rauschecker, 1998; Zatorre et al., 2002; Hackett, 2008; Rauschecker and Scott, 2009; Woods and Alain, 2009; Recanzone and Cohen, 2010).

Characterizing responses to stimuli from typical auditory categories such as music, voices, animal, or environmental sounds have provided important information about the cortical specialization for auditory processing. However, this classification may not fully account for the range of stimulus variability encountered across neuroimaging studies, as most stimuli do not fit neatly into one auditory category. For instance, an amplitude modulated tone can vary in ways that cannot be adequately characterized using typical categories. However, its characteristics can easily be described in terms of variations in time (temporal dimension) and frequency (spectral dimension), suggesting an alternative approach to stimulus classification. Accordingly, any auditory stimulus can be described with respect to its sound complexity characteristics specified with respect to changes in time and frequency. This approach represents a comprehensive characterization of sounds that is not limited to specific categories. Therefore, complexity might represent an alternative organizing principle along which to represent auditory cortical response specialization. In this conceptualization, a single frequency sinusoidal wave (pure tone), constant over time, can be classified as simple, and a sound containing multiple components can be classified as complex with respect to the frequency domain. Examples of sounds with high spectral complexity are musical notes or sustained vowels. Similarly, a sound with acoustical structure varying over time can be classified as complex with respect to the time domain. Examples of stimuli with high temporal complexity are frequency or amplitude modulated sounds or sound sequences. Natural sounds can be complex with regards to both their frequency composition and temporal variation. Phonemes, the basic units of speech, contain multiple frequency components, called formants, which may be combined over time to produce syllables and words. Similarly, musical sequences are composed of complex changes in fundamental frequency and harmonic structure that unfold over time. Additionally, speech processing is mainly dependent on temporal information (Shannon et al., 1995), while spectral composition is most relevant for music perception (Warrier and Zatorre, 2002). Hence, acoustic complexity is not independent of sound categories and the two classification methods explored here should not be considered as mutually exclusive.

As previously proposed, an auditory stimulus can be categorized in more than one way; either based on a priori knowledge about the characterizing features of the sound source or on the basis of a sound's acoustic pattern in the frequency and time domain (Griffiths and Warren, 2004). Additionally, some studies suggest that auditory cortex activation to sounds of a given category could reflect a specialized response to the acoustic components characterizing sounds within this category (Lewis et al., 2005, 2009). This suggests a certain level of interaction between the cortical processes involved in the analysis of acoustic features and those showing sensitivity to sound categories. However, recently Leaver and Rauschecker (2010) demonstrated categorical effects of speech and music stimuli even when controlling for changes in spectral and temporal dimensions. The two classification approaches are therefore not mutually exclusive and both methods seem relevant and can complement each other in revealing different aspects of cortical auditory specialization. In vision, cortical representation of stimulus complexity has been described with simple (first-order) information being analyzed within primary visual cortex (V1) and complex (second-order) information processing involving both primary and non-primary visual cortex (V2/V3; Chubb and Sperling, 1988; Larsson et al., 2006). Given that parallels have often been drawn between visual and auditory cortical functional organization (Rauschecker and Tian, 2000), we were interested in examining how well characterization of sounds by their acoustic complexity might reflects new insights into regional functional specialization.

Given that auditory neuroimaging studies exhibit a high degree of stimulus and task heterogeneity, their individual cortical activity patterns are not easily integrated to obtain an unambiguous picture of typical human auditory cortical organization. Neuroimaging meta-analysis offers a potential solution to this problem as it estimates the consistency of regional brain activity across similar stimuli and tasks, providing a quantitative summary of the state of research in a specific cognitive domain (Fox et al., 1998), estimating the replicability of effects across different scanners, tasks, stimuli, and research groups. By revealing consistently activated voxels across a set of experiments, meta-analysis can characterize the cortical response specificity associated with a particular type of task or stimulus (Wager et al., 2009). Activation Likelihood Estimation (ALE) is a voxel-wise meta-analysis method that provides a quantitative summary of task-related activity consistency across neuroimaging studies (Turkeltaub et al., 2002).

In the current study, we use quantitative ALE meta-analysis to examine the spatial consistency of human auditory processing, classified using either conventional sound categories or acoustic complexity. Given the focus of our study on stimulus complexity effects, we excluded studies of spatial auditory processes including localization, and inter-aural delay, as well as those including complex tasks.

First, we classified sounds using typical auditory categories to examine the evidence supporting hierarchically and hemispherically lateralized functional organization for auditory cortical processing. Hierarchical auditory processing has been described as sensitivity to stimulus complexity increasing from primary to non-primary auditory cortex, with simpler perceptual features represented at primary levels (Wessinger et al., 2001; Hall et al., 2002; Scott and Johnsrude, 2003). Relative hemispheric specialization is reflected by predominantly left-hemisphere processing for speech sounds and stronger right-hemisphere responses to music (for a review see Zatorre et al., 2002). We used typical sound categories, such as pure tones, noise, music, and vocal sounds, to classify auditory material to see if simple sound processing is associated with activity in primary auditory cortex while complex sound processing is associated with activity including both primary and non-primary auditory cortex. We were also interested in examining whether there was meta-analytic evidence for distinctive patterns of hemispheric specialization for music and vocal sounds.

Next, we more closely examined vocal stimuli and a particular subcategory of vocal sounds: intelligible speech. Vocalizations constitute an ecologically central sound category that includes all sounds with a vocal quality irrespective of phonetic or lexical content. Examples include speech in various languages, non-speech affective vocalizations (e.g., laughter), and laboratory-engineered sounds, such as time-reversed speech, that exhibit distinctly vocal qualities. Vocal sounds include, but are not limited to, intelligible speech. Based on previous findings, we expected to observe bilateral superior temporal gyrus (STG) and the superior temporal sulcus (STS) activity related to vocal sounds (Belin et al., 2000, 2002; Kriegstein and Giraud, 2004), and anterior STG and STS activity on the left related to speech intelligibility (Benson et al., 2006; Uppenkamp et al., 2006).

Finally, we examined whether acoustic complexity, estimated from variations in time (temporal) and frequency (spectral) dimensions, represents a relevant organizing principle for functional response specificity in human auditory cortex. In terms of spectral composition, stimuli can have single or multiple frequency components. In the temporal dimension, stimuli can be characterized as unchanging or, for those containing temporal changes, having either regular or irregular changes. Using this classification, we characterized the cortical response related to each level of acoustic complexity. Then, by comparing the “multiple” to the “single” categories, independent of the temporal changes, and the “changing” to the “unchanging” categories, independent of the spectral composition, we isolated the cortical activity related to variations in the frequency and time dimensions, respectively.

Materials and Methods

Inclusion of studies

A preliminary list of articles was identified using several Medline database searches including both articles published prior to March 2010 [keywords: positron emission tomography (PET), functional magnetic resonance imaging (fMRI), auditory, sound, hear?, speech, and music] and lists of citations within those articles. Studies were included if they fulfilled specific inclusion criteria: (1) the study was published in a peer-reviewed journal; (2) the study involved a group of healthy typical adult participants with no history of hearing, psychiatric, neurological, or other medical disorders, (3) the subjects were not trained musicians; (4) the auditory stimuli were delivered binaurally, with no inter-aural delay because of our focus on non-spatial auditory processing; (5) the task-related activity coordinates were reported in standardized anatomical space; (6) the study used whole-brain imaging and voxel-wise analysis. As our main goal was to determine the spatial distribution within auditory cortical regions, the few studies using incomplete brain coverage, but that included the temporal cortex were not excluded (Binder et al., 1996, 2000; Belin et al., 1999; Celsis et al., 1999; Hugdahl et al., 2003; Stevens, 2004; Schönwiesner et al., 2005; Zaehle et al., 2007). Additionally, some studies specifically included subcortical structures (Griffiths et al., 1998; Hwang et al., 2007; Mutschler et al., 2010); and (7) the study had to include passive listening or a simple response task, such as a button press at the end of each sound to assess the participants’ attentive state, task characteristics that tended to minimize the inclusion of activity related to top-down processes or task difficulty (Dehaene-Lambertz et al., 2005; Dufor et al., 2007; Sabri et al., 2008).

Of over 7000 articles retrieved, 58 (19 PET and 39 fMRI) satisfied all inclusion criteria and were included in the analysis (Table 1). Several studies reported activity from multiple task and control conditions. For our analysis, only conditions incorporating either no overt task or a simple task used to maintain attention were considered. To maintain consistency among the control conditions, only task contrasts with a low-level baseline (silence, tone, or noise) were included. For some studies, more than one contrast satisfied our criteria and all were included in the analysis. This procedure was employed to maximize the sensitivity of the analysis, but could potentially bias the results toward samples for which more than one contrast was included.

Table 1

ReferenceTable/figureCategoryFrequency compositionChanges over timeStimuli of interestBaselineFoci
Belin et al. (1998)Table 1, p. 537MultipleIrregularRapid formant transitionSilence3
MultipleIrregularExtended formant transitionSilence2
Belin et al. (1999)Text, p. 422–423MultipleUnchangingHarmonic complex soundSilence4
Benson et al. (2001)Table 1, p. 372MusicNote/chord/chord progressionRest27
Table 2, p. 373VocalMultipleIrregularVowel/consonant–vowels/syllablesRest24
Binder et al. (1996)Table 1, p. 1244VocalMultipleIrregularWordsTones sequence1
Binder et al. (2000)Table 2, p. 518MultipleRegularSequence of tonesNoise11
VocalMultipleIrregularWordsSequence of tones8
VocalMultipleIrregularPseudowordsSequence of tones8
VocalMultipleIrregularReversed wordsSequence of tones7
Binder et al. (2008)AppendixVocalMultipleIrregularWordsRest12
VocalMultipleIrregularWordsTones2
Blood et al. (1999)Table 2, p. 384MusicMelodiesNoise3
MusicMelodiesNoise3
Brown et al. (2004)Table 1, p. 2035MusicMultipleIrregularMusicRest21
Burton et al. (2000)Table 2, p. 682VocalMultipleIrregularWordsTone3
VocalMultipleIrregularWordsTone6
Burton and Small (2006)Table 2, p. 647VocalMultipleIrregularWordsTone4
Celsis et al. (1999)Table 1, p. 138ToneSingleUnchangingToneRest1
ToneSingleUnchangingToneRest1
VocalMultipleIrregularSyllableRest3
ToneSingleUnchangingToneRest3
MultipleUnchangingSquare wave toneRest3
VocalMultipleIrregularSyllableRest3
Chen et al. (2008)Table 2, p. 2849MusicMultipleIrregularMusical rhythmsSilence9
Dalla Barba et al. (1998)Table 1, p. 552VocalMultipleIrregularWordsRest5
De Nil et al. (2008)Table 2, p. 119VocalMultipleIrregularWordsSilence1
Démonet et al. (1992)Table 4, p. 1758VocalMultipleIrregularPhonemes (three or four syllables non-words)Tones7
VocalMultipleIrregularWordsTones14
Engelien et al. (2006)Table 1, p. 603MultipleIrregularMeaningful non-verbal soundsRest8
MultipleIrregularMeaningless non-verbal soundsRest14
Fernandez-Espejo et al. (2008)Table 2, p. 887VocalMultipleIrregularNarrative sentencesSilence2
Gaab et al. (2007a)Table 1, p. 710VocalMultipleIrregularWordsSilence7
VocalMultipleIrregularWordsSilence9
Gaab et al. (2007b)Table 1, p. 727VocalMultipleIrregularWordsSilence5
MultipleIrregularRecorded scanner noiseSilence2
Giraud et al. (2004)Table 1, p 250VocalSpeech (sentences)Temporally matched complex noise5
Griffiths et al. (1998)Table 1, p. 424MusicMelodiesSequence of tones4
Hall et al. (2000)Table 1, p. 114SingleRegularAmplitude and frequency modulated toneStatic tone2
Hall et al. (2002)Table 2, p. 144MultipleUnchangingHarmonic complex toneSingle tone5
SingleRegularFrequency modulated toneStatic tone10
Hart et al. (2003)Text, p. 778SingleRegularAmplitude and frequency modulated toneUnmodulated tone2
Hart et al. (2004)Table 1, p. 182ToneSingleUnchangingToneSilence4
SingleRegularFrequency modulated toneSilence6
Hertrich et al. (2010)Table S1VocalMultipleIrregularSyllablesSilence10
SingleRegularSingle formants sweepSilence8
Hugdahl et al. (2003)Table 3, p. 43VocalMultipleIrregularVowelSilence4
Table 4, p. 43VocalMultipleIrregularPseudowordsSilence3
Table 5, P. 43VocalMultipleIrregularWordsSilence2
Hwang et al. (2007)Table 2, p. 289VocalMultipleIrregularSentencesQuiet7
Jäncke et al. (1998)Table 1, p. 878VocalMultipleIrregularSyllablesSilence9
VocalMultipleIrregularSyllablesSilence9
VocalMultipleIrregularSyllablesSilence9
Table 2, p. 878MultipleRegularTone sequenceSilence4
MultipleRegularTone sequenceSilence6
MultipleRegularTone sequenceSilence6
Lillywhite et al. (2010)Table 1, p. 875VocalMultipleIrregularStory listeningWhite noise2
Müller et al. (1999)Table 3, p. 24MultipleRegularSequence of tonesRest6
VocalMultipleIrregularSentencesRest5
Mummery et al. (1999)Table 1, p. 452VocalSpeech (words)Signal correlated noise7
Mutschler et al. (2010)Supp. MaterialMusicMultipleIrregularMelodiesRest94
O'Leary et al. (1996)Table 2, p. 27VocalMultipleIrregularWordsTone2
MultipleIrregularEnvironmental soundsTone2
Obleser et al. (2006)Table 2, p. 566VocalMultipleIrregularVowelBand-pass noise4
Paulesu et al. (1995)Table 2, p. 667VocalMultipleIrregularWordsTones5
Peretz et al. (2009)Table 1, p. 263MusicMelodiesRandom tone sequence8
Petersen et al. (1988)Table 2, p. 585VocalMultipleIrregularWordsSilence (fixation point)6
Price et al. (1992)Figure 1, p. 180VocalMultipleIrregularNounsRest6
Reynolds Losin et al. (2009)Table 1, p. 374VocalMultipleIrregularStory listeningRest3
Rimol et al. (2005)Table 1, p. 1063VocalMultipleIrregularConsonantRest8
VocalMultipleIrregularSyllablesRest6
NoiseMultipleUnchangingNoise (white, brown, pink)Rest12
NoiseMultipleUnchangingNoise (white, brown, pink)Rest8
Table 2, p. 1064VocalMultipleIrregularConsonantNoise1
VocalMultipleIrregularSyllableNoise3
Schwarzbauer et al. (2006)Table 1, p. 779VocalSentencesSignal correlated noise12
VocalSentencesSignal correlated noise6
Schönwiesner et al. (2005)Text, p. 1523–1525SingleRegularParametric increase in temporal rate changeSilence2
MultipleUnchangingParametric increase in spectral componentsSilence2
Specht and Reul (2003)Table 1, p. 1950ToneSingleUnchangingTonesRest5
MultipleIrregularSounds of animal and instrumentsRest9
VocalMultipleIrregularWordsRest18
Table 2, p. 1951MultipleIrregularSounds of animal and instrumentsTones7
VocalMultipleIrregularWordsTones12
Steinbrink et al. (2009)Table 1, p. 2406VocalMultipleIrregularIncrease rate of syllable presentationSilence7
MultipleRegularIncrease rate of click sequenceSilence8
Stevens (2004)Table 1, p. 166VocalMultipleIrregularVoiceTone13
VocalMultipleIrregularWordsTone5
Thivard et al. (2000)Table 1, p. 2971MultipleUnchangingStatic complex soundSilence6
MultipleIrregularComplex sound modulated (motion)Silence6
SingleIrregularComplex sound modulated (motion)Static complex sound4
Vouloumanos et al. (2001)Table 1, p. 998VocalMultipleIrregularWordsTones8
MultipleIrregularNon-speech complex sinewave soundsTones5
Warren and Griffiths (2003)Table 1, p. 5802MusicChanging pitch sequencesFixed pitch sequences4
Warrier et al. (2009)Table 2, p. 65SingleRegularParametric increase in temporal rate changeSilence6
MultipleUnchangingParametric increase in spectral componentsSilence6
Wise et al. (1991)Table 2, p. 1808VocalMultipleIrregularNounsRest6
Zaehle et al. (2007)Table 1, p. 1201VocalMultipleIrregularSentencesSilence2
VocalMultipleIrregularSentencesSilence2
ToneSingleUnchangingTonesSilence2
ToneSingleUnchangingTonesSilence2
VocalMultipleIrregularSentencesTones2
VocalMultipleIrregularSentencesTones2
Table 2, p. 1201VocalMultipleIrregularSentencesSilence2
VocalMultipleIrregularSentencesSilence2
ToneSingleUnchangingTonesSilence2
ToneSingleUnchangingTonesSilence2
VocalMultipleIrregularSentencesTones2
VocalMultipleIrregularSentencesTones2
Zatorre et al. (1992)Table 2, p. 847NoiseMultipleUnchangingWhite noise burstsSilence8
VocalMultipleIrregularSyllablesNoise6
Zatorre et al. (1994)Table 2, p. 1911MusicMelodiesNoise2
Zatorre et al. (1996)Table 6, p. 26VocalMultipleIrregularSyllablesNoise6
Zatorre et al. (2002)Table 1, p. 907NoiseMultipleIrregularNoise (reversed environmental sounds)Silence3
Zatorre and Belin (2001)Table 1, p. 948SingleRegularTone sequence with increasing rate of presentationSilence2
MultipleUnchangingTone sequence with increasing number of spectral elementsSilence3
Zevin and McCandliss (2005)Table 1VocalMultipleIrregularSyllablesSilence8

Neuroimaging studies included in the meta-analysis.

Contrast classification procedure

One hundred seventeen contrasts, including 768 foci, met the inclusion criteria. These contrasts were classified first by typical sound categories and then according to their variation along either the frequency or time dimension (Table 1).

For the first method, each contrast was classified with respect to one of the typical sound categories: simple sounds or pure tones (9 contrasts, 22 foci), noise (4 contrasts, 31 foci), music (10 contrasts, 175 foci), and vocal sounds (62 contrasts, 370 foci). The pure tones category included only contrasts of single tones vs. silence; the noise category included white, pink, and brown noise (Rimol et al., 2005), noise bursts (Zatorre et al., 1992), and the combination of multiple reversed environmental sounds (Zatorre et al., 2002). Melodies, notes, chords, and chord progressions were classified as music. Finally, all sounds with a vocal quality (syllables, words, voices, reversed words, or pseudowords) were included in the vocal sounds category. Ideally, we would have included other commonly used sound categories such as animal or environmental sounds; however the number of contrasts falling into these categories was not sufficient for quantitative meta-analysis, with only one contrast presenting environmental sounds and only two falling into the animal sound category. The remaining contrasts (30/117), including modulated tones, frequency sweep, harmonic tones, or recorded noise, were not included in this analysis because they did not neatly fit into one sound category.

For the second method, we classified the stimuli with respect to their acoustic features. Two levels of complexity were defined using the frequency dimension (single and multiple frequency components) and three levels in the time domain (unchanging, regular periodic change, or irregular change). Therefore, task contrasts were classified in one of six complexity levels depending on their frequency- and time-related acoustic features (Table 1; Figure 5A): (1) “single, unchanging” (single tone; 9 contrasts, 22 foci), (2)“single, regular change” (frequency or amplitude modulated tone, single formant frequency sweep, parametric variation of modulation rate or rate of presentation; 8 contrasts, 38 foci), (3) “single, irregular change” (1 contrast, 4 foci), (4) “multiple, unchanging” (harmonic tone, square wave tone, vowel, noise, or parametrically increasing spectral component numbers; 10 contrasts, 57 foci), (5) “multiple, regular change” (tone sequences and increasing click rate sequences; 6 contrasts, 41 foci), or (6) “multiple, irregular change” (vocal sounds, music, or environmental sounds; 70 contrasts, 517 foci). Each task contrast was classified using the stimulus description provided in each study. Contrasts resulting from covariate effects of a parameter of interest were classified according to parameter complexity. For instance, effects related to parametric increases in temporal modulation rate were assigned to the “single, regular change” complexity level (Schönwiesner et al., 2005). Ambiguous contrasts were excluded from analysis. For example, we did not classify contrasts that used comparison stimuli that had acoustic complexity comparable to the stimuli of interest (Zatorre et al., 1994; Griffiths et al., 1998; Blood et al., 1999; Mummery et al., 1999; Warren and Griffiths, 2003; Giraud et al., 2004; Schwarzbauer et al., 2006; Peretz et al., 2009) nor those using stimuli that could be assigned to more than one complexity level, such as notes, chords, or chord progressions (i.e., stimuli including note/chord/chord progression; Benson et al., 2001).

ALE meta-analysis

After the task-related activity maxima were classified, ALE maps (Turkeltaub et al., 2002) were computed using GingerALE 1.1 (Laird et al., 2005). Coordinates reported in MNI space were converted to Talairach space using the Lancaster transform icbm2tal (Lancaster et al., 2007). ALE models uncertainty in localization of each activation focus as a Gaussian probability distribution, yielding a statistical map in which each voxel value represents an estimate of the likelihood of activity at that location, utilizing a fixed effects model for which inferences should be limited to the studies under examination. Critical thresholds for the ALE maps were determined using a Monte Carlo style permutation analysis of sets of randomly distributed foci. A FWHM of 10 mm was selected for the estimated Gaussian probability distributions. Critical thresholds were determined using 5000 permutations, corrected for multiple comparisons (p < 0.01 false discovery rate, FDR; Laird et al., 2005) with a cluster extent of greater than 250 mm3. In order to present results in the format most commonly used in the current literature, the ALE coordinate results were transformed into MNI standard space using the Lancaster transform (Lancaster et al., 2007), while ALE maps were transformed by applying spatial normalization parameters obtained from mapping from Talairach to MNI space.

Analysis using classification by typical auditory categories

First, ALE maps were computed for each of the four typical auditory categories: pure tones, noise, music and vocal sounds. Each resulting map shows regions exhibiting consistent activity across studies for each sound category. For example, the “music” map shows the voxel-wise probability of activity for all “musical stimuli vs. baseline” contrasts.

Next, we examined hemispheric specialization effects by directly comparing the “music” and “vocal” sound categories. We directly compared a random subsample of the “vocal” sounds category (Table A1 in Appendix; 20 contrasts, 156 foci) to the “music” category (10 contrasts, 175 foci). This procedure ensured that the resulting ALE maps would reflect activity differences between studies rather than the imbalance in coordinate numbers between those categories (Laird et al., 2005). Then, as lateralization effects are reported for intelligible speech rather than vocal sounds, only contrasts using intelligible speech with semantic content, such as words or sentences, were included. The “music” and the “speech” categories were directly compared to investigate the expected lateralization effects. Given that many contrasts fell into the intelligible speech category, we selected only one contrast per study (Table A1 in Appendix), including a total of 27 contrasts (166 foci).

Finally, we assessed cortical auditory specialization for processing intelligible speech. Given that specialized auditory processes can be more easily isolated when the contrasting stimuli are as close as possible to the stimuli of interest in terms of acoustic complexity (Binder et al., 2000; Uppenkamp et al., 2006), contrasts containing unintelligible spectrally and temporally complex sounds were used as for comparison. Thirteen contrasts (76 foci, see Table A1 in Appendix) selected included reversed words, pseudowords, recorded scanner noise, single formant, environmental sounds, and modulated complex sounds. We directly compared the intelligible speech and complex non-speech sound categories.

Analysis using classification by auditory complexity

To investigate the relevance of acoustic complexity as a stimulus property predicting functional auditory specialization, we computed ALE maps for each level of complexity. Given that only one contrast fell into the “single, irregular change” dimension, this analysis was not conducted. Moreover, as most of the contrasts were classified as “multiple frequencies, irregular modulation” (70 contrasts, 517 foci), a randomly selected subsample of 10 contrasts (70 foci, see Table A1 in Appendix) were selected from this level of complexity to facilitate comparison of activity extent between levels.

Next, we examined effects related to auditory complexity. For the frequency domain, all contrasts falling in the “multiple” level (26 contrasts, 168 foci) were directly compared to those in the “single” level (18 contrasts, 64 foci), independent of their variation over time, (Figure 5A, bottom row vs. top row, green arrow). For the time dimension, comparisons were made between the contrasts including stimulus changes over time (regular and irregular; 25 experiments, 153 foci) and those who did not (unchanging; 19 contrasts, 79 foci), independent of their frequency composition (Figure 5A, middle and right column vs. left column, blue arrow).

Results

Stimulus classification using typical auditory categories

We observed different patterns of activity corresponding to the typical sound categories of pure tones, noise, music, and vocal sounds (Figure 1; Table 2). For all the categories, the strongest effects were found in auditory cortex (Brodmann areas 41, 42, and 22). For the pure tone map, high ALE values were found bilaterally in medial Heschl's gyri (HG). The noise map revealed effects in right medial HG and bilaterally in STG posterior and lateral to HG. Effects related to music were seen in HG, anterior and posterior STG. Finally, vocal sounds elicited large bilateral clusters of activity in HG as well as anterior, posterior, and lateral aspects of the STG. While pure tone effects were restricted to auditory cortex, effects outside temporal cortex were observed for the other categories. Additional activity was seen in frontal cortex for noise (BA 6, 9), music (BA 4, 6, 44, 45, 46), and vocal sounds (BA 45). Effects were observed in cerebellum for noise and music as well as in the anterior cingulate gyrus for vocal sounds.

Figure 1

Table 2

Coordinates
RegionBAxyzVolume (mm3)ALE (×10-3)
PURE TONES
Temporal
   Heschl's gyrus4153−171386426.00
   Heschl's gyrus41−51−201260016.85
NOISE
Temporal
   Superior temporal gyrus2268−235154412.98
   Superior temporal gyrus42−50−2915138415.58
   Superior temporal gyrus2251−24−164811.30
Frontal
   Precentral gyrus9−4612386648.67
   Superior frontal gyrus613145755211.36
   Inferior frontal gyrus945142049610.28
Sub cortical
   Cerebellum33−65−283607.83
MUSIC
Temporal lobe
   Superior temporal gyrus2253−112674421.07
   Superior temporal gyrus22−51−12−7484828.52
   Superior temporal gyrus225512−9109618.21
   Superior temporal gyrus22−493−1325611.58
Frontal lobe
   Inferior frontal gyrus45513418126414.91
   Inferior frontal gyrus45−422412124014.04
   Precentral gyrus44−5717436012.41
   Medial frontal gyrus6−2−36735214.39
   Precentral gyrus454−25128012.31
Sub cortical
   Cerebellum−31−62−2091214.10
VOCAL SOUNDS
Temporal
   Superior temporal gyrus22−59−12−522648134.64
   Superior temporal gyrus2262−13−418088101.69
Frontal
   Inferior frontal gyrus45−481726211221.16
Limbic
   Cingulate gyrus24−5134658418.72

Category classification.

MNI coordinates of the locations of significant ALE maxima for each sound category (p < 0.01, k = 250 voxels).

Effects related to typical sound categories were lateralized. Qualitative examination revealed larger clusters in right auditory cortex for music and in left auditory cortex for vocal sounds (Table 2). The direct comparisons between the musical and vocal sounds and between the musical and speech sounds yielded similar findings (Figure 2; Table 3). Greater activity related to music was observed bilaterally in posterior and anterolateral HG, the planum polare, and the most anterior parts of the right STG. We also observed effects related to music processing outside the temporal lobe, in inferior frontal gyrus (BA 45), the middle frontal gyrus (BA 6), and the left cerebellum (lobule IV). On the other hand, the reverse comparisons revealed stronger activity for vocal sounds as well as for speech in lateral HG, extending to lateral and anterior STG. For the vocal sounds, the extent of auditory activity was greater on the left (10312 voxels) than on the right (4952 voxels), however the ALE values were similar on the left (45.66 × 10−3) and on the right (42.24 × 10−3). As for the speech sounds, both the volume of activity and the corresponding ALE were greater on the left (11112 voxels, 61.39 × 10−3) than the right (5736 voxels, 38.21 × 10−3) hemisphere.

Figure 2

Table 3

Coordinates
RegionBAxyzVolume (mm3)ALE (×10-3)
MUSIC > VOICES
Temporal
   Heschl's gyrus41−49−11492022.64
   Heschl's gyrus41−40−301080015.25
   Heschl's gyrus4151−9068017.15
   Superior temporal gyrus225312−958416.18
   Heschl's gyrus4142−271133613.18
Frontal
   Inferior frontal gyrus4551341877614.75
   Inferior frontal gyrus45−40271438413.00
   Middle frontal gyrus6−2−36737614.39
Subcortical
   Cerebellum−31−65−2039213.04
VOICES > MUSIC
Temporal
   Superior temporal gyrus22−59−10−51031245.66
   Superior temporal gyrus2259−19−1495242.24
   Middle frontal gyrus9−481926103217.67
MUSIC > SPEECH
Temporal
   Superior temporal gyrus225512−972817.31
   Heschl's gyrus41−49−9469620.77
   Heschl's gyrus41−38−28848012.75
   Heschl's gyrus4151−9240814.44
   Heschl's gyrus4145−271133615.08
Frontal
   Inferior frontal gyrus45−402714104813.32
   Inferior frontal gyrus4551341856014.37
   Middle frontal gyrus6−2−56830414.33
Subcortical
   Cerebellum−31−62−2072814.09
SPEECH > MUSIC
Temporal
   Superior temporal gyrus22−59−12−51111261.39
   Superior temporal gyrus2262−13−4573638.21
Frontal
   Medial frontal gyrus8−16373733614.13

Lateralization effects.

MNI coordinates of the locations of significant ALE maxima resulting from the comparison between speech and music (p < 0.01, k = 250 voxels).

We observed specialization for speech processing in auditory cortex. The comparison between intelligible speech and complex non-speech sounds, including vocal sounds without intelligible content, is shown in Table 4 and Figure 3A. Speech was associated with greater activity in non-primary (BA 22) and associative (BA 39) auditory areas, lateral STG, bilateral anterior and middle STS, and the planum temporale (PT). These clusters were larger and had higher ALE values in the left hemisphere. We also observed stronger left prefrontal cortical activity (BA 8) for speech sounds. The reverse comparison yielded stronger activity related to complex non-speech sounds in the right PT (x = 68, y = −27, z = 8, 128 voxels; Figure 3A). The ALE maps associated with speech intelligibility had overlap with the vocal sound category maps (Figure 3B). While large bilateral clusters were observed along the STG and STS for the vocal sounds, there was specific sensitivity to speech intelligibility in the left anterior STG.

Table 4

Coordinates
RegionBAxyzVolume (mm3)ALE (×10-3)
INTELLIGIBLE SPEECH > COMPLEX NON-SPEECH
Temporal
   Superior temporal gyrus22−59−10−5876039.48
   Superior temporal gyrus2259−11−5441628.91
   Superior temporal gyrus39−48−533727213.40
Frontal
   Medial frontal gyrus8−16373742414.21
COMPLEX NON-SPEECH > INTELLIGIBLE SPEECH
No suprathreshold voxels

Functional specialization for speech sounds.

MNI coordinates of the locations of significant ALE maxima resulting from the comparison between speech and complex non-speech contrasts (p < 0.01, k = 250 voxels).

Figure 3

Stimulus classification using auditory complexity

Classification of sounds with respect to their spectral and temporal complexity revealed effects in the temporal lobe (Table 5; Figure 4). The “single, unchanging” stimulus class was associated with two clusters centered on medial HG (BA 41). The “single, regular change” stimulus class was associated with two large bilateral clusters of activity in medial and lateral HG, extending around HG into the anterolateral STG. On the left, we observed one additional peak of activity in posterior STG. For the “multiple, unchanging” stimulus class, temporal lobe activity was centered on medial HG and posterior STG. Effects for the “multiple, regular change” stimulus class were observed in HG, extending to the posterolateral STG. Finally, the “multiple, irregular change” stimulus class was associated with large bilateral effects in, and posterior to, HG. The complexity level maps revealed effects outside the temporal lobe, in frontal cortex areas BA 6, 9, 36, and 47 for the “multiple, unchanging” and “multiple, regular change” stimulus classes. We also observed effects in the cerebellum for the “single, regular change” and “multiple, irregular change” stimulus classes.

Table 5

Coordinates
RegionBAxyzVolume (mm3)ALE (×10-3)
COMPLEXITY LEVELS
Single unchanging
Temporal
   Heschl's gyrus4153−171388026.00
   Heschl's gyrus41−51−180248816.85
Single regular change
Temporal
   Superior temporal gyrus2257−112620031.94
   Superior temporal gyrus22−57−155408024.31
Sub cortical
   Cerebellum23−62−213527.98
Multiple unchanging
Temporal
   Heschl's gyrus41−40−2710724018.91
   Heschl's gyrus4153−211610417.69
Frontal
   Inferior frontal gyrus9−4612384808.67
   Superior frontal gyrus613145744811.36
   Inferior frontal gyrus945152040810.28
Multiple regular change
Temporal
   Superior temporal gyrus4266−215625622.56
   Superior temporal gyrus42−59−2112609621.90
Frontal
   Inferior frontal gyrus466235784815.86
   Inferior frontal gyrus47−4227−468812.63
Multiple irregular change
Temporal
   Superior temporal gyrus2257−191834424.12
   Superior temporal sulcus22−57−14−2604023.61
Subcortical
   Cerebellum18−58−2467212.44
−31−64−1855210.27
COMPARISONS BETWEEN COMPLEXITY LEVELS
Multiple > single
Temporal
   Superior temporal gyrus22−59−2310835233.26
   Superior temporal gyrus2268−215825642.55
Frontal
   Inferior frontal gyrus466235744815.86
   Inferior frontal gyrus47−4227−438413.50
Subcortical
   Cerebellum29−56−2687214.98
   Thalamus14−19235212.62
Changing > unchanging
Temporal
   Superior temporal gyrus2259−112880829.96
   Superior temporal gyrus22−57−9−3555233.22
   Superior temporal sulcus22−64−34733610.54
Frontal
   Inferior frontal gyrus466235744015.85
   Inferior frontal gyrus47−4227−436013.50
Subcortical
   Cerebellum23−58−24107217.22

Complexity classification.

MNI coordinates of the locations of significant ALE maxima for each level of acoustic complexity and comparison between levels (p < 0.01, k = 250 voxels).

Figure 4

Effects related to stimulus spectral and temporal variations were identified by comparing, respectively, the multiple to the single stimulus classes (independent of changes over time; Figure 5B, GREEN) and the changing to the unchanging stimulus classes (independent of the number of frequency components; Figure 5B, BLUE). The coordinates of the effects related to increasing auditory complexity are reported in Table 5. Overlapping sensitivity to spectral and temporal effects was observed in the lateral portion of HG. Increasing numbers of frequency components were associated with greater effects in posterior and lateral non-primary auditory fields, specifically bilateral posterolateral STG and PT. Modulatory effects were also seen in inferior frontal gyrus (BA 45, 47). In contrast, the effects related to temporal modulations compared to their absence were observed in HG, anterior STG, anterior STS, inferior frontal cortex (BA 46, 47), and right cerebellum (lobule IV).

Figure 5

Discussion

Summary of findings

In a quantitative meta-analysis of 58 neuroimaging studies, we examined the functional specialization of human auditory cortex using two different strategies for classifying sounds. The first strategy employed typical categories, such as pure tones, noise, music, and vocal sounds. The second strategy categorized sounds according to their acoustical (spectral and temporal) complexity.

Activation Likelihood Estimation maps computed for each typical sound category included simple (pure tones) and complex (noise, voices, and music) sounds. This analysis gave results consistent with models describing hierarchical functional organization of the human auditory cortex, with simple sounds eliciting activity in the primary auditory cortex and complex sound processing engaging additional activity in non-primary fields. We observed an expected leftward hemispheric specialization for intelligible speech, while right-hemisphere specialization for music was less evident. Additionally, the comparison of intelligible speech to complex non-speech stimuli yielded bilateral effects along the STG and STS, with higher sensitivity to speech intelligibility in the left anterior STG.

Examining an alternative classification based on stimulus variation along spectral and temporal dimensions, we observed a within-hemisphere functional segregation, with spectral effects strongest in posterior STG and temporal modulations strongest in anterior temporal STG. We suggest that acoustic complexity might represent a valid alternative classificatory scheme to describe a novel within-hemisphere dichotomy regarding the functional organization for auditory processing in temporal cortex.

Hierarchically and hemispherically specialized architectures for auditory processing

Originally elaborated on the basis of non-human primate studies, the hierarchical functional organization scheme in auditory cortex incorporates three levels of processing: core (primary area), belt and parabelt (non-primary areas). Simple sound processing is thought to solely recruit the core region whereas complex sounds are believed to elicit activity in core, belt, and parabelt areas. While belt region responses are thought to be sensitive to acoustic feature variations, the parabelt, and more anterior temporal regions, show greater sensitivity to complex sounds such as vocalizations (Rauschecker, 1998; Hackett, 2008; Rauschecker and Scott, 2009; Woods and Alain, 2009). Our quantitative meta-analysis using typical sound classes confirmed that hierarchical processing is a feature that can adequately describe human auditory cortical organization.

Using an ALE analysis of pure tone processing to investigate the correspondence between the core region and activity related to simple sound processing, we observed ALE extrema values bilaterally in medial HG, the putative location of primary auditory cortex. This finding is consistent with previous electrophysiological (Hackett et al., 2001), cytoarchitectural (Sweet et al., 2005), and functional imaging (Lauter et al., 1985; Bilecen et al., 1998; Lockwood et al., 1999; Wessinger et al., 2001) studies of the human auditory cortex that have localized the core region to medial HG. Our findings confirm the existence of functional specialization for simple sound processing in the human core homolog. Consequently, the statistical probability maps obtained here could serve to functionally define primary auditory cortex in a region of interest analysis of functional neuroimaging data.

In contrast, we expected ALE analyses of the complex sound categories to show activity in all three levels of the processing hierarchy. We observed overlapping activity among the complex sound maps in medial HG (core) as well as stronger activity related to complex sound processing in regions surrounding medial HG, corresponding to the areas described as the auditory belt/parabelt in primates (Rauschecker, 1998; Kaas and Hackett, 2000; Rauschecker and Scott, 2009; Recanzone and Cohen, 2010) and humans (Rivier and Clarke, 1997; Wallace et al., 2002; Sweet et al., 2005). The fact that the complex sound maps showed effects in medial HG activity supports the notion that primary auditory regions participate in the early stages of processing upon which further complex processing is built.

Outside primary auditory cortex, noise elicited activity in posterior temporal non-primary fields such as PT. The spatial pattern was similar to that observed in relation to broadband noise, stimuli that have been used to demonstrate the hierarchical organization of human auditory cortex (Wessinger et al., 2001). The PT is generally believed to be involved in complex sound analysis and participate in both language and other cognitive functions (Griffiths and Warren, 2002).

For music, in addition to primary auditory cortex activity, we observed activity in non-primary auditory fields along the STG bilaterally. This result is consistent with the idea that simple extraction and low-level ordering of pitch information involves processes within primary auditory fields, while higher-level processing for tone patterns and melodies involve non-primary auditory fields and association cortex (Zatorre et al., 2002). Moreover, non-primary regions in anterior and posterior STG are thought to process melody pitch intervals (Patterson et al., 2002; Tramo et al., 2002; Warren and Griffiths, 2003). Music also elicited strong inferior frontal cortex activity, a region thought to process musical syntax (Zatorre et al., 1994; Maess et al., 2001; Koelsch et al., 2002).

For vocal sounds, we observed strong bilateral temporal lobe activity in anterior and posterior parts of dorsal STG and the STS, findings consistent with earlier studies (Binder et al., 1994; Belin, 2006). STG activity in response to vocal sounds has previously been interpreted as a neural correlate of the rapid and efficient processing of the complex frequency patterns and temporal variations characterizing speech. The human STG is thought to subserve complex auditory processing, such as vocalizations, as is the STG in non-human primates (Rauschecker et al., 1995). Belin and colleagues (Belin et al., 2000, 2002; Fecteau et al., 2004) reported cortical responses to voices along the upper bank of the middle and anterior STS. The anterior STS is selectively responsive to human vocal sounds (Belin et al., 2000). Response specificity to vocal sounds and their rich identity and affective information content is of crucial importance, as it reflects a set of high-level auditory cognitive abilities that can be directly compared between human and non-human primates. The regions described as “Temporal Voice Areas” in humans (Belin et al., 2000) are thought to be functionally homologous to the temporal voice regions recently described in macaques (Petkov et al., 2008). Our meta-analysis using typical sound categories demonstrates that, in humans, simple sound processing elicits activity limited to the core area while complex sounds elicit effects in all three cortical processing levels.

In addition to the hierarchical organization of auditory cortex, we expected hemispheric asymmetries for music and speech, and observed the expected left lateralization of auditory cortex responses to vocal sounds and intelligible speech. For vocal sounds, lateralization effects were observed only as a larger volume of auditory activity on the left while, for the speech sounds, the left auditory cortical responses were larger and stronger (higher ALE values) than the right-hemisphere responses. Greater lateralization effects for intelligible speech is in agreement with previous independent imaging studies, not included in this meta-analysis, reporting that intelligible speech sounds elicit strong activity in left STG and STS (e.g., Scott et al., 2000; Liebenthal et al., 2005; Obleser et al., 2007). Conversely, we did not see the expected right response lateralization related to music. Possibly, the small number of experiments included in the music category limited the power of this analysis and could have prevented us from observing the expected rightward auditory response. ALE maps derived from small samples are more sensitive to between-study cohort heterogeneity that could limit the detection of hemispheric effects. It is also possible that the right hemisphere is sensitive to particular features of musical stimuli such as fine pitch changes (Hyde et al., 2008) or to specific task demands like contextual pitch judgment (i.e., contextual pitch judgment Warrier and Zatorre, 2004) which were not present in our sample.

Response specificity to speech intelligibility

Within the general category of vocal sounds, a human-specific category of intelligible speech can be further distinguished. Response specificity to speech intelligibility is an important part of understanding the human-specific neural network underlying speech comprehension, and ultimately human language and communication.

In order to identify speech-specific processes, we directly compared intelligible speech to complex non-speech contrasts that included unintelligible spectro-temporally complex sounds. This comparison yielded stronger speech-related activity in lateral non-primary superior temporal regions, specifically in posterior STG, and anterior and middle STS. The effects were stronger and larger in the left hemisphere. Similar effects have been reported in independent studies examining specialization for processing speech sound that did not fulfill our inclusion criteria for this analysis (Scott et al., 2000; Davis and Johnsrude, 2003; Narain et al., 2003; Thierry et al., 2003; Liebenthal et al., 2005). Consistent with the present finding, these previous reports emphasized that speech-specific STS responses are more left-lateralized.

Beyond the auditory cortex, we observed activity in left inferior frontal and prefrontal cortex. These findings support an expanded hierarchical model of speech processing that originates in primary auditory areas and extends to non-auditory regions, mainly within frontal cortex, in a range of motor, premotor, and prefrontal regions (Davis and Johnsrude, 2007; Hickok and Poeppel, 2007; Rauschecker and Scott, 2009). In non-human primates, based on reports of high level of connection between the auditory and frontal cortex, it has been proposed that frontal regions responsive to auditory material should be considered as part of the auditory system (Hackett et al., 1999; Kaas et al., 1999; Romanski et al., 1999).

Functional specialization of the auditory cortex response: acoustic complexity effects

As an alternative to the classical division of auditory stimuli into typical categories like pure tones, noise, voices, and music, we explored how acoustic variations along the temporal and spectral dimensions were represented at the cortical level. This approach for defining auditory material is an efficient and comprehensive characterization of sounds that can be considered as a complement to the more typically studied categorical effects. Possibly, certain aspects of human auditory processes might be better characterized in terms of their capacity to analyze acoustic features rather than having differential sensitivity to typical sound categories. In a meta-analysis Rivier and Clarke (1997) found no clear functional specialization in non-primary auditory fields for a range of complex sound categories, showing that processing sounds of different categories such as noise, words, and music, elicited activity in multiple non-primary fields around HG with no emergence of a specific organizational pattern. Similarly, Griffiths and Warren (2002) reported that activity within the PT, an auditory association region, is not spatially organized according to sound categories such as music, speech or, environmental sounds.

By classifying sounds according to their variations in time and frequency, we isolated different levels of auditory complexity, suggesting a within-hemisphere functional segregation with anterior STG and STS more sensitive to changes in the temporal domain and posterior regions (PT and posterolateral STG) more sensitive to changes in along the spectral dimension. Interestingly, a partial overlap was observed between regions sensitive to temporal and spectral changes in lateral HG, suggesting great sensitivity to variations in acoustic properties within this region, consistent with a recent report of strongest sensitivity to stimulus acoustic features within HG (Okada et al., 2010).

Our observation of differential sensitivity to temporal and spectral features can be interpreted in the light of previous findings. First, in the animal literature, a within-hemisphere model of spectral and temporal processing in the auditory cortex has been proposed (Bendor and Wang, 2008). This scheme suggests two streams of processing originating from primary auditory cortex; an anterior pathway sensitive to temporal changes and a lateral pathway responsive to spectral changes. More precise temporal coding is seen as one progresses from primary to anterior auditory regions in primates (Bendor and Wang, 2007) and greater sensitivity to temporal modulations in anterior non-primary auditory fields is also observed in cats (Tian and Rauschecker, 1994). Possibly, a longer integration window in anterior auditory fields could underlie complex temporal processing (Bendor and Wang, 2008). As regards spectral processing, increasing sensitivity to broadband spectrum noise compared to single tones has been observed in lateral and posterior auditory fields in non-human primates (Rauschecker and Tian, 2004; Petkov et al., 2006). Furthermore, given that the neurons within these regions show strong tuning to bandwidth and frequency, some have suggested their involvement in the early stages of spectral analysis of complex sounds (Rauschecker and Tian, 2004). In our study, sensitivity to temporal changes was observed in anterior temporal regions, while, in response to changes along the spectral dimension, we mainly observed response selectivity in posterolateral auditory fields. Our results therefore seem to be consistent with previous animal studies.

Second, cortical response specificity to spectral and temporal processing has also been studies in humans. Whereas some studies reported no clear functional segregation between responses to spectral and temporal cues (Hall et al., 2002) or observed neuronal populations tuned to specific combinations of spectro-temporal cues (Schönwiesner and Zatorre, 2009), other studies found the sorts of specific sensitivity to spectral vs. temporal features in human auditory cortex we observed in our meta-analysis. For instance, lateral HG and anterolateral PT activity have been reported in association with fine spectral structure analysis (Warren et al., 2005) and change detection of complex harmonic tones involved the posterior STG and lateral PT (Schönwiesner et al., 2007). Additionally, recent studies examining effective connectivity effects among auditory regions reported that spectral envelope analysis follows a serial pathway from HG to PT and then to the STS (Griffiths et al., 2007; Kumar et al., 2007). Conversely for temporal complexity effects, a stream of processing from primary auditory cortex to anterior STG has been observed for auditory pattern analysis such as dynamic pitch variation (Griffiths et al., 1998). Similarly, significant effects of temporal modulation have been reported in anterior non-primary auditory fields (Hall et al., 2000). Some studies therefore report patterns of activity consistent with the current findings, albeit separately for spectral and temporal features.

A more frequently observed feature of spectral vs. temporal processing is between-hemisphere functional specialization. Most studies observed slight but significant lateralization effects with a left-lateralized response to temporal information and right-lateralized activity to spectral information (Zatorre and Belin, 2001; Schönwiesner et al., 2005; Jamison et al., 2006; Obleser et al., 2008). In the current study, lateralization effects were not seen with regard to complexity. However, at higher processing levels, leftward lateralization for speech was observed. Others studies failing to demonstrate the expected lateralization proposed that early stages of processing involve bilateral auditory cortex and that higher cognitive functions, such as speech processing, also rely on these regions but involve more extensive regions in the dominant hemisphere (Langers et al., 2003). Alternatively, Tervaniemi and Hugdahl (2003) reviewed studies showing that response lateralization within the auditory cortex is dependent on sound structure as well as the acoustic background they are presented in. For instance, reduced or absent hemispheric specialization for speech sounds has been reported when the amount of formant structure is not sufficient to establish phoneme categorization (Rinne et al., 1999) or when sounds are presented in noise (Shtyrov et al., 1998). Stimulus heterogeneity among the different experiments included in our meta-analysis could explain why we did not observe asymmetrical hemispheric effects.

To summarize, our meta-analysis demonstrates a clear within-hemisphere functional segregation related to spectral and temporal processing in human auditory cortex, consistent with the known organization of non-human primate auditory system. That such clear spectral vs. temporal complexity gradients are observed (Figure 5), while very few of the included studies have explicitly addressed this issue, illustrates the power of the meta-analysis approach for human neuroimaging studies. Based on the observed regional functional segregation, we argue that acoustic complexity could well represent a relevant stimulus dimension upon which to identify response segregation within the auditory system. Complexity and categorical effects could therefore be considered as two complementary approaches to more fully characterizing the underlying nature of auditory regional functional specialization.

Statements

Acknowledgments

We would like to thank Dr. Laurent Mottron and Dr. Valter Ciocca for providing comments on the manuscript and suggestions regarding the stimulus categorizations. This work was supported by Canadian Institutes for Health Research (grant MOP-84243) as well as a doctoral award from Natural Sciences and Engineering Research Council of Canada to Fabienne Samson.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    BelinP. (2006). Voice processing in human and non-human primates. Philos. Trans. R. Soc. Lond., B, Biol. Sci.361, 20912107.

  • 2

    BelinP.ZatorreR. J.AhadP. (2002). Human temporal-lobe response to vocal sounds. Brain Res. Cogn. Brain Res.13, 1726.

  • 3

    BelinP.ZatorreR. J.HogeR.EvansA. C.PikeB. (1999). Event-related fMRI of the auditory cortex. Neuroimage10, 417429.10.1006/nimg.1999.0480

  • 4

    BelinP.ZatorreR. J.LafailleP.AhadP.PikeB. (2000). Voice-selective areas in human auditory cortex. Nature403, 309312.10.1038/35002078

  • 5

    BelinP.ZilboviciusM.CrozierS.ThivardL.FontaineA.MasureM.-C.SamsonY. (1998). Lateralization of speech and auditory temporal processing. J. Cogn. Neurosci.10, 4, 536540.

  • 6

    BendorD.WangX. (2007). Differential neural coding of acoustic flutter within primate auditory cortex. Nat. Neurosci.10, 763771.

  • 7

    BendorD.WangX. (2008). Neural response properties of primary, rostral, and rostrotemporal core fields in the auditory cortex of marmoset monkeys. J. Neurophysiol.100, 888906.10.1152/jn.00884.2007

  • 8

    BensonR. R.RichardsonM.WhalenD. H.LaiS. (2006). Phonetic processing areas revealed by sinewave speech and acoustically similar non-speech. Neuroimage31, 342353.10.1016/j.neuroimage.2005.11.029

  • 9

    BensonR. R.WhalenD. H.RichardsonM.SwainsonB.ClarkV. P.LaiS.LibermanA. M. (2001). Parametric dissociating speech and nonspeech perception in the brain using fMRI. Brain Lang.78, 364396.10.1006/brln.2001.2484

  • 10

    BilecenD.SchefflerK.SchmidN.TschoppK.SeeligJ. (1998). Tonotopic organization of the human auditory cortex as detected by BOLD-FMRI. Hear. Res.126, 1927.

  • 11

    BinderJ. R.FrostJ. A.HammekeT. A.BellgowanP. S. F.SpringerJ. A.KaufmanJ. N.PossingE. T. (2000). Human temporal lobe activation by speech and nonspeech sounds. Cereb. Cortex10, 512528.10.1093/cercor/10.5.512

  • 12

    BinderJ. R.FrostJ. A.HammekeT. A.RaoS. M.CoxR. W. (1996). Function of the left planum temporale in auditory and linguistic processing. Brain119(Pt 4), 12391247.10.1093/brain/119.4.1239

  • 13

    BinderJ. R.RaoS. M.HammekeT. A.YetkinF. Z.JesmanowiczA.BandettiniP. A.WongE. C.EstkowskiL. D.GoldsteinM. D.HaughtonV. M. (1994). Functional magnetic resonance imaging of human auditory cortex. Ann. Neurol.35, 662672.

  • 14

    BinderJ. R.SwansonS. J.HammekeT. A.SabsevitzD. S. (2008). A comparison of five fMRI protocols for mapping speech comprehension systems. Epilepsia49, 19801997.

  • 15

    BloodA. J.ZatorreR. J.BermudezP.EvansA. C. (1999). Emotional responses to pleasant and unpleasant music correlate with activity in paralimbic brain regions. Nat. Neurosci.2, 382387.

  • 16

    BrownS.MartinezM. J.ParsonsL. M. (2004). Passive music listening spontaneously engages limbic and paralimbic systems. Neuroreport15, 20332037.10.1097/00001756-200409150-00008

  • 17

    BurtonM. W.SmallS. L. (2006). Functional neuroanatomy of segmenting speech and nonspeech. Cortex42, 644651.10.1016/S0010-9452(08)70400-3

  • 18

    BurtonM.W.; SmallS. L.BlumsteinS. E. (2000). The role of segmentation in phonological processing: an fMRI investigation. J. Cogn. Neurosci.12, 679690.

  • 19

    CelsisP.BoulanouarK.DoyonB.RanjevaJ. P.BerryI.NespoulousJ. L.CholletF. (1999). Differential fMRI responses in the left posterior superior temporal gyrus and left supramarginal gyrus to habituation and change detection in syllables and tones. Neuroimage9, 135144.10.1006/nimg.1998.0389

  • 20

    ChenJ. L.PenhuneV. B.ZatorreR. J. (2008). Listening to musical rhythms recruits motor regions of the brain. Cereb. Cortex18, 28442854.10.1093/cercor/bhn042

  • 21

    ChubbC.SperlingG. (1988). Drift-balanced random stimuli: a general basis for studying non-Fourier motion perception. J. Opt. Soc. Am. A5, 19862007.

  • 22

    Dalla BarbaG.ParlatoV.JobertA.SamsonY.PappataS. (1998). Cortical networks implicated in semantic and episodic memory: common or unique?Cortex34, 547561.10.1016/S0010-9452(08)70513-6

  • 23

    DavisM. H.JohnsrudeI. S. (2003). Hierarchical processing in spoken language comprehension. J. Neurosci.23, 34233431.

  • 24

    DavisM. H.JohnsrudeI. S. (2007). Hearing speech sounds: top-down influences on the interface between audition and speech perception. Hear. Res.229, 132147.

  • 25

    Dehaene-LambertzG.PallierC.SerniclaesW.Sprenger-CharollesL.JobertA.DehaeneS. (2005). Neural correlates of switching from auditory to speech perception. Neuroimage24, 2133.10.1016/j.neuroimage.2004.09.039

  • 26

    DémonetJ.-F.CholletF.RamsayS.CardebatD.NespoulousJ.WiseR.RascolA.FrackowiakR. (1992). The anatomy of phonological and semantic processing in normal subjects. Brain115, 17531768.10.1093/brain/115.6.1753

  • 27

    De NilL. F.BealD. S.LafailleS. J.KrollR. M.CrawleyA. P.GraccoV. L. (2008). The effects of simulated stuttering and prolonged speech on the neural activation patterns of stuttering and nonstuttering adults. Brain Lang.107, 114123.10.1016/j.bandl.2008.07.003

  • 28

    DuforO.SerniclaesW.Sprenger-CharollesL.DémonetJ. F. (2007). Top-down processes during auditory phoneme categorization in dyslexia: a PET study. Neuroimage34, 16921707.10.1016/j.neuroimage.2006.10.034

  • 29

    EngelienA.TüscherO.HermansW.IsenbergN.EidelbergD.FrithC.SternE.SilbersweigD. (2006). Functional neuroanatomy of non-verbal semantic sound processing in humans. J. Neural Transm.113, 599608.10.1007/s00702-005-0342-0

  • 30

    FecteauS.ArmonyJ. L.JoanetteY.BelinP. (2004). Is voice processing species-specific in human auditory cortex? An fMRI study. Neuroimage23, 840848.10.1016/j.neuroimage.2004.09.019

  • 31

    Fernandez-EspejoD.JunqueC.VendrellP.BernabeuM.RoigT.BargalloN.MercaderJ. M. (2008). Cerebral response to speech in vegetative and minimally conscious states after traumatic brain injury. Brain Inj.22, 882890.10.1080/02699050802403573

  • 32

    FoxP. T.ParsonsL. M.LancasterJ. L. (1998). Beyond the single study: function/location metanalysis in cognitive neuroimaging. Curr. Opin. Neurobiol.8, 178187.

  • 33

    GaabN.GabrieliJ. D. E.GloverG. H. (2007a). Assessing the influence of scanner background noise on auditory processing. I. An fMRI study comparing three experimental designs with varying degrees of scanner noise. Hum. Brain Mapp.28, 703720.10.1002/hbm.20298

  • 34

    GaabN.GabrieliJ. D. E.GloverG. H. (2007b). Assessing the influence of scanner background noise on auditory processing. II. An fMRI study comparing auditory processing in the absence and presence of recorded scanner noise using a sparse design. Hum. Brain Mapp.28, 721732.10.1002/hbm.20299

  • 35

    GiraudA. L.KellC.ThierfelderC.SterzerP.RussM. O.PreibischC.KleinschmidtA. (2004). Contributions of sensory input, auditory search and verbal comprehension to cortical activity during speech processing. Cereb. Cortex14, 247255.10.1093/cercor/bhg124

  • 36

    GriffithsT. D.WarrenJ. D. (2002). The planum temporale as a computational hub. Trends Neurosci.25, 348353.10.1016/S0166-2236(02)02191-4

  • 37

    GriffithsT. D.WarrenJ. D. (2004). What is an auditory object?Nat. Rev. Neurosci.5, 887892.

  • 38

    GriffithsT. D.BuchelC.FrackowiakR. S.PattersonR. D. (1998). Analysis of temporal structure in sound by the human brain. Nat. Neurosci.1, 422427.

  • 39

    GriffithsT. D.KumarS.WarrenJ. D.StewartL.StephanK. E.FristonK. J. (2007). Approaches to the cortical analysis of auditory objects. Hear. Res.229, 4653.

  • 40

    HackettT. A. (2008). Anatomical organization of the auditory cortex. J. Am. Acad. Audiol.19, 774779.

  • 41

    HackettT. A.PreussT. M.KaasJ. H. (2001). Architectonic identification of the core region in auditory cortex of macaques, chimpanzees, and humans. J. Comp. Neurol.441, 197222.

  • 42

    HackettT. A.StepniewskaI.KaasJ. H. (1999). Prefrontal connections of the parabelt auditory cortex in macaque monkeys. Brain Res.817, 4558.10.1016/S0006-8993(98)01182-2

  • 43

    HallD. A.HaggardM. P.AkeroydM. A.SummerfieldA. Q.PalmerA. R.ElliottM. R.BowtellR. W. (2000). Modulation and task effects in auditory processing measured using fMRI. Hum. Brain Mapp.10, 107119.10.1002/1097-0193(200007)10:3<107::AID-HBM20>3.0.CO;2-8

  • 44

    HallD. A.JohnsrudeI. S.HaggardM. P.PalmerA. R.AkeroydM. A.SummerfieldA. Q. (2002). Spectral and temporal processing in human auditory cortex. Cereb. Cortex12, 140149.10.1093/cercor/12.2.140

  • 45

    HartH. C.PalmerA. R.HallD. A. (2003). Amplitude and frequency modulated stimuli activate common regions of human auditory cortex. Cereb. Cortex13, 773781.10.1093/cercor/13.7.773

  • 46

    HartH. C.PalmerA. R.HallD. A. (2004). Different areas of human non-primary auditory cortex are activated by sounds with spatial and nonspatial properties. Hum. Brain Mapp.21, 178190.10.1002/hbm.10156

  • 47

    HertrichI.DietrichS.AckermannH. (2010). Cross-modal interactions during perception of audiovisual speech and nonspeech signals: an fMRI study. J. Cogn. Neurosci.23, 221237.

  • 48

    HickokG.PoeppelD. (2007). The cortical organization of speech processing. Nat. Rev. Neurosci.8, 393402.

  • 49

    HugdahlK.ThomsenT.ErslandL.RimolL. M.NiemiJ. (2003). The effects of attention on speech perception: an fMRI study. Brain Lang.85, 3748.10.1016/S0093-934X(02)00500-X

  • 50

    HwangJ. H.LiC. W.WuC. W.ChenJ. H.LiuT. C. (2007). Aging effects on the activation of the auditory cortex during binaural speech listening in white noise: an fMRI study. Audiol. Neurootol.12, 285294.10.1159/000103209

  • 51

    HydeK. L.PeretzI.ZatorreR. J. (2008). Evidence for the role of the right auditory cortex in fine pitch resolution. Neuropsychologia46, 632639.10.1016/j.neuropsychologia.2007.09.004

  • 52

    JamisonH. L.WatkinsK. E.BishopD. V.MatthewsP. M. (2006). Hemispheric specialization for processing auditory nonspeech stimuli. Cereb. Cortex16, 12661275.

  • 53

    JänckeL.ShahN. J.PosseS.Grosse-RyukenM.Müller-GärtnerH.-W. (1998). Intensity coding of auditory stimuli: an fMRI study. Neuropsychologia36, 875883.10.1016/S0028-3932(98)00019-0

  • 54

    KaasJ. H.HackettT. A. (2000). Subdivisions of auditory cortex and processing streams in primates. Proc. Natl. Acad. Sci. U.S.A.97, 1179311799.

  • 55

    KaasJ. H.HackettT. A.TramoM. J. (1999). Auditory processing in primate cerebral cortex. Curr. Opin. Neurobiol.9, 164170.

  • 56

    KoelschS.GunterT. C.v CramonD. Y.ZyssetS.LohmannG.FriedericiA. D. (2002). Bach speaks: a cortical “language-network” serves the processing of music. Neuroimage17, 956966.10.1006/nimg.2002.1154

  • 57

    KriegsteinK. V.GiraudA. L. (2004). Distinct functional substrates along the right superior temporal sulcus for the processing of voices. Neuroimage22, 948955.10.1016/j.neuroimage.2004.02.020

  • 58

    KumarS.StephanK. E.WarrenJ. D.FristonK. J.GriffithsT. D. (2007). Hierarchical processing of auditory objects in humans. PLoS Comput. Biol.3, e100.10.1371/journal.pcbi.0030100

  • 59

    LairdA. R.FoxP. M.PriceC. J.GlahnD. C.UeckerA. M.LancasterJ. L.TurkeltaubP. E.KochunovP.FoxP. T. (2005). ALE meta-analysis: controlling the false discovery rate and performing statistical contrasts. Hum. Brain Mapp.25, 155164.10.1002/hbm.20136

  • 60

    LancasterJ. L.Tordesillas-GutierrezD.MartinezM.SalinasF.EvansA.ZillesK.MazziottaJ. C.FoxP. T. (2007). Bias between MNI and Talairach coordinates analyzed using the ICBM-152 brain template. Hum. Brain Mapp.28, 11941205.10.1002/hbm.20345

  • 61

    LangersD. R.BackesW. H.van DijkP. (2003). Spectrotemporal features of the auditory cortex: the activation in response to dynamic ripples. Neuroimage20, 265275.10.1016/S1053-8119(03)00258-1

  • 62

    LarssonJ.LandyM. S.HeegerD. J. (2006). Orientation-selective adaptation to first- and second-order patterns in human visual cortex. J. Neurophysiol.95, 862881.10.1152/jn.00668.2005

  • 63

    LauterJ. L.HerscovitchP.FormbyC.RaichleM. E. (1985). Tonotopic organization in human auditory cortex revealed by positron emission tomography. Hear. Res.20, 199205.

  • 64

    LeaverA. M.RauscheckerJ. P. (2010). Cortical representation of natural complex sounds: effects of acoustic features and auditory object category. J. Neurosci.30, 76047612.

  • 65

    LewisJ. W.BrefczynskiJ. A.PhinneyR. E.JanikJ. J.DeYoeE. A. (2005). Distinct cortical pathways for processing tool versus animal sounds. J. Neurosci.25, 51485158.10.1523/JNEUROSCI.0419-05.2005

  • 66

    LewisJ. W.TalkingtonW. J.WalkerN. A.SpirouG. A.JajoskyA.FrumC.Brefczynski-LewisJ. A. (2009). Human cortical organization for processing vocalizations indicates representation of harmonic structure as a signal attribute. J. Neurosci.29, 22832296.10.1523/JNEUROSCI.4145-08.2009

  • 67

    LiebenthalE.BinderJ. R.SpitzerS. M.PossingE. T.MedlerD. A. (2005). Neural substrates of phonemic perception. Cereb. Cortex15, 16211631.10.1093/cercor/bhi040

  • 68

    LillywhiteL. M.SalingM. M.DemutskaA.MastertonR.FarquharsonS.JacksonG. D. (2010). The neural architecture of discourse compression. Neuropsychologia48, 873879.10.1016/j.neuropsychologia.2009.11.004

  • 69

    LockwoodA. H.SalviR. J.CoadM. L.ArnoldS. A.WackD. S.MurphyB. W.BurkardR. F. (1999). The functional anatomy of the normal human auditory system: responses to 0.5 and 4.0 kHz tones at varied intensities. Cereb. Cortex9, 6576.10.1093/cercor/9.1.65

  • 70

    MaessB.KoelschS.GunterT. C.FriedericiA. D. (2001). Musical syntax is processed in Broca's area: an MEG study. Nat. Neurosci.4, 540545.

  • 71

    MüllerR. A.BehenM. E.RothermelR. D.ChuganiD. C.MuzikO.MangnerT. J.ChuganiH. T. (1999). Brain-mapping of language and auditory perception in high-functioning autistic adults: a PET study. J. Autism Dev. Disord.29, 1931.

  • 72

    MummeryC. J.AshburnerJ.ScottS. K.WiseR. J. S. (1999). Functional neuroimaging of speech perception in six normal and two aphasic subjects. J. Acoust. Soc. Am.106, 449457.

  • 73

    MutschlerI.WieckhorstB.SpeckO.Schulze-BonhageA.HennigJ.SeifritzE.BallT. (2010). Time scales of auditory habituation in the amygdala and cerebral cortex. Cereb. Cortex. 20, 25312539.10.1093/cercor/bhq001

  • 74

    NarainC.ScottS. K.WiseR. J.RosenS.LeffA.IversenS. D.MatthewsP. M. (2003). Defining a left-lateralized response specific to intelligible speech using fMRI. Cereb. Cortex13, 13621368.10.1093/cercor/bhg083

  • 75

    ObleserJ.BoeckerH.DrzezgaA.HaslingerB.HennenlotterA.RoettingerM.EulitzC.RauscheckerJ. P. (2006). Vowel sound extraction in anterior superior temporal cortex. Hum. Brain Mapp.27, 562571.10.1002/hbm.20201

  • 76

    ObleserJ.EisnerF.KotzS. A. (2008). Bilateral speech comprehension reflects differential sensitivity to spectral and temporal features. J. Neurosci.28, 81168123.10.1523/JNEUROSCI.1290-08.2008

  • 77

    ObleserJ.ZimmermannJ.Van MeterJ.RauscheckerJ. P. (2007). Multiple stages of auditory speech perception reflected in event-related FMRI. Cereb. Cortex17, 22512257.

  • 78

    OkadaK.RongF.VeneziaJ.MatchinW.HsiehI. H.SaberiK.SerencesJ. T.HickokG. (2010). Hierarchical organization of human auditory cortex: evidence from acoustic invariance in the response to intelligible speech. Cereb. Cortex. 20, 24862495.10.1093/cercor/bhp318

  • 79

    O'LearyD.AndreasenN. C.HurtigR. R.HichwaR. D.WatkinsL.Boles PontoL. L.RogersM.KirchnerP. T. (1996). A positron emission tomography study of binaurally and dichotically presented stimuli: effects of level of language and directed attention. Brain Lang.53, 2039.10.1006/brln.1996.0034

  • 80

    PattersonR. D.UppenkampS.JohnsrudeI. S.GriffithsT. D. (2002). The processing of temporal pitch and melody information in auditory cortex. Neuron36, 767776.10.1016/S0896-6273(02)01060-7

  • 81

    PaulesuE.HarrisonJ.Bron-CohenS.WatsonJ. D. G.GoldsteinL.HeatherJ.FrackowiakR. S. J.FrithC. D. (1995). The physiology of coloured hearing. A PET activation study of colour-word synaesthesia, Brain118, 661676.

  • 82

    PeretzI.GosselinN.BelinP.ZatorreR. J.PlaillyJ.TillmannB. (2009). Music lexical networks: the cortical organization of music recognition. Ann. N. Y. Acad. Sci.1169, 256265.

  • 83

    PetersenS. E.FoxP. T.PosnerM. I.MintunM.RaichleM. E. (1988). Positron emission tomographic studies of the cortical anatomy of single-word processing. Nature331, 585589.10.1038/331585a0

  • 84

    PetkovC. I.KayserC.AugathM.LogothetisN. K. (2006). Functional imaging reveals numerous fields in the monkey auditory cortex. PLoS Biol.4, e215.10.1371/journal.pbio.0040215

  • 85

    PetkovC. I.KayserC.SteudelT.WhittingstallK.AugathM.LogothetisN. K. (2008). A voice region in the monkey brain. Nat. Neurosci.11, 367374.

  • 86

    PriceC.WiseR.RamsayS.FristonK.HowardD.PattersonK.FrackowiakR. (1992). Regional response differences within the human auditory cortex when listening to words. Neurosci. Lett.146, 179182.

  • 87

    RauscheckerJ. P. (1998). Cortical processing of complex sounds. Curr. Opin. Neurobiol.8, 516521.

  • 88

    RauscheckerJ. P.ScottS. K. (2009). Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat. Neurosci.12, 718724.

  • 89

    RauscheckerJ. P.TianB. (2000). Mechanisms and streams for processing of “what” and “where” in auditory cortex. Proc. Natl. Acad. Sci. U.S.A.97, 1180011806.

  • 90

    RauscheckerJ. P.TianB. (2004). Processing of band-passed noise in the lateral auditory belt cortex of the rhesus monkey. J. Neurophysiol.91, 25782589.10.1152/jn.00834.2003

  • 91

    RauscheckerJ. P.TianB.HauserM. (1995). Processing of complex sounds in the macaque nonprimary auditory cortex. Science268, 111114.10.1126/science.7701330

  • 92

    RecanzoneG. H.CohenY. E. (2010). Serial and parallel processing in the primate auditory cortex revisited. Behav. Brain Res.206, 17.10.1007/s00221-010-2312-5

  • 93

    Reynolds LosinE. A.RiveraS. M.O'HareE. D.SowellE. R.PinterJ. D. (2009). Abnormal fMRI activation pattern during story listening in individuals with Down syndrome. Am. J. Intellect. Dev. Disabil.114, 369380.

  • 94

    RimolL. M.SpechtK.WeisS.SavoyR.HugdahlK. (2005). Processing of sub-syllabic speech units in the posterior temporal lobe: an fMRI study. Neuroimage26, 10591067.10.1016/j.neuroimage.2005.03.028

  • 95

    RinneT.AlhoK.AlkuP.HoliM.SinkkonenJ.VirtanenJ.BertrandO.NäätänenR. (1999). Analysis of speech sounds is left-hemisphere predominant at 100–150 ms after sound onset. Neuroreport10, 11131117.10.1097/00001756-199904060-00038

  • 96

    RivierF.ClarkeS. (1997). Cytochrome oxidase, acetylcholinesterase, and NADPH-diaphorase staining in human supratemporal and insular cortex: evidence for multiple auditory areas. Neuroimage6, 288304.10.1006/nimg.1997.0304

  • 97

    RomanskiL. M.TianB.FritzJ.MishkinM.Goldman-RakicP. S.RauscheckerJ. P. (1999). Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nat. Neurosci.2, 11311136.

  • 98

    SabriM.BinderJ. R.DesaiR.MedlerD. A.LeitlM. D.LiebenthalE. (2008). Attentional and linguistic interactions in speech perception. Neuroimage39, 14441456.10.1016/j.neuroimage.2007.09.052

  • 99

    SchönwiesnerM.NovitskiN.PakarinenS.CarlsonS.TervaniemiM.NaatanenR. (2007). Heschl's gyrus, posterior superior temporal gyrus, and mid-ventrolateral prefrontal cortex have different roles in the detection of acoustic changes. J. Neurophysiol.97, 20752082.10.1152/jn.01083.2006

  • 100

    SchönwiesnerM.RubsamenR.von CramonD. Y. (2005). Hemispheric asymmetry for spectral and temporal processing in the human antero-lateral auditory belt cortex. Eur. J. Neurosci.22, 15211528.

  • 101

    SchönwiesnerM.ZatorreR. J. (2009). Spectro-temporal modulation transfer function of single voxels in the human auditory cortex measured with high-resolution fMRI. Proc. Natl. Acad. Sci. U.S.A.106, 1461114616.

  • 102

    SchwarzbauerC.DavisM. H.RoddJ. M.JohnsrudeI. (2006). Interleaved silent steady state (ISSS) imaging: a new sparse imaging method applied to auditory fMRI. Neuroimage29, 774782.10.1016/j.neuroimage.2005.08.025

  • 103

    ScottS. K.BlankC. C.RosenS.WiseR. J. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain123(Pt 12), 24002406.10.1093/brain/123.12.2400

  • 104

    ScottS. K.JohnsrudeI. S. (2003). The neuroanatomical and functional organization of speech perception. Trends Neurosci.26, 100107.10.1016/S0166-2236(02)00037-1

  • 105

    ShannonR. V.ZengF. G.KamathV.WygonskiJ.EkelidM. (1995). Speech recognition with primarily temporal cues. Science270, 303304.10.1126/science.270.5234.303

  • 106

    ShtyrovY.KujalaT.AhveninenJ.TervaniemiM.AlkuP.IlmoniemiR. J.NäätänenR. (1998). Background acoustic noise and the hemispheric lateralization of speech processing in the human brain: magnetic mismatch negativity study. Neurosci. Lett.251, 141144.

  • 107

    SpechtK.ReulJ. (2003). Functional segregation of the temporal lobes into highly differentiated subsystems for auditory perception: an auditory rapid event-related fMRI-task. Neuroimage20, 19441954.10.1016/j.neuroimage.2003.07.034

  • 108

    SteinbrinkC.AckermannH.LachmannT.RieckerA. (2009). Contribution of the anterior insula to temporal auditory processing deficits in developmental dyslexia. Hum. Brain Mapp.30, 24012411.10.1002/hbm.20674

  • 109

    StevensA. A. (2004). Dissociating the cortical basis of memory for voices, words and tones. Brain Res. Cogn. Brain Res.18, 162171.

  • 110

    SweetR. A.Dorph-PetersenK. A.LewisD. A. (2005). Mapping auditory core, lateral belt, and parabelt cortices in the human superior temporal gyrus. J. Comp. Neurol.491, 270289.

  • 111

    TervaniemiM.HugdahlK. (2003). Lateralization of auditory-cortex functions. Brain Res. Brain Res. Rev.43, 231246.

  • 112

    ThierryG.GiraudA. L.PriceC. (2003). Hemispheric dissociation in access to the human semantic system. Neuron38, 499506.10.1016/S0896-6273(03)00199-5

  • 113

    ThivardL.BelinP.ZilboviciusM.PolineJ. B.SamsonY. (2000). A cortical region sensitive to auditory spectral motion. Neuroreport11, 29692972.10.1097/00001756-200009110-00028

  • 114

    TianB.RauscheckerJ. P. (1994). Processing of frequency-modulated sounds in the cat's anterior auditory field. J. Neurophysiol.71, 19591975.

  • 115

    TramoM. J.ShahG. D.BraidaL. D. (2002). Functional role of auditory cortex in frequency processing and pitch perception. J. Neurophysiol.87, 122139.

  • 116

    TurkeltaubP. E.EdenG. F.JonesK. M.ZeffiroT. A. (2002). Meta-analysis of the functional neuroanatomy of single-word reading: method and validation. Neuroimage16(Pt 1), 765780.10.1006/nimg.2002.1131

  • 117

    UppenkampS.JohnsrudeI. S.NorrisD.Marslen-WilsonW.PattersonR. D. (2006). Locating the initial stages of speech-sound processing in human temporal cortex. Neuroimage31, 12841296.10.1016/j.neuroimage.2006.01.004

  • 118

    VouloumanosA.KiehlK. A.WerkerJ. F.LiddleP. F. (2001). Detection of sounds in the auditory stream: event-related fMRI evidence for differential activation to speech and nonspeech. J. Cogn. Neurosci.13, 9941005.

  • 119

    WagerT. D.LindquistM. A.NicholsT. E.KoberH.Van SnellenbergJ. X. (2009). Evaluating the consistency and specificity of neuroimaging data using meta-analysis. Neuroimage45(Suppl.1), S210S221.

  • 120

    WallaceM. N.JohnstonP. W.PalmerA. R. (2002). Histochemical identification of cortical areas in the auditory region of the human brain. Exp. Brain Res.143, 499508.10.1007/s00221-002-1014-z

  • 121

    WarrenJ. D.GriffithsT. D. (2003). Distinct mechanisms for processing spatial sequences and pitch sequences in the human auditory brain, J. Neurosci.23, 57995804.

  • 122

    WarrenJ. D.JenningsA. R.GriffithsT. D. (2005). Analysis of the spectral envelope of sounds by the human brain. Neuroimage24, 10521057.10.1016/j.neuroimage.2004.10.031

  • 123

    WarrierC.WongP.PenhuneV.ZatorreR.ParrishT.AbramsD.KrausN. (2009). Relating structure to function: Heschl's gyrus and acoustic processing. J. Neurosci.29, 6169.10.1523/JNEUROSCI.3489-08.2009

  • 124

    WarrierC. M.ZatorreR. J. (2002). Influence of tonal context and timbral variation on perception of pitch. Percept. Psychophys.64, 198207.

  • 125

    WarrierC. M.ZatorreR. J. (2004). Right temporal cortex is critical for utilization of melodic contextual cues in a pitch constancy task. Brain127(Pt 7), 16161625.10.1093/brain/awh183

  • 126

    WessingerC. M.VanMeterJ.TianB.Van LareJ.PekarJ.RauscheckerJ. P. (2001). Hierarchical organization of the human auditory cortex revealed by functional magnetic resonance imaging. J. Cogn. Neurosci.13, 17.

  • 127

    WiseR.CholletF.HadarU.FristonK.HoffnerE.FrackowiakR. (1991). Distribution of cortical networks involved in word comprehension and word retrieval. Brain114, 18031817.10.1093/brain/114.4.1803

  • 128

    WoodsD. L.AlainC. (2009). Functional imaging of human auditory cortex. Curr. Opin. Otolaryngol. Head Neck Surg.17, 407411.

  • 129

    ZaehleT.SchimidtC. F.MeyerM.BaumannS.BaltesC.BoesigerP.JanckeL. (2007). Comparison of “silent” clustered and sparse temporal fMRI acquisitions in tonal and speech perception tasks. Neuroimage37, 11951204.10.1016/j.neuroimage.2007.04.073

  • 130

    ZatorreR. J.BelinP. (2001). Spectral and temporal processing in human auditory cortex. Cereb. Cortex11, 946953.10.1093/cercor/11.10.946

  • 131

    ZatorreR. J.BelinP.PenhuneV. B. (2002). Structure and function of auditory cortex: music and speech. Trends Cogn. Sci.6, 3746.

  • 132

    ZatorreR. J.BouffardM.AhadP.BelinP. (2002). Where is ‘where’ in the human auditory cortex?Nat. Neurosci.5, 905909.

  • 133

    ZatorreR. J.EvansA. C.MeyerE. (1994). Neural mechanisms underlying melodic perception and memory for pitch. J. Neurosci.14, 19081919.

  • 134

    ZatorreR. J.EvansA. C.MeyerA.GjeddeA. (1992). Lateralization of phonetic and pitch discrimination in speech processing. Science256, 846849.10.1126/science.1589767

  • 135

    ZatorreR. J.MeyerE.GjeddeA.EvansA. C. (1996). PET studies of phonetic processing of speech: review, replication and reanalysis. Cereb. Cortex6, 2130.10.1093/cercor/6.1.21

  • 136

    ZevinJ.McCandlissB. (2005). Dishabituation of the BOLD response to speech sounds. Behav. Brain Funct.1, 4.

Appendix

Table A1

ReferenceTable/figureCategoryFrequency compositionChanges over timeStimuli of interestBaselineFociVocalIntelligible speechComplex non-speechMultiple, irregular change
Belin et al. (1998)Table 1, p. 537MultipleIrregularRapid formant transitionSilence3XX
MultipleIrregularExtended formant transitionSilence2X
Belin et al. (1999)Text, p. 422–423MultipleUnchangingHarmonic complex soundSilence4
Benson et al. (2001)Table 1, p. 372MusicNote/chord/chord progressionRest27
Table 2, p. 373VocalMultipleIrregularVowel/consonant–vowels/syllablesRest24X
Binder et al. (1996)Table 1, p. 1244VocalMultipleIrregularWordsTones sequence1X
Binder et al. (2000)Table 2, p. 518MultipleRegularSequence of tonesNoise11
VocalMultipleIrregularWordsSequence of tones8X
VocalMultipleIrregularPseudowordsSequence of tones8XX
VocalMultipleIrregularReversed wordsSequence of tones7XX
Binder et al. (2008)AppendixVocalMultipleIrregularWordsRest12X
VocalMultipleIrregularWordsTones2
Blood et al. (1999)Table 2, p. 384MusicMelodiesNoise3
MusicMelodiesNoise3
Brown et al. (2004)Table 1, p. 2035MusicMultipleIrregularMusicRest21
Burton et al. (2000)Table 2, p. 682VocalMultipleIrregularWordsTone3X
VocalMultipleIrregularWordsTone6X
Burton and Small (2006)Table 2, p. 647VocalMultipleIrregularWordsTone4X
Celsis et al. (1999)Table 1, p. 138ToneSingleUnchangingToneRest1
ToneSingleUnchangingToneRest1
VocalMultipleIrregularSyllableRest3X
ToneSingleUnchangingToneRest3
MultipleUnchangingSquare wave toneRest3
VocalMultipleIrregularSyllableRest3
Chen et al. (2008)Table 2, p. 2849MusicMultipleIrregularMusical rhythmsSilence9X
Dalla Barba et al. (1998)Table 1, p. 552VocalMultipleIrregularWordsRest5X
De Nil et al. (2008)Table 2, p. 119VocalMultipleIrregularWordsSilence1X
Démonet et al. (1992)Table 4, p. 1758VocalMultipleIrregularPhonemes (three or four syllables non-words)Tones7X
VocalMultipleIrregularWordsTones14X
Engelien et al. (2006)Table 1, p. 603MultipleIrregularMeaningful non-verbal soundsRest8XX
MultipleIrregularMeaningless non-verbal soundsRest14X
Fernandez-Espejo et al. (2008)Table 2, p. 887VocalMultipleIrregularNarrative sentencesSilence2X
Gaab et al. (2007a)Table 1, p. 710VocalMultipleIrregularWordsSilence7XX
VocalMultipleIrregularWordsSilence9X
Gaab et al. (2007b)Table 1, p. 727VocalMultipleIrregularWordsSilence5X
MultipleIrregularRecorded scanner noiseSilence2X
Giraud et al. (2004)Table 1, p 250VocalSpeech (sentences)Temporally matched complex noise5X
Griffiths et al. (1998)Table 1, p. 424MusicMelodiesSequence of tones4
Hall et al. (2000)Table 1, p. 114SingleRegularAmplitude and frequency modulated toneStatic tone2
Hall et al. (2002)Table 2, p. 144MultipleUnchangingHarmonic complex toneSingle tone5
SingleRegularFrequency modulated toneStatic tone10
Hart et al. (2003)Text, p. 778SingleRegularAmplitude and frequency modulated toneUnmodulated tone2
Hart et al. (2004)Table 1, p. 182ToneSingleUnchangingToneSilence4
SingleRegularFrequency modulated toneSilence6
Hertrich et al. (2010)Table S1VocalMultipleIrregularSyllablesSilence10
SingleRegularSingle formants sweepSilence8
Hugdahl et al. (2003)Table 3, p. 43VocalMultipleIrregularVowelSilence4
Table 4, p. 43VocalMultipleIrregularPseudowordsSilence3XX
Table 5, p. 43VocalMultipleIrregularWordsSilence2X
Hwang et al. (2007)Table 2, p. 289VocalMultipleIrregularSentencesQuiet7XX
Jäncke et al. (1998)Table 1, p. 878VocalMultipleIrregularSyllablesSilence9X
VocalMultipleIrregularSyllablesSilence9
VocalMultipleIrregularSyllablesSilence9
Table 2, p. 878MultipleRegularTone sequenceSilence4
MultipleRegularTone sequenceSilence6
MultipleRegularTone sequenceSilence6
Lillywhite et al. (2010)Table 1, p. 875VocalMultipleIrregularStory listeningWhite noise2XX
Müller et al. (1999)Table 3, p. 24MultipleRegularSequence of tonesRest6
VocalMultipleIrregularSentencesRest5XX
Mummery et al. (1999)Table 1, p. 452VocalSpeech (words)Signal correlated noise7X
Mutschler et al. (2010)Supp. MaterialMusicMultipleIrregularMelodiesRest94
O'Leary et al. (1996)Table 2, p. 27VocalMultipleIrregularWordsTone2X
MultipleIrregularEnvironmental soundsTone2X
Obleser et al. (2006)Table 2, p. 566VocalMultipleIrregularVowelBand-pass noise4X
Paulesu et al. (1995)Table 2, p. 667VocalMultipleIrregularWordsTones5X
Peretz et al. (2009)Table 1, p. 263MusicMelodiesRandom tone sequence8
Petersen et al. (1988)Table 2, p. 585VocalMultipleIrregularWordsSilence (fixation point)6X
Price et al. (1992)Figure 1, p. 180VocalMultipleIrregularNounsRest6XX
Reynolds Losin et al. (2009)Table 1, p. 374VocalMultipleIrregularStory listeningRest3X
Rimol et al. (2005)Table 1, p. 1063VocalMultipleIrregularConsonantRest8X
VocalMultipleIrregularSyllablesRest6
NoiseMultipleUnchangingNoise (white, brown, pink)Rest12
NoiseMultipleUnchangingNoise (white, brown, pink)Rest8
Table 2, p. 1064VocalMultipleIrregularConsonantNoise1
VocalMultipleIrregularSyllableNoise3
Schwarzbauer et al. (2006)Table 1, p. 779VocalSentencesSignal correlated noise12XX
VocalSentencesSignal correlated noise6
Schönwiesner et al. (2005)Text, p. 1523–1525SingleRegularParametric increase in temporal rate changeSilence2
MultipleUnchangingParametric increase in spectral componentsSilence2
Specht and Reul (2003)Table 1, p. 1950ToneSingleUnchangingTonesRest5
MultipleIrregularSounds of animal and instrumentsRest9XX
VocalMultipleIrregularWordsRest18XX
Table 2, p. 1951MultipleIrregularSounds of animal and instrumentsTones7X
VocalMultipleIrregularWordsTones12
Steinbrink et al. (2009)Table 1, p. 2406VocalMultipleIrregularIncrease rate of syllable presentationSilence7X
MultipleRegularIncrease rate of click sequenceSilence8
Stevens (2004)Table 1, p. 166VocalMultipleIrregularVoiceTone13X
VocalMultipleIrregularWordsTone5
Thivard et al. (2000)Table 1, p. 2971MultipleUnchangingStatic complex soundSilence6
MultipleIrregularComplex sound modulated (motion)Silence6X
SingleIrregularComplex sound modulated (motion)Static complex sound4
Vouloumanos et al. (2001)Table 1, p. 998VocalMultipleIrregularWordsTones8X
MultipleIrregularNon-speech complex sinewave soundsTones5XX
Warren and Griffiths (2003)Table 1, p. 5802MusicChanging pitch sequencesFixed pitch sequences4
Warrier et al. (2009)Table 2, p. 65SingleRegularParametric increase in temporal rate changeSilence6
MultipleUnchangingParametric increase in spectral componentsSilence6
Wise et al. (1991)Table 2, p. 1808VocalMultipleIrregularNounsRest6XX
Zaehle et al. (2007)Table 1, p. 1201VocalMultipleIrregularSentencesSilence2X
VocalMultipleIrregularSentencesSilence2
ToneSingleUnchangingTonesSilence2
ToneSingleUnchangingTonesSilence2
VocalMultipleIrregularSentencesTones2
VocalMultipleIrregularSentencesTones2
Table 2, p. 1201VocalMultipleIrregularSentencesSilence2
VocalMultipleIrregularSentencesSilence2
ToneSingleUnchangingTonesSilence2
ToneSingleUnchangingTonesSilence2
VocalMultipleIrregularSentencesTones2
VocalMultipleIrregularSentencesTones2
Zatorre et al. (1992)Table 2, p. 847NoiseMultipleUnchangingWhite noise burstsSilence8
VocalMultipleIrregularSyllablesNoise6X
Zatorre et al. (1994)Table 2, p. 1911MusicMelodiesNoise2
Zatorre et al. (1996)Table 6, p. 26VocalMultipleIrregularSyllablesNoise6X
Zatorre et al. (2002)Table 1, p. 907NoiseMultipleIrregularNoise (reversed environmental sounds)Silence3
Zatorre and Belin (2001)Table 1, p. 948SingleRegularTone sequence with increasing rate of presentationSilence2
MultipleUnchangingTone sequence with increasing number of spectral elementsSilence3
Zevin and McCandliss (2005)Table 1VocalMultipleIrregularSyllablesSilence8XX

Selection of foci for the subsamples: Vocal sounds (20 contrasts, 165 foci) Intelligible speech (27 contrasts, 166 foci), Complex non-speech (13 contrasts, 76 foci) and Multiple frequency, irregular change (10 contrasts, 70 foci).

Summary

Keywords

fMRI, category, time, frequency, hierarchy

Citation

Samson F, Zeffiro TA, Toussaint A and Belin P (2011) Stimulus Complexity and Categorical Effects in Human Auditory Cortex: An Activation Likelihood Estimation Meta-Analysis. Front. Psychology 1:241. doi: 10.3389/fpsyg.2010.00241

Received

16 September 2010

Accepted

23 December 2010

Published

17 January 2011

Volume

1 - 2010

Edited by

Josef P. Rauschecker, Georgetown University School of Medicine, USA

Reviewed by

Jonas Obleser, Max Planck Institute for Human Cognitive and Brain Sciences, Germany; Josef P. Rauschecker, Georgetown University School of Medicine, USA

Copyright

*Correspondence: Fabienne Samson, Hôpital Rivière-des-Prairies, Recherche TN, 7070 Boulevard Perras, Montréal, QC, Canada H1E 1A4. e-mail:

This article was submitted to Frontiers in Auditory Cognitive Neuroscience, a specialty of Frontiers in Psychology.

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics