GENERAL COMMENTARY article

Front. Hum. Neurosci., 03 June 2015

Sec. Speech and Language

Volume 9 - 2015 | https://doi.org/10.3389/fnhum.2015.00330

A Commentary on: “Neural overlap in processing music and speech”

  • RK

    Richard Kunert 1,2*

  • LR

    L. Robert Slevc 3

  • 1. Neurobiology of Language, Max Planck Institute for Psycholinguistics Nijmegen, Netherlands

  • 2. Neurobiology of Language, Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen Nijmegen, Netherlands

  • 3. Language and Music Cognition Lab, Department of Psychology, University of Maryland College Park, MD, USA

Evidence for neural overlap in processing music and speech?

There is growing interest in whether the brain networks responsive to music and language are separate after basic sensory processing or whether they share neural resources. Peretz et al.'s (2015) review on the available brain imaging evidence is a good moment to reflect on the field. We agree that “the question of overlap between music and speech processing must still be considered as an open question.” (p. 16) However, even though their review was not intended to be exhaustive, Peretz et al. (2015) have arguably focused too narrowly on neuroimaging results to give a fair assessment of current knowledge about music-language relationships.

Firstly though, it is worth re-iterating the limitations of neuroimaging studies. The fact that music experiments and language experiments reveal common brain regions (e.g., Koelsch et al., 2002; Herdener et al., 2014) is insufficient evidence for shared neural circuitry, as domain-specific neural populations might be intermingled within the same brain regions (especially given the resolution of noninvasive brain-imaging techniques). Similarly, different cognitive processes might underlie common activation sites, especially in pre-frontal areas. As just one example, attending to music over scanner noise might draw particularly strongly on prefrontal mechanisms of focused attention, compared to language perception, which might be more robust (especially in non-musicians). Therefore, Peretz et al. (2015) propose more sophisticated methods such as multivariate pattern analysis (MVPA) and adaptation paradigms. However, even these methods give equivocal interpretations: different patterns of activation in common brain areas (as revealed by MVPA) might reflect separate music-or-language neural populations within the same region (Rogalsky et al., 2011) or indicate the same neural population reacting differently to music and language (Abrams et al., 2011) possibly due to changes in functional connectivity. And while fMRI adaptation paradigms hold promise, it remains to be seen how they can be applied to this question (for two very different attempts see Steinbeis and Koelsch, 2008a; Sammler et al., 2010). Thus, the current brain imaging literature is indeed equivocal. However, looking beyond fMRI can be beneficial.

Beyond fMRI: the interference paradigm in brain and behavior

Although Peretz et al. (

2015

) nicely describe the current state and limitations of functional neuroimaging evidence on music-language overlap, they ignore a large body of behavioral and electrophysiological evidence for interactive processes

1

. Much of this work relies on interference paradigms, for example, Slevc et al. (

2009

) asked participants to read

garden path

sentences like the following, segment by segment (while measuring reading time as a proxy for processing cost):

  • After | the trial | the attorney | advised | the defendant | was | likely | to commit | more crimes.

  • After | the trial | the attorney | advised

    that

    | the defendant | was | likely | to commit | more crimes.

Resolving the temporary syntactic ambiguity in (a), where “defendant” is initially misinterpreted as a direct object, causes slower reading of “was” than in (b), where “that” signals the correct interpretation. This syntactic garden path effect was augmented when hearing a task-irrelevant, harmonically unexpected chord during the reading of “was” (compared to a harmonically expected chord). This is unlikely to be due to the chord's acoustic unexpectancy, since a timbrally unexpected chord (i.e., new instrument) had no such effect. Slevc et al. (2009) interpreted their result as evidence for shared music-language resources which process structural relations. When these resources are taxed by a harmonically unexpected chord, they sub-optimally process challenging syntactic relations as in (a). See Table 1 for similar studies.

Table 1

ReferencesPrimary outcome
BEHAVIOR
Fedorenko et al., 2009Melodic unexpectancy worsens the comprehension of syntactically complex sentences, volume unexpectancy without effect
Slevc et al., 2009Harmonic unexpectancy slows the resolution of syntactic ambiguities but not of semantically unexpected words; timbre without effect
Hoch et al., 2011Harmonic unexpectancy slows the word judgment time of syntactically unexpected words, but not of semantically unexpected words
Perruchet and Poulin-Charronnat, 2013Harmonic unexpectancy slows the resolution of semantic ambiguities but not of semantically unexpected words
Fiveash and Pammer, 2014Harmonic unexpectancy worsens sentence recall but not word list recall; timbral unexpectancy without effect
ELECTRO ENCEPHALOGRAPHY (EEG)
Besson et al., 1998Melodic unexpectancy does not affect the event-related potential (ERP) to a semantic manipulation (N400)
Koelsch et al., 2005Harmonic unexpectancy affects the syntax-related left anterior negativity (LAN) but not the N400
Steinbeis and Koelsch, 2008bHarmonic unexpectancy affects the LAN but not the N400; language syntactic violations affect the harmony-related early right anterior negativity (ERAN) while language semantic anomalies affect the harmony-related N500
Carrus et al., 2011Harmonic unexpectancy affects the oscillatory response to language syntax (delta-theta bands), but not vice versa; no interaction with semantics
Carrus et al., 2013Melodic unexpectancy affects the LAN but not the N400

Overview of ten representative music-language interference studies.

Behavioral and electrophysiological interference studies offer compelling evidence for shared musico-linguistic resources but were not discussed by Peretz et al. (2015). For illustration, we list five behavioral and five electro-encephalographical studies.

These interference effects are compelling evidence for shared resources. While the aforementioned fMRI paradigms investigate whether shared neural circuitry is extensive enough to be visible in fMRI, studies like those in Table 1 investigate the functional relevance of shared resources (e.g., in terms of behavioral outcomes). Given the support for the latter, an important debate has centered on the functional role of shared resources, such as involvement in structural processing (Patel, 2003), general attention (e.g., Perruchet and Poulin-Charronnat, 2013), or cognitive control (Slevc and Okada, 2015). This debate would surely benefit from a variety of approaches which reveal the time-course, oscillatory, and network dynamics (e.g., via electrophysiological measures of brain activity), as well as the causal role of associated brain areas (e.g., via transcranial magnetic stimulation). Targeted fMRI studies informed by the entirety of the neural as well as the behavioral literature are needed to complement these approaches.

Toward an inter-disciplinary science of music and language processing

Peretz et al. (2015) are certainly right when they write that “converging evidence from several methodologies is needed.” We have tried to sketch the impressive extent of the evidence that is already available. However, there are still open questions. For example, the interference paradigm has so far not been used with linguistic processes beyond syntax and semantics (e.g., phonology, morphology, and prosody) and musical processes beyond melody, harmony, and timbre (e.g., rhythm).

Greater insights into music and language offer great potential for example in terms of clinical applications. Specifically, syntactic processing problems found in Broca's aphasia (see Patel et al., 2008) and specific language impairment (Jentschke et al., 2008) could be helped by melody-harmony interventions given evidence for shared resources for syntax and harmony, see Table 1. Progress with such clinical applications requires us first to understand how music and language relate to each other. This understanding can only emerge when going beyond a focus on any one method and, instead, viewing the field as an inter-disciplinary challenge.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Statements

Acknowledgments

This work was supported by a PhD grant from the Max Planck Society to RK and a grant from the GRAMMY Foundation to LRS.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Footnotes

1.^Peretz et al. (2015) focus on music and speech, not language as we do here. However, the former has also been investigated with the interference paradigm with stimuli sung a cappella (Besson et al., 1998; Fedorenko et al., 2009).

References

  • 1

    AbramsD. A.BhataraA.RyaliS.BalabanE.LevitinD. J.MenonV. (2011). Decoding temporal structure in music and speech relies on shared brain resources but elicits different fine-scale spatial patterns. Cereb. Cortex21, 15071518. 10.1093/cercor/bhq198

  • 2

    BessonM.FaïtaF.PeretzI.BonnelA.-M.RequinJ. (1998). Singing in the brain: independence of lyrics and tunes. Psychol. Sci. 9, 494498. 10.1111/1467-9280.00091

  • 3

    CarrusE.KoelschS.BhattacharyaJ. (2011). Shadows of music–language interaction on low frequency brain oscillatory patterns. Brain Lang. 119, 5057. 10.1016/j.bandl.2011.05.009

  • 4

    CarrusE.PearceM. T.BhattacharyaJ. (2013). Melodic pitch expectation interacts with neural responses to syntactic but not semantic violations. Cortex49, 21862200. 10.1016/j.cortex.2012.08.024

  • 5

    FedorenkoE.PatelA.CasasantoD.WinawerJ.GibsonE. (2009). Structural integration in language and music: evidence for a shared system. Mem. Cogn. 37, 19. 10.3758/MC.37.1.1

  • 6

    FiveashA.PammerK. (2014). Music and language: do they draw on similar syntactic working memory resources?Psychol. Music42, 190209. 10.1177/0305735612463949

  • 7

    HerdenerM.HumbelT.EspositoF.HabermeyerB.Cattapan-LudewigK.SeifritzE. (2014). Jazz drummers recruit language-specific areas for the processing of rhythmic structure. Cereb. Cortex24, 836843. 10.1093/cercor/bhs367

  • 8

    HochL.Poulin-CharronnatB.TillmannB. (2011). The influence of task-irrelevant music on language processing: syntactic and semantic structures. Front. Psychol. 2:112. 10.3389/fpsyg.2011.00112

  • 9

    JentschkeS.KoelschS.SallatS.FriedericiA. D. (2008). Children with specific language impairment also show impairment of music-syntactic processing. J. Cogn. Neurosci. 20, 19401951. 10.1162/jocn.2008.20135

  • 10

    KoelschS.GunterT. C.von CramonD. Y.ZyssetS.LohmannG.FriedericiA. D. (2002). Bach speaks: a cortical “language-network” serves the processing of music. Neuroimage17, 956966. 10.1016/S1053-8119(02)91154-7

  • 11

    KoelschS.GunterT. C.WittfothM.SammlerD. (2005). Interaction between syntax processing in language and in music: an ERP study. J. Cogn. Neurosci. 17, 15651577. 10.1162/089892905774597290

  • 12

    PatelA. D. (2003). Music, Language, and the Brain. Oxford: Oxford University Press.

  • 13

    PatelA. D.IversenJ. R.WassenaarM.HagoortP. (2008). Musical syntactic processing in agrammatic Broca's aphasia. Aphasiology22, 776789. 10.1080/02687030701803804

  • 14

    PeretzI.VuvanD.LagroisM.-É.ArmonyJ. L. (2015). Neural overlap in processing music and speech. Philol. Trans. R. Soc. Lond. B Biol. Sci. 370:20140090. 10.1098/rstb.2014.0090

  • 15

    PerruchetP.Poulin-CharronnatB. (2013). Challenging prior evidence for a shared syntactic processor for language and music. Psychon. Bull. Rev. 20, 310317. 10.3758/s13423-012-0344-5

  • 16

    RogalskyC.RongF.SaberiK.HickokG. (2011). Functional anatomy of language and music perception: temporal and structural factors investigated using functional magnetic resonance imaging. J. Neurosci. 31, 38433852. 10.1523/JNEUROSCI.4515-10.2011

  • 17

    SammlerD.BairdA.ValabrègueR.ClémentS.DupontS.BelinP.et al. (2010). The relationship of lyrics and tunes in the processing of unfamiliar songs: a functional magnetic resonance adaptation study. J. Neurosci. 30, 35723578. 10.1523/JNEUROSCI.2751-09.2010

  • 18

    SlevcL. R.OkadaB. M. (2015). Processing structure in language and music: a case for shared reliance on cognitive control. Psychon. Bull. Rev. 22, 637652. 10.3758/s13423-014-0712-4

  • 19

    SlevcL. R.RosenbergJ. C.PatelA. D. (2009). Making psycholinguistics musical: self-paced reading time evidence for shared processing of linguistic and musical syntax. Psychon. Bull. Rev. 16, 374381. 10.3758/16.2.374

  • 20

    SteinbeisN.KoelschS. (2008a). Comparing the processing of music and language meaning using EEG and fMRI provides evidence for similar and distinct neural representations. PLoS ONE3:e2226. 10.1371/journal.pone.0002226

  • 21

    SteinbeisN.KoelschS. (2008b). Shared neural resources between music and language indicate semantic processing of musical tension-resolution patterns. Cereb. Cortex18, 11691178. 10.1093/cercor/bhm149

Summary

Keywords

neural overlap, music, harmony, speech, language

Citation

Kunert R and Slevc LR (2015) A Commentary on: “Neural overlap in processing music and speech”. Front. Hum. Neurosci. 9:330. doi: 10.3389/fnhum.2015.00330

Received

20 February 2015

Accepted

22 May 2015

Published

03 June 2015

Volume

9 - 2015

Edited by

Lutz Jäncke, University of Zurich, Switzerland

Reviewed by

Eckart Altenmüller, Hannover University of Music, Drama and Media, Germany; Stefan Elmer, University of Zurich, Switzerland

Copyright

*Correspondence: Richard Kunert,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics