OPINION article

Front. Psychol., 28 June 2018

Sec. Psychology of Language

Volume 9 - 2018 | https://doi.org/10.3389/fpsyg.2018.01109

Why We Should Study Multimodal Language

  • School of Humanities, University of Brighton, Brighton, United Kingdom

Article metrics

View details

168

Citations

26,6k

Views

4,5k

Downloads

What do we study when we study language? Our theories of language, and particularly our theories of the cognitive and neural underpinnings of language, have developed primarily from the investigation of spoken language. Moreover, spoken language has been studied primarily as a unichannel phenomenon, i.e., as just speech or text. However, contexts of face-to-face interaction form the primary ecological niche of language, both spoken and signed, and the primary contexts in which language is used, is learned and has evolved (Levinson and Enfield, 2006; Vigliocco et al., 2014). In such contexts, a multitude of cues, both vocal and visual, contribute to utterance construction. Should we thus not turn our attention to the study of language as multimodal language? The position that language can be appropriately studied as just speech or text essentially aligns with a conception of language based on Chomsky's competence or Saussure's langue: it is the linguistic code and the formalization of phonological, morphological, and syntactic structure that is of interest. Even functional, usage-based theories of language, which see linguistic structure as shaped by language use and the function of language in cultural and communicative contexts (e.g., Fillmore, 1982; Givón, 1984; Goldberg, 1995), have focused on the linguistic code and have thus also mainly regarded language as speech or text (but see e.g., Tomasello, 1999; Diessel, 2006). The argument put forward here is that we should study language as its multimodal manifestation in contexts of face-to-face interaction. As such, our object of study should subsume information expressed in both vocal and visual channels, including prosody, gesture, facial expression, body movement, which invariably accompany linguistic expression in face-to-face contexts.

The thought experiment proposed by Vigliocco et al. (2014) offers a window onto this approach by asking: What if the study of language had started with the study of signed language rather than spoken language? If the study of language had started with signed language, the multichannel/multimodal nature of language would have stood center stage from the beginning. Questions that have become matters of serious inquiry and debate only recently, in particular concerning the status and interplay of iconicity and arbitrariness (Perniss et al., 2010; Perniss and Vigliocco, 2014; Dingemanse et al., 2015) and gradience and categoricity (see Goldin-Meadow and Brentari, 2017 and peer commentary, e.g., Occhino and Wilcox, 2017, for review) in language, may have been discussed earlier and answered in different ways. This brings to the fore the relevance of thinking about language in a more unified way: encompassing spoken and signed language; considering multiple channels of expression; and conceptualizing language with respect to its communicative functions.

What have been considered to be non-linguistic aspects of communication—including gesture, facial expression, body movement—have largely been studied separately from language proper. Multimodality studies, for example, are often framed as offering analyses of social interaction, studying something that is around language, but not studying language as such (see Mondada, 2016 for an overview). Pioneering scholars in the field of gesture studies have long advocated for a conception of gesture that is part and parcel of language (McNeill, 1985, 1992; Kendon, 2004). Nevertheless, this conception has not been adopted on a large scale. In advocating for a multimodal conception of “language,” it is important to bear in mind the extent to which our objects of study are constructed by an interplay of the present state of theory, technology and discourse (Kuhn, 1962; Foucault, 1972). This point is made by McNeill (1985: 350) when he writes that the division between speech and gesture (or “body language”) is “a cultural artifact, an arbitrary limitation derived from a particular historical evolution”—they are studied separately, though McNeill considers them to be “parts of a single psychological structure.” The conception that “language” is that which is linguistic, while communication is something different—essentially, the Saussurean and Chomskyan heritage—is not given by necessity. As such, it is time to reconceptualize our object of study and to usher in a new paradigm of language theory, a paradigm that focuses on multimodal language, that aligns with the real world use of language and focuses on doing language (Andresen, 2014; Kendon, 2014).

The study of sign language and gesture, as communicative expression in the visual modality, has been instrumental in widening the lens of investigation regarding the question of our object of study when we study language. Signed language highlights the fundamental multimodality and semiotic diversity of language. Moreover, the study of sign language, and its comparisons with speech and/or gesture, has highlighted the difficulties of maintaining a principled distinction between the linguistic and non-linguistic, and shown the need for developing analyses that admit a combination of categorical (considered linguistic) and gradient (considered non-linguistic) aspects of language (Liddell, 2003; Johnston, 2013; Kendon, 2014; Vigliocco et al., 2014; Goldin-Meadow and Brentari, 2017). Similarly, gesture and multimodality research has shown that, like signers, speakers make use of a wide range of semiotic resources, combining vocal and visible action in meaning making and utterance construction (e.g., Kendon, 2004; Mondada, 2016). The study of sign and gesture expose our current models of language as too narrowly conceived. The new paradigm for the study of language must acknowledge a range of semiotic practices (exhibiting iconicity, arbitrariness, gradience, categoricity) as fundamental to and constitutive of communicative expression. Below, I outline developments in contemporary research that further attest to the need for incorporating multimodality into our theories of language.

The neuroscientific investigation of language processing is one area in which the distinction between “language” and “communication,” and between “linguistic” and “non-linguistic” elements has been undermined. Recent research has been unable to find strong evidence supporting this distinction in language use. In addition, there is evidence that the brain does not privilege linguistic information in processing. Rather all kinds of context, including multimodal cues, are processed simultaneously and immediately (Hagoort and van Berkum, 2007). Numerous studies have provided evidence for similar processing of gesture and speech in terms of semantic and temporal integration (Özyürek et al., 2007; Hubbard et al., 2009; Straube et al., 2009; Habets et al., 2011; Dick et al., 2014; Yang et al., 2015; Peeters et al., 2017), as well as in terms of perceiving conventionalized meaning (Andric et al., 2013; Wolf et al., 2017). In addition, there is evidence that prosodic information from visual and vocal channels is treated similarly by the brain, with gestural beats functioning as visual prosody complementary to speech prosody (Biau et al., 2016). Studies also suggest that the use of different cues from context, including co-speech gesture (Skipper, 2014; Weisberg et al., 2017) and visible mouth movements (van Wassenhove et al., 2005), may speed up processing, aiding interpretation through improved prediction, and requiring less allocation of neural resources and thus conserving metabolic resources. Similar processing of semantically meaningful information, regardless of the modality of presentation has, crucially, also been shown for processing of signed and spoken language (MacSweeney et al., 2004) as well as for integration of pictures with sentence context (Willems et al., 2008). Thus, recent evidence from neuroimaging studies does not support a principled divide between linguistic and non-linguistic elements as the legacy of studying language as competence or langue presupposes. Instead, the evidence suggests that the brain is specially attuned to doing language or languaging (Andresen, 2014; Kendon, 2014).

Additional evidence supporting a multimodal view of language comes from recent research that suggests that what has traditionally been considered to be non-linguistic may in fact be subsumable under grammar and susceptible of grammatical description. Floyd (2016), describing the obligatory incorporation of celestial pointing gestures for time-of-day reference, discusses the possibility of modality hybrid grammars, which would incorporate gestural forms into the grammar. Recent work by Schlenker and Chemla (2017), aims to provide evidence for the grammar-like nature of gestures. Similarly, Ginzburg and Poesio (2016) offer a formalization of intrinsically interactional aspects of language, including gestures as well as disfluencies and non-sentential utterances, with the goal of demonstrating their grammatical, rule-governed behavior. This resonates with work by gesture researchers who have sought to define multimodal approaches to grammar (e.g., Mittelberg, 2006; Fricke, 2012), and who have studied aspects of conventionality in gesture, identifying varying degrees of conventionality in form-meaning pairings in gesture, used consistently across speakers within language communities for conveying certain meanings (e.g., Kendon, 1995, 2004; Calbris, 2011; Bressem and Müller, 2017; Bressem et al., 2017; Müller, 2017). Similarly, elements in the vocal modality not traditionally considered to be linguistic have been found to exhibit systematic behavior in terms of discursive and interactional function, e.g., research on the use of clicks and percussives (Wright, 2011; Ogden, 2013) and “filled pauses” like uh and um (Clark and Fox Tree, 2002).

Technological advances in experimental paradigms, data collection and analysis further motivate the need for a new paradigm in the study of language. The need for experimental control has meant that ecological validity, and the study of language in more real-world settings, has often been sacrificed (Hasson and Honey, 2012). Experimental limitations in the past have thus constrained researchers to the study of certain aspects of language. These aspects have happened to align with a langue/competence-type object of study, best represented as individual words (spoken or written lexemes) and combinations of words (spoken or written sentences). “Non-linguistic” elements, e.g., gradient and iconic elements which naturally occur in parallel and simultaneously with the abstractable, formal linguistic elements, were excluded from study (Tromp et al., 2017). In addition, the wider so-called extra-linguistic context, given by the environment—full of visual and acoustic cues—in which language typically occurs was likewise excluded from study (Knoeferle, 2015). However, new methodologies, and in particular, combinations of methodologies (e.g., Virtual Reality environments with ERP, Tromp et al., 2017; eye-tracking with ERP, Knoeferle, 2015) can improve the interpretation of data from a single methodology. Overall, the development of these technologies will support the construction of multimodal language (in the active sense of doing language) as the new object of study, which more resembles real-world use of language, rather than being restricted to just one aspect of it (Kendon, 2009). These technologies will allow investigation of the use and processing of language in more ecologically valid, contextually rich and communicatively real-world settings.

Renewed interest in the evolutionary origins of language also points toward a focus on the multimodality of language. One question that has dominated the discourse on theories of language evolution concerns the modality of early communication. Adherents to the “gesture-first” theory of language (e.g., Corballis, 2002, 2017; Tomasello, 2008; Arbib, 2012) claim that symbolic communication originated in the visual-manual modality, and that there was, over time, a transition to the vocal channel as the main carrier of linguistic function. However, eminent gesture researchers like McNeill (1992, 2012) and Kendon (2009, 2017) have claimed that expression in the vocal and visual modalities must have characterized communication from the very start (see also Perlman, 2017). The explanation of a “switch” from the visual to the vocal modality is difficult to motivate, and the tight semantic and temporal orchestration of multiple channels of expression and semiotic resources observable today (from corpus to neuroimaging studies) suggests that utterance construction has always shown this entanglement of modes. In addition, the evidence supporting tight hand-mouth coordination and links between kinesis (e.g., grip) and vocalization (Gentilucci et al., 2001; Kendon, 2009; Vainio et al., 2013) further support a view that gives the “speech-kinesis ensemble” (Kendon, 2009) pride of place in the phylogenetic evolution of language. Interesting perspectives for the interplay of visual and vocal communication supporting language emergence ab initio comes from comparative psychology and animal cognition (Leavens, 2003; Gillespie-Lynch et al., 2014) and from the suggestion by Larsson (2015) that the sounds of tool use and locomotion may have contributed to language evolution in a similar way as visible action and motion. Taking “multimodal language” as our object of study would allow a straightforward reconciliation of such findings.

Finally, developments in the fields of multilingualism research and language documentation offer illustrative guides to the changes that need to be generalized in language theory more broadly. The field of multilingualism research has recently been transformed through the notion of translanguaging. Researchers no longer conceive of code-switching or even code-mixing as an adequate account of the language behavior of bi-/multilingual speakers (Li, 2017). Bi-/multilingual speakers do not switch between or mix different “codes,” as formal systems of language. Rather, they engage in flexible use of diverse semiotic repertoires. Kusters et al. (2017) note that in translanguaging studies, researchers focus on multilingual communication, but without paying attention to multimodal communicative resources; while in multimodality studies, researchers do not attend to multilingual communication. Given the parallels with respect to the focus on a diverse semiotic repertoire and dynamic language practice, Kusters et al. (2017) note the benefits of bringing the fields together, and suggest that the language practices of signers can offer unique insight into the use and negotiation of both multimodal and multilingual repertoires.

Many linguists, especially those studying endangered languages, have adopted practices consistent with the linguistic subdiscipline of language documentation (Himmelmann, 2006). The goal of language documentation goes beyond the production of a (written) grammar of a language. Rather, the goal is documentation of language use and practice in order to create a “lasting, multipurpose record of a language” (Himmelmann, 2006, p. 1). Technological advances have been a boon here as well. Language documentation demands video-recordings of language use on as broad a scale as possible, including different varieties of use, domains of use, and social interaction. This necessarily includes the multimodality of language, and attention to multichannel and semiotically diverse modes of communication. The recognition that the majority of the world is multilingual is also important here, in that it points to the inadequacy of characterizing knowledge of language as residing in an idealized, monolingual speaker in a homogenous language community (Chomsky, 1965). Ansaldo (2010, p. 622) suggests that lessons from monolingual language use and transmission may represent such “exotic communicative ecologies in the history of human language evolution [that] the lessons derived from their study, albeit significant, could well end up being potentially exceptional, maybe even peripheral to the construction of general theories of language.”

Similarly, our models of language need to be based on ecologically valid contexts of multimodal language use (contexts of doing language)—and not on the “exotic communicative ecologies” represented by just speech or text. The development of our hitherto dominant models of language has been based on only a part of language, the abstractable, linguistic part best exemplified by written form (McNeill, 1985). A multimodal language model includes the full complement of fundamental modes of communication, including depiction, description, and indexing (Clark, 1996, 2016), and the wider context in which utterances are constructed and interpreted (Kendon, 2014; Vigliocco et al., 2014; Knoeferle, 2015). In various and interconnected ways, the studies reviewed above suggest that we are already on the threshold of a new paradigm. They point to the large range of elements, both vocal and visual, that contribute in systematic ways to language use and communicative expression and which we should not exclude a priori from the study of language (See Andrén (2014) for discussion of the nature of the problem of delineating the “lower limit of gesture”—the problem of drawing a line between what aspects of “visible action as utterance” Kendon (2004) to include or exclude from study.). We must remind ourselves that science often progresses precisely through a redefinition of the object of study. By redefining the nature and parameters of our concept of “language” we will be capable of forging this new paradigm adequate to a unified conception of language as communication, and basing our theories of language on language as a multimodal phenomenon.

Statements

Author contributions

The author confirms being the sole contributor of this work and approved it for publication.

Acknowledgments

I thank the editor and reviewers for helpful comments on an earlier version of the article. I thank the School of Humanities, University of Brighton for providing the funds to cover open access publishing fees.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    AndrénM. (2014). On the lower limit of gesture, in From Gesture in Conversation to Visible Action as Utterance, eds SeyfeddinipurM.GullbergM. (Amsterdam: John Benjamins), 153174.

  • 2

    AndresenJ. (2014). Linguistics and Evolution: A Developmental Approach. Cambridge: Cambridge University Press.

  • 3

    AndricM.SolodkinA.BuccinoG.Goldin-MeadowS.RizzolattiG.SmallS. (2013). Brain function overlaps when people observe emblems, speech, and grasping. Neuropsychologia51, 16191629. 10.1016/j.neuropsychologia.2013.03.022

  • 4

    AnsaldoU. (2010). Identity alignment and language creation in multilingual communities. Lang. Sci. 32, 615623. 10.1016/j.langsci.2010.08.003

  • 5

    ArbibM. (2012). How the Brain Got Language: The Mirror Neuron Hypothesis. Oxford: Oxford University Press.

  • 6

    BiauE.Morís FernándezL.HolleH.AvilaC.Soto-FaracoS. (2016). Hand gestures as visual prosody: BOLD responses to audio-visual alignment are modulated by the communicative nature of the stimuli. Neuroimage132, 129137. 10.1016/j.neuroimage.2016.02.018

  • 7

    BressemJ.MüllerC. (2017). The “Negative-Assessment-Construction” – a multimodal pattern based on a recurrent gesture?Ling. Vang. 3, 19. 10.1515/lingvan-2016-0053

  • 8

    BressemJ.SteinN.WegenerC. (2017). Multimodal language use in Savosavo: refusing, excluding and negating with speech and gesture. Pragmatics2, 173206. 10.1075/prag.27.2.01bre

  • 9

    CalbrisG. (2011). Elements of Meaning in Gesture. Amsterdam: John Benjamins.

  • 10

    ChomskyN. (1965). Aspects of the Theory of Syntax. Boston, MA: MIT Press.

  • 11

    ClarkH. (1996). Using Language. Cambridge: Cambridge University Press.

  • 12

    ClarkH. (2016). Depicting as a method of communication. Psychol. Rev. 3, 324347. 10.1037/rev0000026

  • 13

    ClarkH. H.Fox TreeJ. E. (2002). Using uh and um in spontaneous speaking. Cognition84, 73111. 10.1016/S0010-0277(02)00017-3

  • 14

    CorballisM. (2002). From Hand to Mouth: The Origins of Language. Princeton, NJ: Princeton University Press.

  • 15

    CorballisM. (2017). A word in the hand: the gestural origins of language, in Neural Mechanisms of Language. Innovations in Cognitive Neuroscience, ed ModyM. (Boston, MA: Springer), 199218.

  • 16

    DickA. S.MokE. H.Raja BeharelleA.Goldin-MeadowS.SmallS.L. (2014). Frontal and temporal contributions to understanding the iconic co-speech gestures that accompany speech. Hum. Brain Mapp. 35, 900917. 10.1002/hbm.22222

  • 17

    DiesselH. (2006). Demonstratives, joint attention, and the emergence of grammar. Cogn. Ling. 17, 463489. 10.1515/COG.2006.015

  • 18

    DingemanseM.BlasiD. E.LupyanG.ChristiansenM. H.MonaghanP. (2015). Arbitrariness, iconicity, and systematicity in language. Trends Cogn. Sci.19, 603615. 10.1016/j.tics.2015.07.013

  • 19

    FillmoreC. (1982). Frame semantics, in Cognitive Linguistics: Basic Readings, ed GeeraertsD. (Berlin: Mouton), 373400.

  • 20

    FloydS. (2016). Modality hybrid grammar? Celestial pointing for time-of-day reference in Nheengatú. Language92, 3164. 10.1353/lan.2016.0013

  • 21

    FoucaultM. (1972). The Archaeology of Knowledge.London: Routledge.

  • 22

    FrickeE. (2012). Grammatik multimodal: Wie Wörter und Gesten Zusammenwirken. Berlin: De Gruyter.

  • 23

    GentilucciM.BenuzziF.GangitanoM.GrimaldiS. (2001). Grasp with hand and mouth: a kinematic study on healthy subjects. J. Neurophys. 86, 16851699. 10.1152/jn.2001.86.4.1685

  • 24

    Gillespie-LynchK.GreenfieldP. M.LynH.Savage-RumbaughS. (2014). Gestural and symbolic development among apes and humans: support for a multimodal theory of language evolution. Front. Psychol.5:1228. 10.3389/fpsyg.2014.01228

  • 25

    GinzburgJ.PoesioM. (2016). Grammar is a system that characterizes talk in interaction. Front. Psychol.7:1938. 10.3389/fpsyg.2016.01938

  • 26

    GivónT. (1984). Syntax: A Functional-Typological Introduction. Vol. I. Amsterdam: John Benjamins.

  • 27

    GoldbergA. (1995). Constructions: A Construction Grammar Approach to Argument Structure. Chicago, IL: Chicago University Press.

  • 28

    Goldin-MeadowS.BrentariD. (2017). Gesture, sign, and language: the coming of age of sign language and gesture studies. Behav. Brain Sci. 40:e46. 10.1017/S0140525X15001247

  • 29

    HabetsB.KitaS.ShaoZ.ÖzyürekA.HagoortP. (2011). The role of synchrony and ambiguity in speech–gesture integration during comprehension. J. Cogn. Neurosci. 23, 18451854. 10.1162/jocn.2010.21462

  • 30

    HagoortP.van BerkumJ. (2007). Beyond the sentence given. Philos. Trans. R. Soc. B362, 801811. 10.1098/rstb.2007.2089

  • 31

    HassonU.HoneyC. (2012). Future trends in neuroimaging: neural processes as expressed within real-life contexts. Neuroimage62, 12721278. 10.1016/j.neuroimage.2012.02.004

  • 32

    HimmelmannN. (2006). Language documentation: what is it and what is it good for? in Essentials of Language Documentation, eds GippertJ.HimmelmannN.MoselU. (Berlin: Mouton), 130.

  • 33

    HubbardA.WilsonS.CallanD.DaprettoM. (2009). Giving speech a hand: gesture modulates activity in auditory cortex during speech perception. Hum. Brain Mapp. 30, 10281037. 10.1002/hbm.20565

  • 34

    JohnstonT. (2013). Towards a comparative semiotics of pointing actions in signed and spoken languages. Gesture13, 109142. 10.1075/gest.13.2.01joh

  • 35

    KendonA. (1995). Gestures as illocutionary and discourse structure markers in Southern Italian conversation. J. Pragm. 23, 247279. 10.1016/0378-2166(94)00037-F

  • 36

    KendonA. (2004). Gesture: Visible Action as Utterance. Cambridge: Cambridge University Press.

  • 37

    KendonA. (2009). Language's matrix. Gesture9, 355372. 10.1075/gest.9.3.05ken

  • 38

    KendonA. (2014). Semiotic diversity in utterance production and the concept of ‘language’. Philos. Trans. R. Soc. Lond. B Biol. Sci.369:20130293. 10.1098/rstb.2013.0293

  • 39

    KendonA. (2017). Reflections on the “gesture-first” hypothesis of language origins. Psychon. Bull. Rev. 24, 163170. 10.3758/s13423-016-1117-3

  • 40

    KnoeferleP. (2015). Language comprehension in rich non-linguistic contexts: combining eye tracking and event-related brain potentials, in Cognitive Neuroscience of Natural Language Use, ed WillemsR. M. (Cambridge: Cambridge University Press), 77100.

  • 41

    KuhnT. (1962). The Structure of Scientific Revolutions.Chicago, IL: Univ. Chic. Press.

  • 42

    KustersA.SpottiM.SwanwickR.TapioE. (2017). Beyond languages, beyond modalities: transforming the study of semiotic repertoires. Int. J. Multiling. 14, 219232. 10.1080/14790718.2017.1321651

  • 43

    LarssonM. (2015). Tool-use-associated sound in the evolution of language. Anim. Cogn. 18, 9931005. 10.1007/s10071-015-0885-x

  • 44

    LeavensD. (2003). Integration of visual and vocal communication: evidence for Miocene origins. Behav.Brain Sci. 26, 232233. 10.1017/S0140525X03480060

  • 45

    LevinsonS.EnfieldN. (2006). The Roots of Human Sociality: Culture, Cognition and Interaction. London: Bloomsbury.

  • 46

    LiW. (2017). Translanguaging as a practical theory of language. Appl. Ling. 39, 930. 10.1093/applin/amx039

  • 47

    LiddellS. (2003). Grammar, Gesture and Meaning in American Sign Language.Cambridge: Cambridge University Press.

  • 48

    MacSweeneyM.CampbellR.WollB.GiampietroV.DavidA. S.McGuireP.et al. (2004). Dissociating linguistic and nonlinguistic gestural communication in the brain. Neuroimage22, 16051618. 10.1016/j.neuroimage.2004.03.015

  • 49

    McNeillD. (1985). So you think gestures are nonverbal?Psych. Rev. 92, 350371. 10.1037/0033-295X.92.3.350

  • 50

    McNeillD. (1992). Hand and Mind.Chicago, IL: Univ. Chic. Press.

  • 51

    McNeillD. (2012). How Language Began. Gesture and Speech in Human Evolution. Cambridge: Cambridge University Press.

  • 52

    MittelbergI. (2006). Metaphor and Metonymy in Language and Gesture: Discourse Evidence for Multimodal Models of Grammar. Ann Arbor, MI: UMI.

  • 53

    MondadaL. (2016). Challenges of multimodality: Language and the body in social interaction. J. Socioling. 20, 336366. 10.1111/josl.1_12177

  • 54

    MüllerC. (2017). How recurrent gestures mean: Conventionalized contexts-of-use and embodied motivation. Gesture16, 277304. 10.1075/gest.16.2.05mul

  • 55

    OcchinoC.WilcoxS. (2017). Gesture or sign? a categorization problem. Behav. Brain Sci. 40, 3637. 10.1017/S0140525X15003015

  • 56

    OgdenR. (2013). Clicks and percussives in English conversation. J. Int. Phonetic Assoc. 43, 299320. 10.1017/S0025100313000224

  • 57

    ÖzyürekA.WillemsR.KitaS.HagoortP. (2007). On-line integration of semantic information from speech and gesture: insights from event-related brain potentials. J. Cogn. Neurosci. 19, 605616. 10.1162/jocn.2007.19.4.605

  • 58

    PeetersD.SnijdersT.HagoortP.ÖzyürekA. (2017). Linking language to the visual world: neural correlates of comprehending verbal reference to objects through pointing and visual cues. Neuropsychologia95, 2129. 10.1016/j.neuropsychologia.2016.12.004

  • 59

    PerlmanM. (2017). Debunking two myths against vocal origins of language: language is iconic and multimodal to the core. Interact. Stud. 18, 376401. 10.1075/is.18.3.05per

  • 60

    PernissP.ViglioccoG. (2014). The bridge of iconicity: from a world of experience to the experience of language. Philos. Trans. R. Soc. B369:20130300. 10.1098/rstb.2013.0300

  • 61

    PernissP.ThompsonR.ViglioccoG. (2010). Iconicity as a general property of language: evidence from spoken and signed languages. Front. Psychol.1:227. 10.3389/fpsyg.2010.00227

  • 62

    SchlenkerP.ChemlaE. (2017). Gestural agreement. Nat. Lang. Linguist. Theory36, 587625. 10.1007/s11049-017-9378-8

  • 63

    SkipperJ. (2014). Echoes of the spoken past: how auditory cortex hears context during speech perception. Philos. Trans. R. Soc. Lond. B Biol. Sci.369:20130297. 10.1098/rstb.2013.0297

  • 64

    StraubeB.GreenA.WeisS.ChatterjeeA.KircherT. (2009). Memory effects of speech and gesture binding: cortical and hippocampal activation in relation to subsequent memory performance. J. Cogn. Neurosci. 21, 821836. 10.1162/jocn.2009.21053

  • 65

    TomaselloM. (1999). The Cultural Origins of Human Cognition.Cambridge: Harvard University Press.

  • 66

    TomaselloM. (2008). The Origins of Human Communication. Cambridge: MIT Press.

  • 67

    TrompJ.PeetersD.MeyerA.HagoortP. (2017). The combined use of virtual reality and EEG to study language processing in naturalistic environments. Behav. Res. Methods50, 862869. 10.3758/s13428-017-0911-9.

  • 68

    VainioL.SchulmanM.TiippanaK.VainioM. (2013). Effect of syllable articulation on precision and power grip performance. PLoS ONE8:e53061. 10.1371/journal.pone.0053061

  • 69

    van WassenhoveV.GrantK.PoeppelD. (2005). Visual speech speeds up the neural processing of auditory speech. Proc. Natl. Acad. Sci. U.S.A.102, 11811186. 10.1073/pnas.0408949102

  • 70

    ViglioccoG.PernissP.VinsonD. (2014). Language as a multimodal phenomenon: implications for language learning, processing and evolution. Philos. Trans. R. Soc. Lond. B Biol. Sci.369:20130292. 10.1098/rstb.2013.0292

  • 71

    WeisbergJ.HubbardA.EmmoreyK. (2017). Multimodal integration of spontaneously produced representational co-speech gestures: an fMRI study. Lang. Cogn. Neurosci. 32, 158174. 10.1080/23273798.2016.1245426

  • 72

    WillemsR.ÖzyürekA.HagoortP. (2008). Seeing and hearing meaning: ERP and fMRI evidence of word versus picture integration into a sentence context. J. Cogn. Neurosci. 20, 12351249. 10.1162/jocn.2008.20085

  • 73

    WolfD.RekittkeL.-M.MittelbergI.KlasenM.MathiakD. (2017). Perceived conventionality in co-speech gestures involves the fronto-temporal language network. Front. Hum. Neurosci.11:573. 10.3389/fnhum.2017.00573

  • 74

    WrightM. (2011). On clicks in English talk-in-interaction. J. Int. Phonetic Assoc. 41, 207229. 10.1017/S0025100311000144

  • 75

    YangJ.AndricM.MathewM. (2015). The neural basis of hand gesture comprehension: a meta-analysis of functional magnetic resonance imaging studies. Neurosci. Biobehav. Rev. 57, 88104. 10.1016/j.neubiorev.2015.08.006

Summary

Keywords

multimodality, language, communication, sign language, gesture

Citation

Perniss P (2018) Why We Should Study Multimodal Language. Front. Psychol. 9:1109. doi: 10.3389/fpsyg.2018.01109

Received

15 December 2017

Accepted

11 June 2018

Published

28 June 2018

Volume

9 - 2018

Edited by

Wendy Sandler, University of Haifa, Israel

Reviewed by

Mark Aronoff, Stony Brook University, United States; Jonathan Ginzburg, Paris Diderot University, France

Updates

Copyright

*Correspondence: Pamela Perniss

This article was submitted to Language Sciences, a section of the journal Frontiers in Psychology

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics