Impact Factor 2.089
2017 JCR, Clarivate Analytics 2018

The world's most-cited Multidisciplinary Psychology journal

Review ARTICLE

Front. Psychol., 07 March 2019 | https://doi.org/10.3389/fpsyg.2019.00523

Sensus Communis: Some Perspectives on the Origins of Non-synchronous Cross-Sensory Associations

  • 1Laboratoire Ethologie, Cognition, Développement, Université Paris Nanterre, Nanterre, France
  • 2STMS Ircam-CNRS-Sorbonne Université, Paris, France
  • 3Laboratoire ETIS, Université Cergy-Pontoise, Cergy-Pontoise, France

Adults readily make associations between stimuli perceived consecutively through different sense modalities, such as shapes and sounds. Researchers have only recently begun to investigate such correspondences in infants but only a handful of studies have focused on infants less than a year old. Are infants able to make cross-sensory correspondences from birth? Do certain correspondences require extensive real-world experience? Some studies have shown that newborns are able to match stimuli perceived in different sense modalities. Yet, the origins and mechanisms underlying these abilities are unclear. The present paper explores these questions and reviews some hypotheses on the emergence and early development of cross-sensory associations and their possible links with language development. Indeed, if infants can perceive cross-sensory correspondences between events that share certain features but are not strictly contingent or co-located, one may posit that they are using a “sixth sense” in Aristotle’s sense of the term. And a likely candidate for explaining this mechanism, as Aristotle suggested, is movement.

Introduction

Everyday experience is multisensory. Even the simplest of activities entails the interaction of multiple sense modalities, and the sense modalities must operate together for perception of events to occur. How is it that our perception of objects, people and events is unitary and coherent? This question dates back to Aristotle (Suzuki, 1952) who proposed a common sense or “sensus communis” responsible for monitoring and coordinating the five senses out of which unified conscious experience arises. Indeed, Aristotle posited the sensus communis as a sixth sense; one that does not depend on specific sensory apparatus, but that is rooted in the possibility or potentiality of movement. Although great advances have been made in the fields of neuroscience, psychology and robotics in answer to Aristotle’s conundrum, the question of the origins and development of cross-sensory perception in infancy is still only partially answered. Specifically, a very limited number of studies have focused on the relations between sensory systems that convey similar or analogous information non-synchronously. A cross-sensory correspondence is the association between an attribute or dimension in one sense modality and an analogous attribute or dimension in another sense modality (Walker-Andrews et al., 1994). The way our brain integrates multimodal information and how this develops is still in debate. Nonetheless, many experiments and studies in the neurobiology of development show the importance of two major cerebral areas involved in cross-sensory integration, namely the superior colliculus (SC) and the parietal cortex (PC) (Stein et al., 2009). These two areas mature at different intervals and speeds before birth till the second-year and also represent multisensory information differently (e.g., egocentric vs. allocentric and reflexive vs. learned). Furthermore, they are not isolated circuits and work in combination with other regions in different sensorimotor loops depending on the types of behavior (automatic or rule-based), the timescale of learning and execution (short or long).

Considering the development of MI (multimodal integration) in the cortex, Keysers and Gazzola (2006) suggested that mirror neurons may be present at birth in the whole cortex. Heyes (2010) proposed that this mirroring mechanism could be based on the reinforcement learning of sensorimotor contingencies which assumes that visual (sensory) representations of action simultaneously seen and executed become linked to motor representations through Hebbian learning.

Many events involve perceiving the same property in different sense modalities. The size and shape of an object can, for example, be processed both visually and haptically, which leads to a unitary percept. Intensity of either light or sound can change over time and thus be perceived as a similar phenomenon. In fact, in everyday situations, contingent associations of different attributes simultaneously changing over time are experienced on a regular basis. A bouncing ball, for example, is perceived as regular up and down motion and as regular sound-silence alternation. Its movement generates a multimodal representation based on synchronous and analogous dynamic time variations in visual and auditory experience. Both the object’s attributes and its particular qualities of motion are thus represented.

One of the first experiences of a newborn infant is of a person talking to him/her, involving both synchronous and non-concurrent analogous changes and shifts in the contours of facial, vocal, postural, and tactile stimulations. Yet, very little is known today about the underlying processes by which very young infants associate analogous expressions occurring non-synchronously across modalities. For example, when the mother raises her eyebrows and then, less than a second later, produces a higher pitched rising-falling utterance, does the infant associate the expressive contours of her behavior across modalities? Infant-addressed behavior is usually exaggerated compared to adult-addressed expression. Furthermore, in social infant-directed expression, synchronous behavior in various modalities fosters high degrees of intersensory redundancy, which is thought to facilitate perception for young infants (Bahrick and Lickliter, 2002). An important body of knowledge has been gathered on the perception of synchronous changes in different modalities by young infants (Bremner et al., 2012). Much less research has been concerned with non-synchronous but analogous changes over time and across different sense modalities.

Some studies in adults have shown that stimuli presented non-synchronously to different sense modalities are perceived as having correspondences when they are matched on certain features such as intensity or duration. More surprisingly, adults also perceive stable correspondences between certain object attributes that are not straightforwardly connected. For example, adults readily associate higher pitch in sound with angularity of shape or acidity in taste and smell (Crisinel and Spence, 2009; Deroy et al., 2013). The existence of these types of correspondences raises important questions. Can these correspondences be considered as analogies? Are they learned through experience? Are they automatic? Are there different kinds of cross-sensory correspondences? When do infants make correspondences between static cross-sensory attributes, between cross-sensory events that change together over time and between cross-sensory events that share certain features but do not co-occur? Do some types of cross-sensory correspondences develop before others? Answering these questions could provide insight into the nature and development of abstraction abilities. The present review will focus on: (1) the development of cross-sensory associations in the first year and (2) its possible role in language development.

Cross-Sensory Associations in the First Year of Life

In the past 20 years or so, the study of cross-sensory perception has largely been based on a developmental model of progressive integration of the senses. This enquiry has intensified in recent years in the broad field of cognitive science. Much research has focused on how adults integrate information perceived in different sense modalities. Although everyday events are perceived as unitary, they are usually thought to involve the integration of information processed independently by each sensory system. Multimodal knowledge would be acquired through repeated experiences of such intersensory integration, which is extended to new situations. For example, large heavy objects usually make deep and loud sounds when they fall. This repeated association may orient representations based on a unimodal experience, so that a loud sound can conjure up a representation of a large object in the absence of visual input. Implicit multimodal knowledge may also be obtained from perceived correspondences between non-concurrent events. This applies particularly for perception of animate beings where multimodal knowledge could be driven by basic principles of sociality. Thus, the fact that young infants are highly motived toward social stimuli, and spend much more time in contact with persons than with objects, might explain why they rapidly learn to associate sensory experience in a given modality with sensory experience in other modalities. It is also possible that infant’s readiness for social interaction is linked to a precocious ability to perceive correspondences between modalities.

The use of multimodal knowledge is therefore a fundamental ability, enabling the identification of events and adaptive responses to these events (Simon, 2008). Many studies investigating the behavioral effects of multisensory experience have shown the benefits it affords (Press et al., 2004; Lippert et al., 2007) and it appears quite clearly to facilitate and optimize learning (Seitz et al., 2006). Yet, the development of such capacities is still unclear. One hypothesis formulated by researchers, the early integration hypothesis, proposes that cross-sensory integration is already present from birth while another hypothesis, the late integration hypothesis, emphasizes the role of experience in the development of cross-sensory associations (Dionne-Dostie et al., 2015). Here, we will focus on how infants sense their world and how they make sense of it, rather than on how the senses operate alone and together in infants. In particular, we will distinguish studies on cross-modal integration from studies on cross-modal transfer.

Studies on the Precocity of Cross-Sensory Integration

Although the literature on cross-sensory integration in infants is fairly recent, the idea that this capacity emerges very early during development is not new. According to Gibson’s (2014) ecological theory of perception, amodal information, that is information not specific to any one modality and that can be conveyed redundantly across many senses, is directly obtained from adaptive interaction between organisms and their environments. Duration, spatial extent, temporal synchrony, shape, and intensity are considered to be important amodal characteristics of objects and events. Gibson (1969) proposed that amodal spatial and temporal dimensions are available to all sensory modalities already from birth.

Based on a large body of experimental research on the role of these amodal characteristics in early perceptual abilities, Bahrick and Lickliter (2000) put forward the “intersensory redundancy hypothesis” to explain how infants perceive coherent, unified multimodal objects and events through different sense modalities. This theory proposes that, in order to be perceptually integrated, the same information must be spatially coordinated and temporally synchronous across two or more sensory modalities and that cross-sensory integration is thus only possible for amodal properties that are not specific to a single sense modality (e.g., shape, rhythm, duration, and intensity). In other words, regardless of which sensory modality is solicited, similar qualities are perceived through the integration of information from diverse sources. For instance, the sound and sight of a bouncing ball are integrated because auditory and visual information is synchronous (shares a common tempo and rhythm) and originates from the same location. Therefore, sensitivity to amodal properties allows young infants to selectively direct their attention to unitary and meaningful events in their environment (Bahrick, 1992).

Prefiguring Bahrick and Lickliter’s hypothesis, Lewkowicz and Turkewitz (1980) were the first to demonstrate that 3-week-old infants can match sound and light intensities. In their seminal study, infants exposed to light spots of different intensities looked preferentially toward a light of intermediate intensity. However, infants first exposed to sounds at various intensity levels and then to light spots of similar intensities preferred the light spot that matched the sound stimulus. Lewkowicz and Turkewitz (1980) concluded from these results that young infants attend to quantitative variations in stimuli. It has also been shown that newborns are able to learn arbitrary auditory-visual associations (e.g., between an oriented colored line and a syllable), but only when the visual and auditory information are presented synchronously (Slater and Kirby, 1998). Furthermore, newborn infants are able to associate objects and sounds on the basis of temporal synchrony (Slater and Kirby, 1998). They can also associate a vocal sound from a non-human primate with the corresponding lip shape of the primate’s face (Lewkowicz et al., 2010) based on temporal synchrony.

Some authors have proposed that the results of these studies can be explained by the importance of synchrony for perception in the first months of life (Bahrick, 1987; Lewkowicz, 1996; Bahrick and Lickliter, 2012). According to this view, early cross-sensory integration would mainly be based on temporal synchrony (e.g., a sound and an image occurring together), and spatial colocation (e.g., coincidence of the location of a sound and an image). Quite paradoxically however, according to this view, an infant would not be surprised to perceive his/her mother’s face with a male voice as long as the lips’ movements are temporally synchronous with the mouth movements.

While temporal synchrony has been recognized as a fundamental dimension for establishing the link between visual and auditory information about an event or an object, it does not appear to be necessary for infants to integrate and make sense of multimodal events. For example, Izard et al. (2009) showed that newborn infants spontaneously associate visuo-spatial arrays of objects with auditory sequences of events based on similarities in numerosity. Despite the absence of synchrony between the objects and the sounds, newborns were able to respond to abstract numerical quantities presented across these two modalities (i.e., auditory and visual). Guellaï et al. (2016) have also shown that newborns make accurate audio-visual associations based on the non-synchronous presentation of stimuli. Infants were presented with two dynamic facial displays uttering two different sentences but where only one of them corresponded to one of the two displays. Both facial displays started and stopped at the same time. Thus infants used cues other than temporal synchrony to match utterances to the corresponding facial movements.

As already stated, the main studies in the literature have shown that infants are able to integrate information from different sense modalities very early on, and primarily based on temporal synchrony. Nonetheless, in everyday social situations, many multimodal events can be considered as sharing common features that occur non-synchronously. For example, the infant can be tickled by his/her mother, and a few seconds later she can say something that prosodically matches her gesture. In order to perceive such an event as unified and to make sense of it, infants must transfer information from one modality (proprioception) to other modalities (audio-visual). Therefore, cross-modal transfer must be an important ability in early infancy.

Studies on the Precocity of Cross-Modal Transfer

Studies on the cross-modal transfer of information from touch to vision have revealed that neonates are able to process and encode shape information about haptically explored objects and to discriminate between subsequently presented visual objects (Streri and Gentaz, 2004). Newborns are also able to visually recognize a texture that they previously touched and to tactually recognize the texture that they previously saw even though they are presented sequentially (Sann and Streri, 2008). Meltzoff and Borton (1979) pioneering study reported that 1-month-old infants show a clear visual preference for objects with which they had been familiarized through oral presentation. Thus, a handful of studies have shown that transfer of information is possible from haptic experience to vision already at birth. In everyday life, infants explore the world haptically and also experience being touched by other people during social interactions and care practices. However, very little is known to date about infants’ perception of being touched and how that particular kind of information could be transferred to other sensory modalities such as vision and audition. One recent study suggest that newborns are capable of associating tactile stimulation on their own body with a congruent visual image of the same tactile stimulus (Filippetti et al., 2015).

In early social engagement with an adult, infants probably experience non-synchronous yet contingent expressions in different modalities. Therefore it is likely that infants are capable of integrating matching non-synchronous events as well as synchronous ones. Nevertheless, it remains unclear why cross-sensory matching abilities are present so early in life. One possibility is that such a process allows infants to perceive invariant aspects of the environments they must rapidly adapt to. Moreover, it has important implications for the development of face and emotion recognition. Studies using a visual preference paradigm in a multimodal context for human faces have reported that as early as 2 months of age, infants can associate phonetic information from voices with lip movements (Kuhl and Meltzoff, 1982; Patterson and Werker, 1999, 2003). More recently, it has been shown that 8-month-olds can associate a speech stream with the corresponding facial movements even when the speech stream is low-pass filtered (Kitamura et al., 2014), suggesting that prosody can be perceived both in the auditory and visual modalities (Esteve-Gibert and Guellaï, 2018). In addition, 4-month-old infants can perceive affect (joy, sadness, or anger) in speech sequences that are supported by audio-visual presentations of faces (Walker-Andrews and Lennon, 1991; Flom and Bahrick, 2007).

Overall, these studies support the idea of a precocious cross-sensory perception capacity, based on both cross-modal transfer and cross-sensory integration. The underlying mechanisms and the development of this capacity is still puzzling and remain largely unknown. Are general associative learning mechanisms sufficient to explain how infants come to pair sensory cues across modalities, or do specific learning processes or constraints guide the acquisition of some (or all) cross-sensory correspondences? Studies using neuroimaging techniques could help identify the underlying mechanisms of cross-sensory perception. Furthermore, research on prenatal perception suggests that at birth infants have some experience of multisensory events. Studies on premature babies could provide compelling data to help understand the foundations and mechanisms of cross-sensory associations. Experimental studies of cross-sensory perception suggest that sense modalities operate from the earliest days of life.

Evidence for the existence of not straightforwardly redundant cross-sensory integration in infants (0 to 12 months old) and toddlers (12 to 24 months old) has begun to be gathered. Cross-sensory correspondences between features of speech sounds and visual shapes have been demonstrated as early as 4-months of age (Peña et al., 2011; Ozturk et al., 2013). Similarly to adults, infants associate specific shapes to particular linguistic sounds. Why do such associations exist? In the next section, we will focus on sound symbolism as a specific form of cross-sensory association that could be an important starting point for the development of language.

The Particular Case of Sound Symbolism

One of the most amazing capacities of our species is its ability to combine symbols in order to communicate specific meanings (Deacon, 1997). This capacity underlies the emergence of all natural languages. While the phylogeny and ontogeny of this skill have been the center of interest of many researchers in the last decades, they have raised vast and as yet unresolved debates across disciplines. For example, there is no clear account of how the human mind came to support a symbolic system that is largely disconnected from direct perception in the first place. In the “Cratylus,” Plato describes Hermogenes as stating “if one substituted one name for another, the latter would be as correct as the former” to which Cratylus answers that there is “for each object a name that is its own and that belongs to it intrinsically, or by its nature.” Socrates then concludes that even if in general the link between a thing and its name is arbitrary, there are, nevertheless, some noble words whose sound reflects their meaning. Yet one of the founders of modern linguistics, De Saussure (1989), proposed that all linguistic symbols are arbitrary, in other words, that there is no natural connection between linguistic form and linguistic meaning.

Even if the later proposition is the dominant view in linguistics today, a number of scholars have noted and studied the occurrence of analogical or iconic relations between words and their references. Some linguists and psychologists (Köhler, 1929; Sapir, 1929) have provided empirical evidence of non-arbitrary links between the signifier and the signified. The idea is that certain sounds are meaningful in themselves. This idea was first proposed by Gestalt psychologists, and is known as “sound symbolism.” One of the most famous experiments on this question was proposed by Köhler (1929). Using a simple forced choice task, adult participants were given two nonsense words, maluma and takete, and two abstract shapes, a rounded shape and an angular one. They were asked to match the words to the shapes. Kohler’s striking result was that almost all subjects identified the rounded shape as the maluma and the angular one as the takete. These findings were the first to show that linguistic sounds can bear an indirect yet natural connection to their referents. Furthermore, this connection is distinct from linguistic sound imitation, or sound analogy which involves association within the same modality. The Köhler task demonstrates a cross-sensory association between two sensory domains, vision (i.e., shape) and sound (i.e., spoken word).

In the past decades Kohler’s naming bias has been replicated cross-culturally (Bremner et al., 2013) and with various stimuli (Knöferle and Spence, 2012; Hanson-Vaux et al., 2013). Ramachandran and Hubbard (2001) used different non-word and visual stimuli in their version of Köhler’s experiment. Their words, bouba and kiki and the term ‘bouba kiki effect’ are now well-known in the literature. In their experiment, the rounded shape is consistently named as bouba and the angular shape as kiki.

Most studies to date on sound symbolism have explored the phenomenon in adult populations. Though the results are quite striking, it is not clear to what extent sound symbolism is dependent on experience. What about infants? One way to resolve this issue is to explore it developmentally, through infancy.

It is known that already from birth infants are sensitive to congruency between audiovisual speech inputs (Guellaï et al., 2016). Nevertheless, only a handful of studies have investigated the emergence of sound symbolism. So far, studies on infants have explored only associations between vision and audition (Peña et al., 2011; Maurer et al., 2013; Ozturk et al., 2013; Imai and Kita, 2014), and, moreover, they present divergent results. Notably, some studies have shown the bouba-kiki effect as early as 4 months (Peña et al., 2011; Ozturk et al., 2013) but these findings have been undermined by other studies that did not find this effect in infants (Fort et al., 2013).

Interestingly, recent studies have shown that this bias is not limited to the auditory-visual domains in adults. It has been found across different sensory modalities such as taste and sound (Knöferle and Spence, 2012) or odor and vision (e.g., a “sweet” odor is associated with a rounded shape and an “acid” one with a spiky shape) (Hanson-Vaux et al., 2013). The reason for such biased associations remains largely unknown but some possible explanations can be proposed. One possibility is that these biases, which are grounded in perception, played an important role in the evolution of the human language capacity. This possibility has received only little interest so far. Some authors proposed that these associations supported the emergence of a small-scale communication system, or protolanguage, mainly based on non-arbitrary associations, from which fully symbolic languages emerged (Cuskley, 2013). How might a non-arbitrary protolanguage have become arbitrary? One way to answer this question is to adopt an ontogenetic perspective and to investigate sound symbolism at different stages of early language development, exploring various types of cross-sensory correspondences across all sense modalities (Walker, 2016).

Conclusion and Perspectives

The present review aimed at presenting a state of the art of studies on cross-sensory perception. It bridges a gap between studies focused on cross-sensory integration which show that spatio-temporal co-occurrence is a crucial context for infants to associate events in different sense modalities and other studies showing that even without strict co-occurrence very young infants can associate events in different sense modalities. These two areas of research together thus suggest that infants are equipped from birth to make sense of their environments through integration and transfer of experience between sense modalities. Yet these lines of research do not satisfactorily inform us on Aristotle’s suggestion of a “sensus communis” and further studies are needed to understand if and how young infants are able to associate non-synchronous events in different sense modalities. Indeed, if infants can perceive cross-sensory correspondences between events that share certain features but are not strictly synchronous or co-located, one may posit that they are using a “sixth sense” in Aristotle’s sense of the term. And a likely candidate for explaining this mechanism, as Aristotle suggested, is movement. Indeed, some researchers have begun to highlight the existence of strong overlap in brain structures involved in the perception of vision, sound and touch together and in motor control and planning (Gallese and Lakoff, 2005). Starting in utero, body movement may thus be a crucial process for developing a multi-modally integrated brain. Furthermore, if infants are able to perceive seemingly complex correspondences, between analogous but non-identical phenomena, one may assume that they are well equipped to make sense of the complex expressions of their social partners and to respond with an equally sophisticated preverbal sense-making ability.

Few studies to date have explored the development of these different types of cross-sensory correspondence longitudinally. An interesting approach would be to use developmental robotics to build and test models of cross-sensory perception pathways. The field of developmental robotics looks toward infant development for inspiration, data, and guidance, in order to build models of learning that may be useful for a better understanding of typical and atypical human development of cross-sensory experience. Recently, Brunetti et al. (2018) evidenced that pitch/size correspondence in adults was relative in nature, that is, adults match sound pitch to the corresponding image size depending on number of trials. Therefore, the role of experience in the development of cross-sensory integration is still unclear. The field of developmental robotics is motivated by the construction of autonomous robots and also by the idea of using the robot as a tool to investigate cognitive models. For example, Thomaz et al. (2005) used a robot head named “Kismet” which was able to: recognize gestures (pointing gestures) and facial expressions; with a head-mounted tracker to evaluate a human’s object of attention; and to respond to vocal stimuli. These types of applications in the field of developmental robotics could help understanding the role of experience in cross-sensory associations.

Author Contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

Funding

We thank the Fyssen Foundation for its financial support.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Bahrick, L. E. (1987). Infants’ intermodal perception of two levels of temporal structure in natural events. Infant Behav. Dev. 10, 387–416. doi: 10.1016/0163-6383(87)90039-7

CrossRef Full Text | Google Scholar

Bahrick, L. E. (1992). Infants’ perceptual differentiation of amodal and modality-specific audio-visual relations. J. Exp. Child Psychol. 53, 180–199. doi: 10.1016/0022-0965(92)90048-B

CrossRef Full Text | Google Scholar

Bahrick, L. E., and Lickliter, R. (2000). Intersensory redundancy guides attentional selectivity and perceptual learning in infancy. Dev. Psychol. 36, 190–201. doi: 10.1037/0012-1649.36.2.190

PubMed Abstract | CrossRef Full Text | Google Scholar

Bahrick, L. E., and Lickliter, R. (2002). “Intersensory redundancy guides early perceptual and cognitive development,” in Advances in Child Development and Behavior, Vol. 30, ed. R. Kail (New York, NY: Academic Press), 153–187.

Google Scholar

Bahrick, L. E., and Lickliter, R. (2012). “The role of intersensory redundancy in early perceptual, cognitive, and social development,” in Multisensory Development, eds A. J. Bremner, D. J. Lewkowicz, and C. Spence (New York, NY: Oxford University Press), 183–206.

Google Scholar

Bremner, A. J., Caparos, S., Davidoff, J., de Fockert, J., Linnell, K. J., and Spence, C. (2013). “Bouba” and “Kiki” in Namibia? A remote culture make similar shape-sound matches, but different shape-taste matches to Westerners. Cognition 126, 165–172. doi: 10.1016/j.cognition.2012.09.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Bremner, A. J., Lewkowicz, D. J., and Spence, C. (eds) (2012). Multisensory Development. Oxford: Oxford University Press. doi: 10.1093/acprof:oso/9780199586059.001.0001

CrossRef Full Text | Google Scholar

Brunetti, R., Indraccolo, A., Del Gatto, C., Spence, C., and Santangelo, V. (2018). Are crossmodal correspondences relative or absolute? Sequential effects on speeded classification. Attent. Percept. Psychophys. 80, 527–534. doi: 10.3758/s13414-017-1445-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Crisinel, A. S., and Spence, C. (2009). Implicit association between basic tastes and pitch. Neurosci. Lett. 464, 39–42. doi: 10.1016/j.neulet.2009.08.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Cuskley, C. F. (2013). Shared Cross-Modal Associations and the Emergence of the Lexicon. Ph.D. thesis, University of Edinburgh, Edinburgh.

Google Scholar

De Saussure, F. (1989). Cours de Linguistique Générale: Édition critique (Vol. 1). Wiesbaden: Harrassowitz Verlag.

Google Scholar

Deacon, T. (1997). The Symbolic Species. New York, NY: W. W. Norton and Company.

Google Scholar

Deroy, O., Crisinel, A. S., and Spence, C. (2013). Crossmodal correspondences between odors and contingent features: odors, musical notes, and geometrical shapes. Psychon. Bull. Rev. 20, 878–896. doi: 10.3758/s13423-013-0397-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Dionne-Dostie, E., Paquette, N., Lassonde, M., and Gallagher, A. (2015). Multisensory integration and child neurodevelopment. Brain Sci. 5, 32–57. doi: 10.3390/brainsci5010032

PubMed Abstract | CrossRef Full Text | Google Scholar

Esteve-Gibert, N., and Guellaï, B. (2018). Prosody in the auditory and visual domains: a developmental perspective. Front. Psychol. 9:338. doi: 10.3389/fpsyg.2018.00338

PubMed Abstract | CrossRef Full Text | Google Scholar

Filippetti, M. L., Orioli, G., Johnson, M. H., and Farroni, T. (2015). Newborn body perception: sensitivity to spatial congruency. Infancy 20, 455–465. doi: 10.1111/infa.12083

PubMed Abstract | CrossRef Full Text | Google Scholar

Flom, R., and Bahrick, L. E. (2007). The development of infant discrimination of affect in multimodal and unimodal stimulation: the role of intersensory redundancy. Dev. Psychol. 43:238. doi: 10.1037/0012-1649.43.1.238

PubMed Abstract | CrossRef Full Text | Google Scholar

Fort, M., Weiß, A., Martin, A., and Peperkamp, S. (2013). “Looking for the bouba-kiki effect in prelexical infants,” in International Conference on Auditory-Visual Speech Processing (AVSP), Norwich.

Google Scholar

Gallese, V., and Lakoff, G. (2005). The brain’s concepts: the role of the sensory-motor system in conceptual knowledge. Cogn. Neuropsychol. 22, 455–479. doi: 10.1080/02643290442000310

PubMed Abstract | CrossRef Full Text | Google Scholar

Gibson, E. J. (1969). Principles of Perceptual Learning and Development. East Norwalk, CT: Appleton-Century-Crofts.

Google Scholar

Gibson, J. J. (2014). The Ecological Approach to Visual Perception: Classic Edition. London: Psychology Press. doi: 10.4324/9781315740218

CrossRef Full Text | Google Scholar

Guellaï, B., Streri, A., Chopin, A., Rider, D., and Kitamura, C. (2016). Newborns’ sensitivity to the visual aspects of infant-directed speech: Evidence from point-line displays of talking faces. J. Exp. Psychol. 42:1275. doi: 10.1037/xhp0000208

PubMed Abstract | CrossRef Full Text | Google Scholar

Hanson-Vaux, G., Crisinel, A. S., and Spence, C. (2013). Smelling shapes: crossmodal correspondences between odors and shapes. Chem. Senses 38, 161–166. doi: 10.1093/chemse/bjs087

PubMed Abstract | CrossRef Full Text | Google Scholar

Heyes, C. (2010). Where do mirror neurons come from? Neurosci. Biobehav. Rev. 30, 575–583. doi: 10.1016/j.neubiorev.2009.11.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Imai, M., and Kita, S. (2014). The sound symbolism bootstrapping hypothesis for language acquisition and language evolution. Philos. Trans. R. Soc. B 369:20130298. doi: 10.1098/rstb.2013.0298

PubMed Abstract | CrossRef Full Text | Google Scholar

Izard, V., Sann, C., Spelke, E. S., and Streri, A. (2009). Newborn infants perceive abstract numbers. Proc. Natl. Acad. Sci. U.S.A. 106, 10382–10385. doi: 10.1073/pnas.0812142106

PubMed Abstract | CrossRef Full Text | Google Scholar

Keysers, C., and Gazzola, V. (2006). Towards a unifying neural theory of social cognition. Prog. Brain Res. 156, 379–401. doi: 10.1016/S0079-6123(06)56021-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Kitamura, C., Guellaï, B., and Kim, J. (2014). Motherese by eye and ear: infants perceive visual prosody in point-line displays of talking heads. PLoS One 9:e111467. doi: 10.1371/journal.pone.0111467

PubMed Abstract | CrossRef Full Text | Google Scholar

Knöferle, K., and Spence, C. (2012). Crossmodal correspondences between sounds and tastes. Psychon. Bull. Rev. doi: 10.3758/s13423-012-0321-z [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

Köhler, W. (1929). Gestalt Psychology. New York, NY: New American Library.

Google Scholar

Kuhl, P. K., and Meltzoff, A. N. (1982). The bimodal perception of speech in infancy. Science 218, 1138–1141. doi: 10.1126/science.7146899

CrossRef Full Text | Google Scholar

Lewkowicz, D., and Turkewitz, G. (1980). Intersensory interaction in newborns: modification of visual preferences following exposure to sound. Child Dev. 52, 827–832. doi: 10.2307/1129083

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewkowicz, D. J. (1996). Perception of auditory–visual temporal synchrony in human infants. J. Exp. Psychol. 22, 1094–1106. doi: 10.1037/0096-1523.22.5.1094

CrossRef Full Text | Google Scholar

Lewkowicz, D. J., Leo, I., and Simion, F. (2010). Intersensory perception at birth: newborns match non human primate faces and voices. Infancy 15, 46–60. doi: 10.1111/j.1532-7078.2009.00005.x

CrossRef Full Text | Google Scholar

Lippert, M., Logothetis, N. K., and Kayser, C. (2007). Improvement of visual contrast detection by a simultaneous sound. Brain Res. 1173, 102–109. doi: 10.1016/j.brainres.2007.07.050

PubMed Abstract | CrossRef Full Text | Google Scholar

Maurer, D., Gibson, L. C., and Spector, F. (2013). “Synesthesia in infants and very young children,” in Oxford Handbook of Synesthesia, eds J. Simner and E. M. Hubbard (Oxford: Oxford University Press), 46–63. doi: 10.1093/oxfordhb/9780199603329.013.0003

CrossRef Full Text | Google Scholar

Meltzoff, A. N., and Borton, R. W. (1979). Intermodal matching by human neonates. Nature 282, 403–404. doi: 10.1038/282403a0

CrossRef Full Text | Google Scholar

Ozturk, O., Krehm, M., and Vouloumanos, A. (2013). Sound symbolism in infancy: evidence for sound–shape cross-modal correspondences in 4-month-olds. J. Exp. Child Psychol. 114, 173–186. doi: 10.1016/j.jecp.2012.05.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Patterson, M. L., and Werker, J. F. (1999). Matching phonetic information in lips and voice is robust in 4.5-month-old infants. Infant Behav. Dev. 22, 237–247. doi: 10.1016/j.cognition.2008.05.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Patterson, M. L., and Werker, J. F. (2003). Two-month-old infants match phonetic information in lips and voice. Dev. Sci. 6, 191–196. doi: 10.1016/j.cognition.2008.05.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Peña, M., Mehler, J., and Nespor, M. (2011). The role of audiovisual processing in early conceptual development. Psychol. Sci. 22, 1419–1421. doi: 10.1177/0956797611421791

PubMed Abstract | CrossRef Full Text | Google Scholar

Press, C., Taylor-Clarke, M., Kennett, S., and Haggard, P. (2004). Visual enhancement of touch in spatial body representation. Exp. Brain Res. 154, 238–245. doi: 10.1007/s00221-003-1651-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ramachandran, V. S., and Hubbard, E. M. (2001). Synaesthesia–a window into perception, thought and language. J. Conscious. Stud. 8, 3–34.

Google Scholar

Sann, C., and Streri, A. (2008). Inter-manual transfer of object texture and shape in human neonates. Neuropsychologia 46, 698–703. doi: 10.1016/j.neuropsychologia.2007.09.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Sapir, E. (1929). A study in phonetic symbolism. J. Exp. Psychol. 12, 225–239. doi: 10.1037/h0070931

CrossRef Full Text | Google Scholar

Seitz, A. R., Kim, R., and Shams, L. (2006). Sound facilitates visual learning. Curr. Biol. 16, 1422–1427. doi: 10.1016/j.cub.2006.05.048

PubMed Abstract | CrossRef Full Text | Google Scholar

Simon, S. A. (2008). Merging of the senses. Front. Neurosci 2, 13–14. doi: 10.3389/neuro.01.019.2008

PubMed Abstract | CrossRef Full Text | Google Scholar

Slater, A., and Kirby, R. (1998). Innate and learned perceptual abilities in the newborn infant. Exp. Brain Res. 123, 90–94. doi: 10.1007/s002210050548

CrossRef Full Text | Google Scholar

Stein, B. E., Stanford, T. R., and Rowland, B. A. (2009). The neural basis of multisensory integration in the midbrain: its organization and maturation. Hear. Res. 258, 4–15. doi: 10.1016/j.heares.2009.03.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Streri, A., and Gentaz, E. (2004). Cross-modal recognition of shape from hand to eyes and handedness in human newborns. Neuropsychologia 42, 1365–1369. doi: 10.1016/j.neuropsychologia.2004.02.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Suzuki, A. I. (1952). The Role Of Sensus Communis in Aristotle, Thomas Aquinas, Locke and Kant. Doctoral dissertation, Boston University, Boston, MA.

Google Scholar

Thomaz, A. L., Berlin, M., and Breazeal, C. (2005). “Robot science meets social science: an embodied computational model of social referencing,” in Proceedings of the Workshop Toward Social Mechanisms of Android Science, Italy, 7–17.

Google Scholar

Walker, P. (2016). Cross-sensory correspondences and symbolism in spoken and written language. J. Exp. Psychol. 42, 1339–1361. doi: 10.1037/xlm0000253

PubMed Abstract | CrossRef Full Text | Google Scholar

Walker-Andrews, A., Lewkowicz, D. J., and Lickliter, R. (1994). “Taxonomy for intermodal relations,” in The Development of Intersensory Perception: Comparative Perspectives, eds D. J. Lewkowicz and R. Lickliter (Hillsdale, NJ: Lawrence Erlbaum Associates, Inc.), 39–56.

Google Scholar

Walker-Andrews, A. S., and Lennon, E. (1991). Infants’ discrimination of vocal expressions: contributions of auditory and visual information. Infant Behav. Dev. 14, 131–142. doi: 10.1016/0163-6383(91)90001-9

CrossRef Full Text | Google Scholar

Keywords: cross-sensory, correspondences, infancy, development, perception

Citation: Guellaï B, Callin A, Bevilacqua F, Schwarz D, Pitti A, Boucenna S and Gratier M (2019) Sensus Communis: Some Perspectives on the Origins of Non-synchronous Cross-Sensory Associations. Front. Psychol. 10:523. doi: 10.3389/fpsyg.2019.00523

Received: 26 July 2018; Accepted: 22 February 2019;
Published: 07 March 2019.

Edited by:

Timothy L. Hubbard, Arizona State University, United States

Reviewed by:

Valerio Santangelo, University of Perugia, Italy
Ric Dalla Volta, Università degli Studi Magna Græcia di Catanzaro, Italy

Copyright © 2019 Guellaï, Callin, Bevilacqua, Schwarz, Pitti, Boucenna and Gratier. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Bahia Guellaï, bahia.guellai@gmail.com Maya Gratier, gratier@gmail.com