<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychol.</journal-id>
<journal-title>Frontiers in Psychology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychol.</abbrev-journal-title>
<issn pub-type="epub">1664-1078</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyg.2019.02339</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Modeling Sensory Preference in Speech Motor Planning: A Bayesian Modeling Framework</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Patri</surname> <given-names>Jean-Fran&#x000E7;ois</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/611234/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Diard</surname> <given-names>Julien</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/69320/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Perrier</surname> <given-names>Pascal</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/143355/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Universit&#x000E9; Grenoble Alpes, CNRS, Grenoble INP, GIPSA-lab</institution>, <addr-line>Grenoble</addr-line>, <country>France</country></aff>
<aff id="aff2"><sup>2</sup><institution>Universit&#x000E9; Grenoble Alpes, CNRS, LPNC</institution>, <addr-line>Grenoble</addr-line>, <country>France</country></aff>
<aff id="aff3"><sup>3</sup><institution>Cognition Motion and Neuroscience Unit, Fondazione Istituto Italiano di Tecnologia</institution>, <addr-line>Genova</addr-line>, <country>Italy</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Pascal van Lieshout, University of Toronto, Canada</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Bernd J. Kr&#x000F6;ger, RWTH Aachen University, Germany; Satrajit S. Ghosh, Massachusetts Institute of Technology, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Pascal Perrier <email>pascal.perrier&#x00040;grenoble-inp.fr</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Language Sciences, a section of the journal Frontiers in Psychology</p></fn></author-notes>
<pub-date pub-type="epub">
<day>25</day>
<month>10</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>10</volume>
<elocation-id>2339</elocation-id>
<history>
<date date-type="received">
<day>30</day>
<month>04</month>
<year>2019</year>
</date>
<date date-type="accepted">
<day>01</day>
<month>10</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2019 Patri, Diard and Perrier.</copyright-statement>
<copyright-year>2019</copyright-year>
<copyright-holder>Patri, Diard and Perrier</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Experimental studies of speech production involving compensations for auditory and somatosensory perturbations and adaptation after training suggest that both types of sensory information are considered to plan and monitor speech production. Interestingly, individual sensory preferences have been observed in this context: subjects who compensate less for somatosensory perturbations compensate more for auditory perturbations, and <italic>vice versa</italic>. We propose to integrate this sensory preference phenomenon in a model of speech motor planning using a probabilistic model in which speech units are characterized both in auditory and somatosensory terms. Sensory preference is implemented in the model according to two approaches. In the first approach, which is often used in motor control models accounting for sensory integration, sensory preference is attributed to the relative precision (i.e., inverse of the variance) of the sensory characterization of the speech motor goals associated with phonological units (which are phonemes in the context of this paper). In the second, &#x0201C;more original&#x0201D; variant, sensory preference is implemented by modulating the sensitivity of the comparison between the predicted sensory consequences of motor commands and the sensory characterizations of the phonemes. We present simulation results using these two variants, in the context of the adaptation to an auditory perturbation, implemented in a 2-dimensional biomechanical model of the tongue. Simulation results show that both variants lead to qualitatively similar results. Distinguishing them experimentally would require precise analyses of partial compensation patterns. However, the second proposed variant implements sensory preference without changing the sensory characterizations of the phonemes. This dissociates sensory preference and sensory characterizations of the phonemes, and makes the account of sensory preference more flexible. Indeed, in the second variant the sensory characterizations of the phonemes can remain stable, when sensory preference varies as a response to cognitive or attentional control. This opens new perspectives for capturing speech production variability associated with aging, disorders and speaking conditions.</p></abstract>
<kwd-group>
<kwd>speech motor control</kwd>
<kwd>Bayesian modeling</kwd>
<kwd>sensory integration</kwd>
<kwd>sensory preference</kwd>
<kwd>speech motor goals</kwd>
</kwd-group>
<contract-sponsor id="cn001">FP7 Ideas: European Research Council<named-content content-type="fundref-id">10.13039/100011199</named-content></contract-sponsor>
<contract-sponsor id="cn002">H2020 European Research Council<named-content content-type="fundref-id">10.13039/100010663</named-content></contract-sponsor>
<counts>
<fig-count count="6"/>
<table-count count="0"/>
<equation-count count="16"/>
<ref-count count="75"/>
<page-count count="14"/>
<word-count count="11171"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>The recent history of research that investigates the links between phonology, production and perception of speech has been marked by vigorous exchanges between proponents of purely acoustic/auditory theories (Stevens, <xref ref-type="bibr" rid="B67">1972</xref>; Stevens and Blumstein, <xref ref-type="bibr" rid="B69">1978</xref>; Blumstein and Stevens, <xref ref-type="bibr" rid="B3">1979</xref>; Lindblom, <xref ref-type="bibr" rid="B43">1990</xref>; Sussman et al., <xref ref-type="bibr" rid="B70">1991</xref>) for whom the physical correlates of phonological units would be exclusively in the acoustic domain, and proponents of theories who rather saw these correlates primarily in the articulatory/somatosensory domain (Fowler, <xref ref-type="bibr" rid="B16">1986</xref>; Saltzman, <xref ref-type="bibr" rid="B61">1986</xref>). These debates were all the more vigorous because they were related to important theoretical issues around phonological theories (Chomsky and Halle, <xref ref-type="bibr" rid="B8">1968</xref>; Clements, <xref ref-type="bibr" rid="B9">1985</xref>; Keyser and Stevens, <xref ref-type="bibr" rid="B35">1994</xref> vs. Browman and Goldstein, <xref ref-type="bibr" rid="B4">1989</xref>, <xref ref-type="bibr" rid="B5">1992</xref>; Goldstein and Fowler, <xref ref-type="bibr" rid="B22">2003</xref>) and cognitive theories of perception (Diehl and Kluender, <xref ref-type="bibr" rid="B11">1989</xref> vs. Gibson, <xref ref-type="bibr" rid="B21">1979</xref> vs. Liberman et al., <xref ref-type="bibr" rid="B42">1967</xref>).</p>
<p>As a consequence, models that were designed to simulate and investigate the process of articulation and sound production from the specification of phonological sequences (we will call these models Speech Production Models henceforth) were split into two main categories: models in which the goals of the speech task were specified in the articulatory domain (Coker, <xref ref-type="bibr" rid="B10">1976</xref>; The Task Dynamics Model: Kelso et al., <xref ref-type="bibr" rid="B33">1986</xref>; Saltzman and Munhall, <xref ref-type="bibr" rid="B62">1989</xref>; The DIVA Model Version 1: Guenther, <xref ref-type="bibr" rid="B23">1995</xref>; Kr&#x000F6;ger et al., <xref ref-type="bibr" rid="B37">1995</xref>; The C/D model: Fujimura, <xref ref-type="bibr" rid="B19">2000</xref>), and models in which the goals were specified in the acoustic domain (The DIVA Model Version 2: Guenther et al., <xref ref-type="bibr" rid="B25">1998</xref>; GEPPETO: Perrier et al., <xref ref-type="bibr" rid="B56">2005</xref>).</p>
<p>A number of experimental studies have been carried out in order to find clear support for one or the other of these theories. The majority of them relied on perturbation paradigms, in which one of the modalities, either acoustic or articulatory, was perturbed. Patterns of behavioral adaptation to perturbation of the jaw with bite-blocks (Gay et al., <xref ref-type="bibr" rid="B20">1981</xref>) or of the lips with lip-tubes (Savariaux et al., <xref ref-type="bibr" rid="B63">1995</xref>) were interpreted as evidence for the specification of the goal in the acoustic/auditory domain, whereas adaptation in response to a perturbation of the jaw with a velocity-dependent force field (Tremblay et al., <xref ref-type="bibr" rid="B72">2003</xref>) supported the hypothesis of a goal in the articulatory/somatosensory domain. In the absence of any evidence supporting undeniably one of these theories, new theories emerged assuming that phonological units could be associated with both auditory and somatosensory goals (see for example the concept of &#x0201C;perceptuo-motor unit&#x0201D; in the Perception-for-Action-Control Theory of Schwartz et al. (<xref ref-type="bibr" rid="B64">2012</xref>); or, for another perspective, the phonological processing of the HFSC model of Hickok (<xref ref-type="bibr" rid="B27">2012</xref>) distributed over an auditory-motor circuit for syllable and over a somatosensory-motor circuit for the phonemes).</p>
<p>Today, the large majority of the Speech Production Models associate both somatosensory and auditory goals to phonological units (Guenther et al., <xref ref-type="bibr" rid="B24">2006</xref>; Kr&#x000F6;ger et al., <xref ref-type="bibr" rid="B38">2009</xref>; Hickok, <xref ref-type="bibr" rid="B27">2012</xref>; Yan et al., <xref ref-type="bibr" rid="B75">2014</xref>; Parrell et al., <xref ref-type="bibr" rid="B48">2018</xref>). In this context, a key-question is the respective weight of each modality in the specification of the goals. Lindblom (<xref ref-type="bibr" rid="B44">1996</xref>) and Stevens (<xref ref-type="bibr" rid="B68">1996</xref>) considered that the articulatory/somatosensory correlates are not primary, but are rather the secondary consequences of the articulatory strategies that have emerged for a correct achievement of the acoustic/auditory goals. In line with these suggestions, we have assumed a hierarchical organization of the goals, with a higher priority for the achievement of the auditory goals (Perrier, <xref ref-type="bibr" rid="B54">2005</xref>). In its recent versions, the DIVA model assumes that speech acquisition is based on purely auditory targets, and that the somatosensory targets are learned in a second stage during speech development as &#x0201C;sensations associated with the sound currently being produced&#x0201D; (Guenther et al., <xref ref-type="bibr" rid="B24">2006</xref>, p. 286), introducing also a hierarchy in the role of the modalities in the specification of the goals. In an experimental study, in which speech production was perturbed both in the auditory domain (with an on-line shift of formant F1) and in the somatosensory one (with an on-line alteration of the jaw opening, which also affects F1), Feng et al. (<xref ref-type="bibr" rid="B15">2011</xref>) found that participants compensated for the auditory perturbation regardless of the direction of the perturbation of the jaw opening. This observation was in support of a dominant role of the auditory modality in the control of speech production.</p>
<p>However, three important experimental findings have contested the validity of the hierarchical hypothesis. The first finding is the fact that, when the auditory feedback is perturbed, the compensation to the perturbation is never complete, with a magnitude commonly being at the most at 1/3 of the perturbation (Houde and Jordan, <xref ref-type="bibr" rid="B29">2002</xref>; Purcell and Munhall, <xref ref-type="bibr" rid="B59">2006</xref>; Villacorta et al., <xref ref-type="bibr" rid="B73">2007</xref>; Cai et al., <xref ref-type="bibr" rid="B6">2010</xref>). A convincing explanation for this phenomenon is the fact that the strength of the specification of the somatosensory goal limits the authorized magnitude of the articulatory changes used to compensate for the auditory perturbation (Villacorta et al., <xref ref-type="bibr" rid="B73">2007</xref>; Katseff et al., <xref ref-type="bibr" rid="B31">2012</xref>). The second finding is that motor learning associated with a perturbation of the auditory feedback generates a shift of the perceptual boundaries between the phonemes of interest (Shiller et al., <xref ref-type="bibr" rid="B66">2009</xref>; Lametti et al., <xref ref-type="bibr" rid="B40">2014</xref>). Using a simplified Bayesian model of speech production, we have shown that the perceptual boundary shift was also in part due to the strength of the somatosensory goals (Patri et al., <xref ref-type="bibr" rid="B52">2018</xref>). The third finding is the observation of &#x0201C;sensory preference&#x0201D; in a speech production task in which both auditory feedback and jaw movement were perturbed on line (Lametti et al., <xref ref-type="bibr" rid="B39">2012</xref>). Indeed Lametti et al. (<xref ref-type="bibr" rid="B39">2012</xref>) found that contrary to the observations of Feng et al. (<xref ref-type="bibr" rid="B15">2011</xref>) not all the participants did compensate in priority for the auditory perturbation: some of them did compensate more for the auditory perturbation, but some others did compensate more for the jaw perturbation, and a significant negative correlation was found between the amounts of compensation to the perturbation in each modality. This completely changed the way to consider the crucial question of the physical domain in which the speech goals are specified in adults speakers for the production of phonological units. The answer to this question would not be generic and only depending on the characteristics of the language, but would be strongly subject-dependent and related to a preference of the subjects for one feedback modality or the other. From a general linguistic point of view, the debate currently moves toward considering speaker-specific characteristics of the way to deal with the constraints of the language. Developing models of such phenomena will open doors for the elaboration of new experimental paradigms to question how speakers deal with the constraints of their language, and to investigate the consequences on speaker behaviors in terms of adaptation, coarticulation, and possibly diachronic phonetic changes.</p>
<p>In this work, we address the question of the &#x0201C;sensory preference&#x0201D; within a Bayesian model of speech motor planning, in which speech units are characterized both in auditory and somatosensory terms. This approach includes internal models predicting the sensory consequences of motor commands, and the definition of the sensory characterization of the motor goals, also called henceforth &#x0201C;sensory targets,&#x0201D; associated with phonemes. These components are described in terms of probability distributions. We show that sensory preference can be implemented in the model in two ways.</p>
<p>In the first variant, sensory preference is attributed to the relative accuracy measured as the precision (i.e., inverse of variance) of the sensory targets. This is inspired from well-acknowledged models of sensory fusion for perception (Ernst and Banks, <xref ref-type="bibr" rid="B12">2002</xref>; Alais and Burr, <xref ref-type="bibr" rid="B1">2004</xref>; Kersten et al., <xref ref-type="bibr" rid="B34">2004</xref>) and of sensorimotor integration (K&#x000F6;rding and Wolpert, <xref ref-type="bibr" rid="B36">2004</xref>). It corresponds in particular to the approach proposed by the DIVA model (Villacorta et al., <xref ref-type="bibr" rid="B73">2007</xref>; Perkell et al., <xref ref-type="bibr" rid="B53">2008</xref>). In this view, sensory preference originates from the level of the stored sensory targets that are intrinsically associated with phonological units. This suggests that sensory preference would be an inflexible property of each individual. We call this modeling approach &#x0201C;Target-based approach.&#x0201D;</p>
<p>In the second, more original variant, sensory preference is implemented by modulating the sensitivity of the comparison between the predicted sensory consequences of motor commands and the sensory characterization of speech motor goals. This approach differs from linear weightings of the error associated with each modality in the computation of the feedback correction signal (see for example the &#x0201C;synaptic weights&#x0201D; in Guenther et al., <xref ref-type="bibr" rid="B24">2006</xref>, Equation 9, p. 286), because of our probabilistic formulation. Indeed, we will see that the probabilistic formulation enables an interesting interpretation of the variation of sensory preference in terms of &#x0201C;clarity&#x0201D; or &#x0201C;sharpness&#x0201D; of the sensory pathway. Furthermore, in this second view, sensory preference is more flexible, as it can be modified without changing the stored sensory targets. Such a modification can then result from cognitive control, attentional processes or features of the task, without affecting the sensory characterization of speech motor goals associated with phonological units. We call this modeling approach &#x0201C;Comparison-based approach.&#x0201D;</p>
<p>The main purpose of the current study is to compare these two variants, in the context of the adaptation to a long-lasting steady-state external sensory perturbation. As we recalled above, numerous experimental studies have used such a perturbation paradigm, and they have shown that perturbation leads to two kinds of compensation depending on the exposure time to the perturbation: first to an almost immediate change of speech articulation aiming at compensating for the unpredicted newly introduced perturbation; second, after a sufficiently long period in presence of the sustained perturbation, to a long-lasting compensation resulting from adaptation. Adaptation has been shown to induce after-effects (Houde and Jordan, <xref ref-type="bibr" rid="B28">1998</xref>; Tremblay et al., <xref ref-type="bibr" rid="B72">2003</xref>) which has been interpreted as evidence for long-lasting changes in the internal representations of the relations between motor commands and sensory outputs (called internal models in this paper). Thus, it is important to distinguish immediate compensation, associated with instantaneous motor control of speech movements, and compensation resulting from adaptation, associated with changes in the planning of speech movements. In this work we focus on the compensation resulting from adaptation, without considering the dynamics of the learning process underlying the transition from immediate compensation to final adaptation.</p>
<p>This paper is structured as follows. In section 2, we introduce all the elements of the modeling framework. We first describe the GEPPETO model, overall, and detail the Bayesian version of its motor planning layer. Then we explain how we simulate sensory perturbations and how we account for the resulting adaptations. Finally, we describe both variants of our model of sensory preference. In section 3, we simulate the two variants, highlighting their equivalence, which we then analyze formally. Finally, we discuss our results and possible extensions in section 4.</p>
</sec>
<sec id="s2">
<title>2. Methods</title>
<sec>
<title>2.1. Overview of the Framework</title>
<sec>
<title>2.1.1. The GEPPETO Model</title>
<p>GEPPETO (see <xref ref-type="fig" rid="F1">Figure 1</xref>) is a model of speech production organized around four main components: (i) a biomechanical model of the vocal tract simulating the activation of muscles and their influence on the postures and the movements of the main oro-facial articulators involved in the production of speech (Perrier et al., <xref ref-type="bibr" rid="B57">2011</xref>); (ii) a model of muscle force generation mechanisms (the &#x003BB; model, Feldman, <xref ref-type="bibr" rid="B13">1986</xref>) that includes the combined effects on motoneurons&#x00027; depolarization of descending information from the Central Nervous System and afferent information arising via short delay feedback loops from muscle spindles (stretch reflex) or mechano-receptors; (iii) a pure feedforward control system that specifies the temporal variation of the control variables (called &#x003BB; variables) of the &#x003BB; model from the specification of the target values inferred in the motor planning phase and of their timing; and (iv) a motor planning system that infers the target &#x003BB; variables associated with the phonemes of the planned speech sequence.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Schematic representation of the GEPPETO model. <bold>(A)</bold> Overview of the four layers of the GEPPETO model. The red dashed box indicates the planning layer on which we focus in the present work and which is the object of the Bayesian modeling. <bold>(B)</bold> Illustration of phoneme sensory target regions in the model. Top plots: ellipses representing auditory target regions in the (<italic>F</italic><sub>2</sub>, <italic>F</italic><sub>1</sub>) acoustic plane (left) and in the first two PCA dimensions of the somatosensory space (right). Colors enable to visualize the distortion of geometry induced by the non-linearity of the relation between the auditory and somatosensory spaces. Dashed boxes indicate the portion of auditory and somatosensory spaces on which we focus for the results presented in this paper. Bottom plots: probabilistic characterization of phoneme target regions in the Bayesian model as multivariate Gaussian distributions.</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0001.tif"/>
</fig>
<p>In the implementation of GEPPETO used in this study, the biomechanical model is a 2-dimensional finite element model of the tongue in the vocal tract, which includes 6 principal tongue muscles as actuators and accounts for mechanical contacts with the vocal tract boundaries. The motor planning layer specifies the target &#x003BB; variables by considering the motor goals associated with the phonemes of the speech utterance to be produced and using an optimal approach. Complete descriptions of GEPPETO, available elsewhere (Perrier et al., <xref ref-type="bibr" rid="B56">2005</xref>; Winkler et al., <xref ref-type="bibr" rid="B74">2011</xref>; Patri et al., <xref ref-type="bibr" rid="B50">2015</xref>, <xref ref-type="bibr" rid="B51">2016</xref>; Patri, <xref ref-type="bibr" rid="B49">2018</xref>), also involve the specification of intended levels of effort. This enables in particular to perform speech sequences at different speaking rates; however, for simplicity, we do not consider this aspect of the model in the current study.</p>
<p>A key hypothesis in GEPPETO is that speech production is planned on the basis of units having the size of the phonemes. The account for larger speech units is given in the model via optimal planning: larger speech units correspond to the span of the phoneme sequence on which optimal planning applies (CV syllables, CVC syllables, VCV sequences, see Perrier and Ma, <xref ref-type="bibr" rid="B55">2008</xref>; Ma et al., <xref ref-type="bibr" rid="B45">2015</xref>). Given the limitations of the biomechanical model used in this study, which only models the tongue and assumes fixed positions for the jaw and the lips, we only consider French vowels that do not crucially involve jaw or lip movements, which are {/i/, /e/, /&#x0025B;/, /a/, /oe/, /&#x00254;/}. GEPPETO further assumes that the motor goals associated with phonemes are defined as particular target regions in the sensory space. These regions are assumed to describe the usual range of variation of the sensory inputs associated with the production of the phonemes. Previous versions of GEPPETO have only considered the auditory space for the definition of these target regions. The auditory space is identified in GEPPETO to the space of the first three formants (F<sub>1</sub>, F<sub>2</sub>, F<sub>3</sub>) and target regions are defined in this space as dispersion ellipsoids of order 2, whose standard-deviations have been determined from measures provided by phoneme production experiments (Calliope, <xref ref-type="bibr" rid="B7">1984</xref>; Robert-Ribes, <xref ref-type="bibr" rid="B60">1995</xref>; M&#x000E9;nard, <xref ref-type="bibr" rid="B46">2002</xref>) and adapted to the acoustic maximal vowel space of the biomechanical model (Perrier et al., <xref ref-type="bibr" rid="B56">2005</xref>; Winkler et al., <xref ref-type="bibr" rid="B74">2011</xref>). The top left part of <xref ref-type="fig" rid="F1">Figure 1B</xref> represents the projection of these target regions in the (<italic>F</italic><sub>2</sub>, <italic>F</italic><sub>1</sub>) plane.</p>
<p>In the present study, we consider an updated version of GEPPETO that includes both auditory and somatosensory characterizations of the phonemes. We call it &#x0201C;Bayesian GEPPETO,&#x0201D; because the planning layer, which is at the core of the present study, is described with a Bayesian model. In this formulation, the somatosensory space only accounts for tongue proprioception. This account is based on the shape of the tongue contour in the mid-sagittal plane. More specifically, the somatosensory space is defined as the space of the first three Principal Components that model the covariation of the 17 nodes of the tongue contour in the Finite Element tongue mesh in the mid-sagittal plane, when the target &#x003BB; variables vary over a large range of values, which covers all possible realistic tongue shapes associated with vowel productions. In line with the idea that auditory goals are primary in speech acquisition and that somatosensory goals are learned as a consequence of the achievement of the auditory goals (Lindblom, <xref ref-type="bibr" rid="B44">1996</xref>; Stevens, <xref ref-type="bibr" rid="B68">1996</xref>; Guenther et al., <xref ref-type="bibr" rid="B24">2006</xref>), GEPPETO assumes that somatosensory target regions characterizing phonemes are dispersion ellipsoids that approximate the projections of the auditory target regions into the somatosensory space. The top right part in <xref ref-type="fig" rid="F1">Figure 1B</xref> illustrates the somatosensory target regions in the plane of the first two principal components. Data points within increasing elliptical rings in the auditory target regions are plotted with identical colors in the auditory and somatosensory spaces, providing an intuitive idea of the geometry distortion resulting from the non-linear relation between the auditory and the somatosensory space.</p>
<p>For a given phoneme sequence, the goal of the motor planning layer of GEPPETO is to find the &#x003BB; target variables that enable to reach the sensory target regions of the phonemes with the appropriate serial-order. In the most recent developments of GEPPETO, this inverse problem is addressed as an inference question formulated in a Bayesian modeling framework (Patri et al., <xref ref-type="bibr" rid="B50">2015</xref>, <xref ref-type="bibr" rid="B51">2016</xref>). It is on this Bayesian component of GEPPETO that we focus in this work.</p>
</sec>
<sec>
<title>2.1.2. Bayesian Modeling of Speech Motor Planning in GEPPETO</title>
<p>The Bayesian model formulates the key ingredients of the motor planning stage of GEPPETO in a probabilistic framework, where key quantities are represented as probabilistic variables and their relations are represented by probability distributions. It is mathematically based on the theoretical concepts defined in the COSMO model of speech communication (Moulin-Frier et al., <xref ref-type="bibr" rid="B47">2015</xref>; Laurent et al., <xref ref-type="bibr" rid="B41">2017</xref>). In previous works we have described our modeling framework in the context of coarticulation modeling, planning of sequences of phonemes (Patri et al., <xref ref-type="bibr" rid="B50">2015</xref>), and the specification of effort levels for the planning of speech at different speaking rates (Patri et al., <xref ref-type="bibr" rid="B51">2016</xref>). However, these previous implementations of the model only considered auditory goals for the phonemes. A novelty in the present work is the integration of both auditory and somatosensory goals in &#x0201C;Bayesian GEPPETO.&#x0201D; This integration is based on modeling principles that we have recently elaborated in the context of a simplified Bayesian model of speech production (Patri et al., <xref ref-type="bibr" rid="B52">2018</xref>), in the aim to study various potential explanations for the shifts of perceptual boundaries observed after speech motor learning (Shiller et al., <xref ref-type="bibr" rid="B66">2009</xref>; Lametti et al., <xref ref-type="bibr" rid="B40">2014</xref>). Note that for simplicity we focus here only on the production of single phonemes. However, the extension of the present formulation to consider sequences of phonemes as in Patri et al. (<xref ref-type="bibr" rid="B50">2015</xref>) is straightforward.</p>
<p>In the case of single-phoneme planning, &#x0201C;Bayesian GEPPETO&#x0201D; includes eight probabilistic variables, described in <xref ref-type="fig" rid="F2">Figure 2</xref> along with their dependencies. The right hand side of the diagram represents variables involved in the definition of the motor goals associated with phonemes: variable &#x003A6; is the variable representing phoneme identity, variables <italic>A</italic><sub>&#x003A6;</sub> and <italic>S</italic><sub>&#x003A6;</sub> are auditory and somatosensory variables involved in the sensory characterization of phonemes (we call them sensory-phonological variables). The left hand side of the diagram represents variables involved in sensory-motor predictions: the 6-dimensional motor control variable <italic>M</italic> represents the six &#x003BB; variables that control muscle activation and then tongue movements in the biomechanical model (<italic>M</italic> &#x0003D; (&#x003BB;<sub>1</sub>, &#x02026;, &#x003BB;<sub>6</sub>)); variables <italic>A</italic><sub><italic>M</italic></sub> and <italic>S</italic><sub><italic>M</italic></sub> are sensory-motor variables representing the auditory and somatosensory consequences of motor variable <italic>M</italic>.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Diagram describing the Bayesian representation of the motor planning layer in GEPPETO. Nodes represent variables in the model and arrows represent their dependencies. The diagram is a graphical representation of the decomposition of the joint probability distribution given in Equation (1).</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0002.tif"/>
</fig>
<p>Motor planning of a single phoneme is achieved in the model by identifying the sensory-motor predictions that match the sensory specification of the intended phoneme. This matching is imposed with two coherence variables <italic>C</italic><sub><italic>A</italic></sub> and <italic>C</italic><sub><italic>S</italic></sub> (Bessi&#x000E8;re et al., <xref ref-type="bibr" rid="B2">2013</xref>), that act as &#x0201C;probabilistic switches,&#x0201D; and can be understood as implementing a matching constraint between the predicted sensory-motor variables and the specified sensory-phonological variables.</p>
<p>The diagram in <xref ref-type="fig" rid="F2">Figure 2</xref> also represents the decomposition of the joint probability distribution of all the variables in the model:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M2"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mi>M</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Each of the factors on the right hand side of Equation (1) corresponds to one particular piece of knowledge involved in motor planning:
<list list-type="simple">
<list-item><p><italic>P</italic>(<italic>M</italic>) and <italic>P</italic>(&#x003A6;) are prior distributions representing prior knowledge about possible values of motor variable <italic>M</italic> and of phoneme variable &#x003A6;. We assume all possible values to be equally probable (no prior knowledge) and thus define <italic>P</italic>(<italic>M</italic>) and <italic>P</italic>(&#x003A6;) as uniform distributions over their domains. The domain of variable <italic>M</italic> is a continuous 6-dimensional support defined by the allowed range of values of each parameter &#x003BB;<sub><italic>i</italic></sub> of the biomechanical model. &#x003A6; is a discrete, categorical variable including the identity of the different phonemes considered in the model.</p></list-item>
<list-item><p><italic>P</italic>(<italic>A</italic><sub>&#x003A6;</sub> | &#x003A6;) and <italic>P</italic>(<italic>S</italic><sub>&#x003A6;</sub> | &#x003A6;) correspond to the auditory and somatosensory characterizations of phonemes. We define them as multivariate Gaussian distributions in the auditory and somatosensory spaces:
<disp-formula id="E2"><label>(2)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>&#x003A6;</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>:=</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p></list-item>
<list-item><p>where <italic>X</italic> refers to the sensory modality (<italic>A</italic> for &#x0201C;Auditory&#x0201D; or <italic>S</italic> for &#x0201C;Somatosensory&#x0201D;), and <inline-formula><mml:math id="M5"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M6"><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> correspond to the parameters specifying the distribution associated to phoneme &#x003D5; in the sensory space <italic>X</italic> (i.e., mean vector <inline-formula><mml:math id="M7"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and covariance matrix <inline-formula><mml:math id="M8"><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>). This definition of the sensory characterizations translates in probabilistic terms the hypothesis that phonemes are characterized by the ellipsoid regions illustrated in <xref ref-type="fig" rid="F1">Figure 1B</xref>. In particular, the mean vector and covariance matrix of each distribution are identified from these ellipsoid regions. The correspondence between these two representations is illustrated in the top and bottom plots of <xref ref-type="fig" rid="F1">Figure 1B</xref>.</p></list-item>
<list-item><p><italic>P</italic>(<italic>A</italic><sub><italic>M</italic></sub> | <italic>M</italic>) and <italic>P</italic>(<italic>S</italic><sub><italic>M</italic></sub> | <italic>M</italic>) correspond to the knowledge relating the motor control variable <italic>M</italic> to its predicted sensory consequences <italic>A</italic><sub><italic>M</italic></sub> and <italic>S</italic><sub><italic>M</italic></sub>, in the auditory and somatosensory space, respectively. We identify this knowledge to sensory-motor internal models in the brain (Kawato et al., <xref ref-type="bibr" rid="B32">1990</xref>; Jordan and Rumelhart, <xref ref-type="bibr" rid="B30">1992</xref>; Tian and Poeppel, <xref ref-type="bibr" rid="B71">2010</xref>). In the current implementation we assume that these internal models are deterministic and we implement them as Dirac probability distributions centered on the outputs of sensory-motor maps, &#x003C1;<sub><italic>a</italic></sub> and &#x003C1;<sub><italic>s</italic></sub>:
<disp-formula id="E3"><label>(3)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>:=</mml:mo><mml:mi>&#x003B4;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p></list-item>
<list-item><p>where <italic>X</italic><sub><italic>m</italic></sub> stands for <italic>A</italic><sub><italic>M</italic></sub> or <italic>S</italic><sub><italic>M</italic></sub>, depending on the modality, &#x003B4; denotes the Dirac distribution (i.e., <italic>P</italic>([<italic>X</italic><sub><italic>M</italic></sub> &#x0003D; <italic>x</italic>] | [<italic>M</italic> &#x0003D; <italic>m</italic>]) is zero unless <italic>x</italic> &#x0003D; &#x003C1;<sub><italic>x</italic></sub>(<italic>m</italic>)). The sensory-motor maps &#x003C1;<sub><italic>a</italic></sub> and &#x003C1;<sub><italic>s</italic></sub> have been created from the results of around 50,000 simulations carried out with the biomechanical model by randomly sampling the space of the &#x003BB; motor control variables. We implemented these sensory maps by learning the relation between the &#x003BB; variables and the sensory variables with Radial Basis Functions (RBF; Poggio and Girosi, <xref ref-type="bibr" rid="B58">1989</xref>) with a usual supervised learning approach.</p></list-item>
<list-item><p><italic>P</italic>(<italic>C</italic><sub><italic>A</italic></sub> | <italic>A</italic><sub><italic>M</italic></sub><italic>A</italic><sub>&#x003A6;</sub>) and <italic>P</italic>(<italic>C</italic><sub><italic>S</italic></sub> | <italic>S</italic><sub><italic>M</italic></sub><italic>S</italic><sub>&#x003A6;</sub>) implement the two sensory matching constraints. <italic>C</italic><sub><italic>A</italic></sub> and <italic>C</italic><sub><italic>S</italic></sub> are both binary variables (taking values 0 or 1) that activate the corresponding matching constraint when their values are set to 1. This is implemented with the following definition:
<disp-formula id="E4"><label>(4)</label><mml:math id="M10"><mml:mrow><mml:mi>P</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mi>C</mml:mi><mml:mi>X</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>|</mml:mo><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mi>M</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>m</mml:mi></mml:msub><mml:mo stretchy='false'>]</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mo>&#x003A6;</mml:mo></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>&#x003D5;</mml:mi></mml:msub><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x000A0;&#x000A0;</mml:mtext><mml:mo>:=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mtable columnalign='left'><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mn>1</mml:mn></mml:mtd><mml:mtd columnalign='left'><mml:mrow><mml:mtext>if&#x000A0;</mml:mtext><mml:msub><mml:mi>x</mml:mi><mml:mi>m</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mi>&#x003D5;</mml:mi></mml:msub></mml:mrow></mml:mtd></mml:mtr><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mn>0</mml:mn></mml:mtd><mml:mtd columnalign='left'><mml:mrow><mml:mtext>otherwise</mml:mtext><mml:mo>.</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mrow></mml:mrow></mml:math></disp-formula></p></list-item>
<list-item><p>where again <italic>X</italic><sub><italic>M</italic></sub> stands for <italic>A</italic><sub><italic>M</italic></sub> or <italic>S</italic><sub><italic>M</italic></sub>, and <italic>X</italic><sub>&#x003A6;</sub> stands for <italic>A</italic><sub>&#x003A6;</sub> or <italic>S</italic><sub>&#x003A6;</sub>.</p></list-item>
</list></p>
</sec>
<sec>
<title>2.1.3. Motor Planning in the Bayesian Model</title>
<p>The goal of the motor planning layer in GEPPETO is to find values of the motor control variable <italic>M</italic> that correctly make the tongue articulate the intended phoneme. The Bayesian model enables to address this question as an inference question that can be formulated in three ways: (i) by activating only the auditory pathway with [<italic>C</italic><sub><italic>A</italic></sub> &#x0003D; 1]; (ii) by activating only the somatosensory pathway with [<italic>C</italic><sub><italic>S</italic></sub> &#x0003D; 1]; (iii) by activating both the auditory and somatosensory pathways with [<italic>C</italic><sub><italic>A</italic></sub> &#x0003D; 1] and [<italic>C</italic><sub><italic>S</italic></sub> &#x0003D; 1] (we call this the &#x0201C;fusion&#x0201D; planning model). These three planning processes are computed analytically, by applying probabilistic calculus to the joint probability distribution <italic>P</italic>(<italic>MA</italic><sub><italic>M</italic></sub><italic>S</italic><sub><italic>M</italic></sub><italic>A</italic><sub>&#x003A6;</sub><italic>S</italic><sub>&#x003A6;</sub> &#x003A6;<italic>C</italic><sub><italic>A</italic></sub><italic>C</italic><sub><italic>S</italic></sub>) specified by Equation (1). The outcome of these computations for each planning process gives:</p>
<disp-formula id="E5"><label>(5)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x003A6;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0221D;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E6"><label>(6)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x003A6;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0221D;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E7"><label>(7)</label><mml:math id="M14"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x003A6;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x0221D;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where the mathematical symbol &#x0201C;&#x0221D;&#x0201D; means &#x0201C;proportional to.&#x0201D;</p>
<p>Equations (5&#x02013;7) give the probability, according to each of the three planning process, that a given value <italic>m</italic> of the motor control variable <italic>M</italic> will actually produce the intended phoneme &#x003A6;. Practically, in order to have for each planning process a reasonable set of values covering the range of variation of the motor control variable with their probability to correctly produce the intended phoneme, we randomly sampled the space of the motor control variable according to these probability distribution. This sampling was implemented to approximate the probability distributions with a standard Markov Chain Monte Carlo algorithm (MCMC) using Matlab&#x00027;s &#x0201C;mhsample&#x0201D; function. The MCMC algorithm performs a random walk in the control space resulting in a distribution of random samples that converges toward the desired probability distribution. The left panels in <xref ref-type="fig" rid="F3">Figure 3</xref> present the dispersion ellipses of order 2 in the auditory and somatosensory spaces of the result obtained from 2.10<sup>4</sup> random samples, taken from 20 independent sampling runs (after removal of the first 10<sup>3</sup> burn-in samples in each chain), for the production of phoneme /&#x00254;/ for each of the three planning processes. It can be observed that all three planning processes correctly achieve the target region in both sensory spaces.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Results of the three planning processes obtained with the model for the production of phoneme /&#x00254;/, in the auditory space (top panels) and the somatosensory space (bottom panels). Results are presented in three conditions: in unperturbed condition (left panels); with auditory perturbation before adaptation (middle panels); and with auditory perturbation once adaptation has been achieved (right panels). Black ellipses indicate the phoneme target regions (see <xref ref-type="fig" rid="F1">Figure 1B</xref>). Colored ellipses present results as dispersion ellipses of order 2 obtained from 2.10<sup>4</sup> samples for each of the three planning processes: auditory planning in red, somatosensory planning in blue and fusion planning in green.</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0003.tif"/>
</fig>
</sec>
</sec>
<sec>
<title>2.2. Implementation of Sensory Perturbations and Adaptation in the Model</title>
<p>Sensory perturbations alter the sensed consequence of motor actions such that the sensory output predicted by the internal model becomes erroneous. When the perturbation is consistently maintained, a new relation between motor control variables and sensory outputs is experienced and the sensory-motor internal models (<italic>P</italic>(<italic>A</italic><sub><italic>M</italic></sub> | <italic>M</italic>) and <italic>P</italic>(<italic>S</italic><sub><italic>M</italic></sub> | <italic>M</italic>)) are updated as a result of motor learning and adaption (Shadmehr and Mussa-Ivaldi, <xref ref-type="bibr" rid="B65">1994</xref>; Houde and Jordan, <xref ref-type="bibr" rid="B28">1998</xref>; Haruno et al., <xref ref-type="bibr" rid="B26">1999</xref>; Tremblay et al., <xref ref-type="bibr" rid="B72">2003</xref>), in order to capture the new sensory-motor relation imposed by the perturbation. We define adaptation, in the model, as the update of the parameters of the internal models.</p>
<p>According to Lametti et al. (<xref ref-type="bibr" rid="B39">2012</xref>), differences in sensory preference lead to differences across speakers in their tolerance to errors in each of the sensory modalities (auditory or somatosensory). This phenomenon has been assumed to explain the observed inter-speaker differences in the amount of compensation after adaptation. The evaluation of our two implementations of sensory preference is based on their capacity to account for these differences in compensation. Importantly, whatever the nature of the sensory perturbation (auditory or somatosensory), compensation induces changes in both the auditory and somatosensory outputs, generating errors in both domains. Hence, the amount of compensation is modulated by sensory preference even if the perturbation affects only one sensory modality. Therefore in this paper, for the sake of simplicity, we only consider auditory perturbations (but see Patri, <xref ref-type="bibr" rid="B49">2018</xref> for results involving somatosensory perturbations).</p>
<sec>
<title>2.2.1. Implementation of Sensory Perturbations</title>
<p>We simulate auditory perturbations in the model by altering the spectral characteristic of the acoustic signal associated with the tongue configurations of the biomechanical model. More specifically, if a tongue configuration <italic>T</italic> produced an acoustic output <italic>a</italic><sup><italic>u</italic></sup> in unperturbed condition, then with the auditory perturbation the same tongue configuration will result in a shifted acoustic output <italic>a</italic><sup>&#x0002A;</sup> &#x0003D; <italic>a</italic><sup><italic>u</italic></sup>&#x0002B;&#x003B4;. The middle panel of <xref ref-type="fig" rid="F3">Figure 3</xref> illustrates the effect of an auditory perturbation that shifts the first formant F1 down by &#x003B4; &#x0003D; &#x02212;100 Hz, during the production of vowel /&#x00254;/ for the three planning processes.</p>
</sec>
<sec>
<title>2.2.2. Implementation of Adaptation</title>
<p>In the context of an auditory perturbation, only the auditory-motor internal model <italic>P</italic>(<italic>A</italic><sub><italic>M</italic></sub> | <italic>M</italic>) becomes erroneous. Hence, we implement adaptation to the auditory perturbation by updating the auditory-motor map &#x003C1;<sub><italic>a</italic></sub> of the auditory-motor internal model <italic>P</italic>(<italic>A</italic><sub><italic>M</italic></sub> | <italic>M</italic>) (see Equation 3). This update is defined in order to capture the new relation between the motor control variable and its auditory consequence. In the case of an auditory perturbation that shifts auditory values by a constant vector &#x003B4;, we assume the resulting update to be complete and perfect, of parameter &#x003B4;<sub><italic>A</italic></sub> &#x0003D; &#x003B4;:</p>
<disp-formula id="E8"><label>(8)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B4;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <inline-formula><mml:math id="M16"><mml:msubsup><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>*</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M17"><mml:msubsup><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> denote the auditory-motor maps in the perturbed and unperturbed condition, respectively. In all simulations involving the perturbation, we choose to shift only the first formant F1 down by &#x02212;100 Hz, such that &#x003B4;<sub><italic>A</italic></sub> &#x0003D; [&#x02212;100, 0, 0].</p>
<p>The right panel of <xref ref-type="fig" rid="F3">Figure 3</xref> illustrates the effect of the auditory perturbation and the outcome of adaptation for each of the three planning processes. In unperturbed conditions (left panels), all three planning processes correctly achieve both the auditory and the somatosensory target regions. In the middle panel, which represents the situation before adaptation occurs, the auditory perturbation induces for the three planning processes a shift in the auditory domain (top middle panel), and obviously not in the somatosensory domain (bottom middle panel), since the perturbation only alters the auditory-motor relations. The right panels illustrate the outcome of the three planning processes after adaptation has been achieved, as implemented by Equation (8). It can be seen that the results corresponding to the somatosensory planning, <italic>P</italic>(<italic>M</italic> | &#x003A6; [<italic>C</italic><sub><italic>S</italic></sub> &#x0003D; 1]), remain unchanged. This is because somatosensory planning does not involve the auditory-motor map &#x003C1;<sub><italic>a</italic></sub> (Equation 6), and is then not concerned by the update of the auditory-motor map induced by the adaptation. On the other hand, and as expected, after the perfect update of the auditory-motor internal model, the auditory planning <italic>P</italic>(<italic>M</italic> | &#x003A6; [<italic>C</italic><sub><italic>A</italic></sub> &#x0003D; 1]) (Equation 5) fully compensates for the perturbation and results in a correct reaching of the auditory target region (top right panel). However, this compensation is achieved by a change in the value of the motor control variable, which results in a tongue posture associated with a somatosensory output that is outside of the somatosensory target region (bottom right panel). Finally, the fusion planning <italic>P</italic>(<italic>M</italic> | &#x003A6; [<italic>C</italic><sub><italic>A</italic></sub> &#x0003D; 1] [<italic>C</italic><sub><italic>S</italic></sub> &#x0003D; 1]) (Equation 7) combines the two previous results: since auditory and somatosensory target regions are no more compatible due to the update of the auditory-motor internal model, fusion planning cannot reach both sensory target regions at the same time, and therefore it makes a compromise between the auditory and the somatosensory constraints. As a result, fusion planning leads to auditory and somatosensory consequences that lie midway between those of a pure auditory or a pure somatosensory planning.</p>
<p>In summary, we have described how the three planning processes achieve similar results in unperturbed condition but generate very different results after adaptation to the sensory perturbation. Intuitively, if we are able to modulate in the model the weight associated with each sensory modality in the fusion planning process, we would be able to achieve a continuum of compensation magnitudes after adaptation. This continuum, representing all the possible patterns of sensory preference, would go from full compensation for the auditory perturbation, when sensory preference induces a full reliance on the auditory modality, to no compensation at all when sensory preference induces a full reliance on the somatosensory modality.</p>
<p>For the evaluation of the two variants of our model of sensory preference, we mainly consider the &#x0201C;fusion&#x0201D; planning, as it is the planning process that combines both auditory and somatosensory pathways, and then enables an account of the sensory preference phenomenon (see Equation 7). However, we will also study the planning processes based on each sensory pathway individually, in order to have them as reference to evaluate the consequences of different sensory preference patterns. The impact of sensory preference on planning will be evaluated by modulating the relative involvement of each sensory pathway in the planning process. In general terms, the involvement of a sensory pathway is related to the magnitude of the mismatch between sensory-motor predictions and the intended target: for example, by increasing the magnitude of this mismatch for the auditory modality we obtain an increase of the involvement of auditory pathway in the planning process.</p>
</sec>
</sec>
<sec>
<title>2.3. Modeling Sensory Preference</title>
<sec>
<title>2.3.1. The Target-Based Approach: Modulating the Precision of Sensory Targets</title>
<p>In the Target-based approach we modulate the involvement of each sensory modality at the level of the target regions associated with phonemes, as illustrated in the left panel of <xref ref-type="fig" rid="F4">Figure 4</xref>. In our model, the target regions result from the sensory characterization of phonemes which is represented by the terms <italic>P</italic>(<italic>A</italic><sub>&#x003A6;</sub> | &#x003A6;) and <italic>P</italic>(<italic>S</italic><sub>&#x003A6;</sub> | &#x003A6;). These terms are specified in Equation (2) as multivariate Gaussian probability distributions with mean vectors <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and covariance matrices <inline-formula><mml:math id="M20"><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M21"><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula>, respectively. We implement sensory preference in the model by modulating the precision of these distributions with the introduction of two additional parameters, respectively &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub> for the auditory and the somatosensory pathway. These parameters multiply the covariance matrices of the corresponding Gaussian distributions:</p>
<disp-formula id="E9"><label>(9)</label><mml:math id="M22"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo stretchy="true">|</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>&#x003A6;</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003BA;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>X</italic>, once more, stands either for the auditory or the somatosensory modality. The left panel of <xref ref-type="fig" rid="F4">Figure 4</xref> illustrates the effect of parameters &#x003BA;<sub><italic>X</italic></sub> on the target distributions in a one-dimensional case: increasing &#x003BA;<sub><italic>X</italic></sub> results in widening the distribution, and as suggested previously this induces a decrease of the involvement of the corresponding sensory modality in the planning process, since larger distributions will less penalize sensory signals that depart from the center of the target region and will thus allow larger errors in this sensory modality. The same reasoning applies to a decrease of &#x003BA;<sub><italic>X</italic></sub>, which will induce a narrowing of the distribution and an increase of the involvement of the corresponding sensory modality.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>(A)</bold> Illustration of the effect in the Target-based approach of parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub> (see text) on the auditory and somatosensory target regions associated with phonemes <italic>P</italic>(<italic>A</italic><sub>&#x003A6;</sub> | &#x003A6;) and <italic>P</italic>(<italic>S</italic><sub>&#x003A6;</sub> | &#x003A6;). The greater the value of &#x003BA; parameter, the wider the target region, and the weaker the contribution of the corresponding sensory pathway to the planning process. <bold>(B)</bold> Results of the fusion planning process after adaptation to the auditory perturbation described in section 2.2.2, for different values of parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub>.</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0004.tif"/>
</fig>
<p>Replacing the forms given by Equation (9) into Equation (7) gives a first formulation of the influence of sensory preference in the fusion planning process:</p>
<disp-formula id="E10"><label>(10)</label><mml:math id="M24"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mo>&#x003A6;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mo>&#x0221D;</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003BA;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>&#x003BA;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
</sec>
<sec>
<title>2.3.2. The Comparison-Based Approach: Modulating the Weight of the Sensory Matching Constraints</title>
<p>In the Comparison-based approach we modulate the involvement of each sensory modality at the level of the comparison between sensory-motor predictions and sensory characterizations of phonemes, as illustrated on the left panel of <xref ref-type="fig" rid="F5">Figure 5</xref>. To do so, we have to slightly modify the definition of the operator that performs the comparison, i.e., the sensory matching constraint defined in Equation (4). Until now we have defined the sensory matching constraint in an &#x0201C;all-or-nothing&#x0201D; manner, where terms are either &#x0201C;1&#x0201D; when values of the variable predicted with the sensory-motor map match exactly the sensory-phonological variables, or &#x0201C;0&#x0201D; when they differ, regardless of the magnitude of the difference (see Equation 4). This definition is very strict, as it requires an extreme accuracy in the achievement of the speech motor task in the sensory domain. Intuitively, if we are able to soften this constraint, we may be able to modulate the strengths of the comparisons and hence the involvement of each sensory pathway in the planning process.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><bold>(A)</bold> Illustration of the effect in the Comparison-based approach of parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub> on their corresponding sensory matching constraints. The smaller the value of &#x003B7;, the sharper the constraint function and the stronger the relative contribution of the corresponding sensory pathway to the planning process. <bold>(B)</bold> Results of the fusion planning process after adaptation to the auditory perturbation described in section 2.2.2, for different values of parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub>.</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0005.tif"/>
</fig>
<p>We relax the sensory-matching constraint by extending its definition given in Equation (4) as follows (Bessi&#x000E8;re et al., <xref ref-type="bibr" rid="B2">2013</xref>):</p>
<disp-formula id="E11"><label>(11)</label><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Here <italic>d</italic><sub><italic>X</italic></sub>(<italic>x</italic><sub>1</sub>, <italic>x</italic><sub>2</sub>) is a distance measure between sensory values <italic>x</italic><sub>1</sub> and <italic>x</italic><sub>2</sub>. Since <italic>e</italic><sup>&#x02212;<italic>x</italic></sup> is a decreasing continuous function of <italic>x</italic>, the function defined in Equation (11) gives high probability of matching for <italic>x</italic><sub>1</sub> and <italic>x</italic><sub>2</sub> values that are close (small distance <italic>d</italic><sub><italic>X</italic></sub>(<italic>x</italic><sub>1</sub>, <italic>x</italic><sub>2</sub>)) and low probability of matching for values that are far from each other. Note that the definition given in Equation (4) can be considered to be a degenerate case of this new expression of the sensory-matching constraint, in which the distance measure would be zero when <italic>x</italic><sub>1</sub> &#x0003D; <italic>x</italic><sub>2</sub> and infinite otherwise. For computational reasons, we choose a distance measure that is quadratic, i.e., <inline-formula><mml:math id="M26"><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. This choice enables to obtain a closed analytic form for the derivation of the motor planning question.</p>
<p>With this new expression of the matching constraint, we implement sensory preference in the model by introducing two additional parameters, respectively &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub>, for the auditory and the somatosensory pathway. These parameters modulate the sensitivity of the distance measures <italic>d</italic><sub><italic>A</italic></sub>(<italic>a</italic><sub>1</sub>, <italic>a</italic><sub>2</sub>) and <italic>d</italic><sub><italic>S</italic></sub>(<italic>s</italic><sub>1</sub>, <italic>s</italic><sub>2</sub>) associated with the sensory pathways:</p>
<disp-formula id="E12"><label>(12)</label><mml:math id="M27"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>With this choice of parametric quadratic measure, Equation (11) becomes:</p>
<disp-formula id="E13"><label>(13)</label><mml:math id="M28"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p><xref ref-type="fig" rid="F5">Figure 5A</xref> illustrates the form of the matching constraint defined by Equations (13) in the Comparison-based approach for different values of parameter &#x003B7;<sub><italic>X</italic></sub>: small values of &#x003B7;<sub><italic>X</italic></sub> lead to sharper matching constraints; large values lead to flatter constraints. Note in particular that for &#x003B7;<sub><italic>X</italic></sub> &#x02192; 0 the rigid constraint formulated in Equation (4) is recovered, while for &#x003B7;<sub><italic>X</italic></sub> &#x02192; &#x0002B;&#x0221E; the constraint function becomes constant, independent of the sensory values, which in fact corresponds to an absence of constraint.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>3. Results</title>
<sec>
<title>3.1. Simulating Sensory Preference</title>
<sec>
<title>3.1.1. Simulation of the Target-Based Approach</title>
<p>We now illustrate results of simulations using the Target-based approach to model sensory preference in the context of the adaptation to the auditory perturbation described above in section 2.2.2. The colored triangles in <xref ref-type="fig" rid="F4">Figure 4</xref> present the mean results computed for different values of parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub> based on 2.10<sup>4</sup> samples in the motor control space. For reference, colored ellipses present the results obtained with the three planning processes of the previous Section [i.e., purely auditory (red color), purely somatosensory (blue color), or &#x0201C;fusion&#x0201D; planning (intermediate color)].</p>
<p>It can be seen that, as expected, progressively increasing parameter &#x003BA;<sub><italic>A</italic></sub> leads to results that progressively drift toward the outcome of the pure somatosensory planning process. Similar results are obtained toward the outcome of the pure auditory planning when progressively increasing &#x003BA;<sub><italic>S</italic></sub>. Hence, parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub> effectively modulate the strength of each sensory pathway. This confirms the possibility of implementing sensory preference in our model in a way similar to previous approaches: modulating the relative precision of sensory target regions effectively modulates the contribution of the corresponding sensory pathway.</p>
</sec>
<sec>
<title>3.1.2. Simulation of the Comparison-Based Approach</title>
<p>We now illustrate the Comparison-based approach to model sensory preference, and study the effect of parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub> in the model in the context of the adaptation to the auditory perturbation described above in section 2.2.2. The colored triangles in <xref ref-type="fig" rid="F5">Figure 5</xref> present the mean results computed for different values of parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub> based on 2.10<sup>4</sup> samples in the motor control space. As in <xref ref-type="fig" rid="F4">Figure 4</xref>, colored ellipses present the results obtained with the three initial planning processes, for reference.</p>
<p>It can be seen that progressively increasing parameter &#x003B7;<sub><italic>A</italic></sub> of the auditory matching constraint leads to results that progressively drift toward the outcome of the somatosensory planning process. Similarly increasing parameter &#x003B7;<sub><italic>S</italic></sub> of the somatosensory matching constraint results in a drift toward the outcome of the auditory planning process. Hence, parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub> successfully enable to modulate the strength of the constraint imposed by the corresponding sensory pathways.</p>
</sec>
</sec>
<sec>
<title>3.2. Equivalence of the Approaches</title>
<p>We have formulated two alternative approaches to implement sensory preference in Bayesian GEPPETO. Although these approaches account for clearly different ways to process sensory variables, simulations with the model have shown that they lead to qualitatively similar results (right panels of <xref ref-type="fig" rid="F4">Figures 4</xref>, <xref ref-type="fig" rid="F5">5</xref>). Increasing parameter &#x003BA;<sub><italic>A</italic></sub> or parameter &#x003B7;<sub><italic>A</italic></sub> decreases in a comparable manner the involvement of the auditory modality in the model, and, thus, the magnitude of the changes induced by the compensation for the auditory perturbation. Thus, at the limit, for very large values of &#x003BA;<sub><italic>A</italic></sub> or &#x003B7;<sub><italic>A</italic></sub>, the magnitude of the compensation for the auditory perturbation tends toward zero, which perfectly matches the results of the pure somatosensory planning process. Conversely, increasing parameter &#x003BA;<sub><italic>S</italic></sub> or parameter &#x003B7;<sub><italic>S</italic></sub> decreases the involvement of the somatosensory modality and induces an increase of the magnitude of the compensation for the auditory perturbation. At the limit, for very large values of &#x003BA;<sub><italic>S</italic></sub> or &#x003B7;<sub><italic>S</italic></sub>, the magnitude of the compensation tends toward the magnitude obtained with the pure auditory planning process.</p>
<p>However, a closer comparison of the results presented in the right panels of <xref ref-type="fig" rid="F4">Figures 4</xref>, <xref ref-type="fig" rid="F5">5</xref> reveals differences in the ways the compensation for the auditory perturbation varies when parameters &#x003BA;<sub><italic>X</italic></sub> or &#x003B7;<sub><italic>X</italic></sub> vary. In the Target-based approach, the sequence of compensatory results follows a slightly more simple and straight path than in the Comparison-based approach.</p>
<p>Despite these slight differences, the qualitative similarity of the results obtained with both approaches can be formally explained. Indeed, let us consider the outcome of the fusion planning <italic>P</italic>([<italic>M</italic> &#x0003D; <italic>m</italic>] | &#x003A6; [<italic>C</italic><sub><italic>A</italic></sub> &#x0003D; 1] [<italic>C</italic><sub><italic>S</italic></sub> &#x0003D; 1]) using the generalized sensory matching constraints given by Equation (11) in the Comparison-based approach. It yields:</p>
<disp-formula id="E14"><label>(14)</label><mml:math id="M30"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mo>&#x003A6;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0221D;</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>|</mml:mo><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>|</mml:mo><mml:mo>&#x003A6;</mml:mo></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo stretchy="false">|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where we have omitted intermediate steps for the sake of brevity. Now, using the definition of sensory targets given in Equation (2) and the quadratic distance in the matching constraints as given in Equation (13), we note that all terms on the right hand side of Equation (14) are Gaussian. Hence, we can rewrite Equation (14) as:</p>
<disp-formula id="E15"><label>(15)</label><mml:math id="M32"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0221D;</mml:mo><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>a</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>s</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where we have denoted by <italic>I</italic><sub><italic>A</italic></sub> and <italic>I</italic><sub><italic>S</italic></sub> the identity matrices in the auditory and somatosensory space, respectively. With the introduction of variable <italic>y</italic> &#x0003D; &#x003C1;<sub><italic>x</italic></sub>(<italic>m</italic>)&#x02212;<italic>x</italic><sub>&#x003A6;</sub>, each of the sums in Equation (15) are in fact the convolution of two Gaussian distributions, one with mean <inline-formula><mml:math id="M33"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and covariance <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula>, the other of mean 0 and covariance <inline-formula><mml:math id="M35"><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>X</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. The convolution of two Gaussian distributions with mean vectors &#x003BC;<sub>1</sub>, &#x003BC;<sub>2</sub> and covariances &#x003A3;<sub>1</sub>, &#x003A3;<sub>2</sub> is known to result in another Gaussian distribution with mean vector &#x003BC;<sub>1</sub>&#x0002B;&#x003BC;<sub>2</sub> and covariance &#x003A3;<sub>1</sub>&#x0002B;&#x003A3;<sub>2</sub>. Hence, the planning process becomes:</p>
<disp-formula id="E16"><label>(16)</label><mml:math id="M37"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>M</mml:mi><mml:mo>=</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x003A6;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtable><mml:mtr><mml:mtd><mml:mo>&#x0221D;</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>S</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="-tex-caligraphic">N</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003C1;</mml:mi></mml:mrow><mml:mrow><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x003A6;</mml:mo></mml:mrow></mml:msubsup><mml:mo>&#x0002B;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mi>I</mml:mi></mml:mrow><mml:mrow><mml:mi>A</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Let us compare Equation (16) and Equation (10): they are almost identical, except for the form of the covariance matrices in auditory and somatosensory spaces. The planning process in the Target-based approach (Equation 10) involves Gaussian distributions with covariance matrices that are modulated multiplicatively by the parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub>, whereas the planning process in the Comparison-based approach (Equation (16)) involves Gaussian distributions with covariance matrices that are modulated additively by parameters &#x003B7;<sub><italic>A</italic></sub> and &#x003B7;<sub><italic>S</italic></sub>. Hence, the effect of parameters &#x003B7;<sub><italic>X</italic></sub> and &#x003BA;<sub><italic>X</italic></sub> are qualitatively similar, as we have illustrated experimentally: they both induce an increase in the covariance of the sensory characterization of phonemes. However, quantitatively, we have shown that parameters &#x003BA;<sub><italic>X</italic></sub> increase them multiplicatively, whereas parameters &#x003B7;<sub><italic>X</italic></sub> increase them additively.</p>
<p>We note that if the auditory and somatosensory spaces would be one-dimensional, both approaches would be exactly equivalent, since any additive increase &#x00393;&#x0002B;&#x003B7; can be written as a multiplicative increase &#x003BA;&#x00393;, with <inline-formula><mml:math id="M38"><mml:mi>&#x003BA;</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x003B7;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x00393;</mml:mo></mml:mrow></mml:mfrac></mml:math></inline-formula>. This is not true anymore in higher dimensions though, since the Target-based approach scales all coefficients of the covariance matrices, whereas the Comparison-based approach only modifies their diagonal terms. More specifically, the Target-based approach increases the size of the target regions while preserving their orientation, whereas the Comparison-based approach stretches the regions along the coordinate axes, inducing a progressive alignment of the main axes of the target regions with the coordinate axes (off-diagonal terms in the covariance matrices become negligible compared to the increased diagonal terms, and the resulting ellipsoid regions progressively lose their orientations). We assume that the slight differences observed above in the consequences on compensation of progressive variations of the &#x003BA;<sub><italic>X</italic></sub> and &#x003B7;<sub><italic>X</italic></sub> parameters find their origins in these changes in target orientations.</p>
<p><xref ref-type="fig" rid="F6">Figure 6</xref> gives an intuitive interpretation of the equivalence of these two approaches. On the one hand, the Target-based approach directly modulates the size of the target regions, while keeping their orientations, as illustrated on the left lens of the glasses in <xref ref-type="fig" rid="F6">Figure 6</xref>. On the other hand, the Comparison-based approach does not change the targets, but modifies the precision of the comparison of the target with the sensory-motor predictions. This is as if the target were seen through a blurring lens, that would &#x0201C;spread&#x0201D; the borders of the target, making it appear bigger. This &#x0201C;blurring effect&#x0201D; is induced by the convolution of the target with a Gaussian term that acts as noise (Equation 15). The larger the value of parameter &#x003B7;<sub><italic>X</italic></sub>, the larger the power of the noise, and the stronger the &#x0201C;blurring&#x0201D; of the target.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Illustrative interpretation of the equivalence between the two implementations of sensory preference. In the Target-based approach (left part of the figure) the size of the auditory or somatosensory target regions are directly modified with parameters &#x003BA;<sub><italic>A</italic></sub> and &#x003BA;<sub><italic>S</italic></sub>. In the Comparison-based approach (right part of the figure) parameters &#x003B7;<sub><italic>S</italic></sub> and &#x003B7;<sub><italic>A</italic></sub> modulate the sensitivity of the corresponding sensory matching constraint, as if target regions were &#x0201C;blurred,&#x0201D; making them appear larger.</p></caption>
<graphic xlink:href="fpsyg-10-02339-g0006.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>The main contribution of our work is to present two different approaches implementing sensory preference in a speech production model that integrates both the auditory and the somatosensory modality. This is done in the context of our Bayesian GEPPETO model for speech motor planning and speech motor control (Perrier et al., <xref ref-type="bibr" rid="B56">2005</xref>; Patri et al., <xref ref-type="bibr" rid="B51">2016</xref>; Patri, <xref ref-type="bibr" rid="B49">2018</xref>), which specifies both auditory and somatosensory constraints to infer motor commands for the production of a given phoneme. We have implemented sensory preference in this model by modulating the relative involvement of sensory modalities with two different approaches: (1) the Target-based approach, which modulates the precision of auditory and somatosensory target regions; (2) the Comparison-based approach, which modulates the sensory-matching constraints between predictions from internal models and sensory target regions. At the core of the evaluation of the two approaches, we have considered the phenomenon of incomplete compensation for sensory perturbations in speech production and its inter-subject variability, which has been evidenced by several experimental studies. Although conceptually different, we have shown in our model that these two approaches are able to account for incomplete compensation variability under the same amount of change in the internal model resulting from adaptation. Furthermore, we have demonstrated the mathematical equivalence of the two approaches in some specific cases, which explains the qualitative similarity of results obtained under both approaches.</p>
<p>In this context, the main outstanding question is whether the two modeling variants are distinguishable. We consider two aspects of this issue: mathematical formulation and experimental evaluation.</p>
<p>Let us compare the mathematical formulations of the two approaches. The Comparison-based approach is less compact and contains more degrees-of-freedom than the Target-based approach. We have also demonstrated that, under certain assumptions, both models behave similarly. On parsimony grounds, then, the Target-based approach certainly wins over the Comparison-based approach. On the other hand the additional degrees of freedom enable the Comparison-based approach to be more flexible.</p>
<p>For further experimental evaluation we consider two possible directions. First, our simulation results illustrate that the particular pattern of partial compensation obtained under both approaches slightly differ. Whether and how these differences could be assessed experimentally is an open question. The main difficulty arises from the fact that the observed differences in partial compensation do not only depend on differences in compensation mechanisms induced by each approach, but also on speaker specific relations between motor commands and sensory variables. Taking into account these speaker specific characteristics would be the main challenge in this experimental evaluation.</p>
<p>The second direction for experimental evaluation, would be related to the different flexibility associated with each approach. Whereas the Target-based approach would predict fixed compensation strategies, ascribing any remaining variability to causes unrelated to sensory preferences or measurement errors, the Comparison-based approach would potentially relate sensory preference with some aspects of the structure of the observed variability. Furthermore, experimentally induced effects (e.g., asking subjects, for a given trial block, to focus especially on somatosensation; introducing a dual-task condition to induce attentional load, etc.) could help discriminating between the predictions of the two models.</p>
<p>Overall, the results of our study provide a new contribution to the understanding of the sensory preference phenomenon. They highlight that two factors could influence sensory preference, that mostly differ by their temporal stability. On the one hand, the Target-based approach represents sensory preference as the precision of target regions. This suggests that sensory preference is learned through language interaction and is stable over time, as the target regions would be used during everyday speech planning. On the other hand, the Comparison-based approach represents sensory preference &#x0201C;elsewhere&#x0201D; in the model, so that it can mathematically be manipulated independently of sensory target regions. Indeed, in this second approach, we have explicitly considered two independent components: (1) the sensory characterization of phonemes, which are mathematically characterized as constraints via the specification of sensory target regions; (2) matching-constraints, which modulate the precision with which sensory predictions from the internal models are compared with phoneme related sensory target regions. This allows a more general and flexible model, as compared to the Target-based approach. This flexibility suggests ways in which sensory preference would be modulated by cognitive control or attentional processes. Such an attentional model would explicitly modulate on the fly sensory preference depending on the context. This modulation could arise, for example, from changes in the access to one of the sensory modality due to disorders, aging, or noise, or from the absence of congruence between the two sensory pathways. A proposal for such an attentional model, as an extension of the Comparison-based model presented here, is outlined in <xref ref-type="supplementary-material" rid="SM1">Supplementary Material</xref>.</p>
<p>Finally, we turn to possible theoretical extensions and applications of our model. So far, the Comparison-based approach of sensory preference we have described here is constrained by the specific hypotheses of the Bayesian-GEPPETO model in which it is included. For instance, it only concerns sensory preference between somatosensory and acoustic descriptions of targets during serial order planning of sequences of vocalic speech sounds. Of course, the application scope could be extended, e.g., toward sensory preference during movement execution and movement correction, with a finer temporal resolution than we have considered so far. This would for instance allow to study time-varying sensory preference, or sensory preference that depends on speech sounds. Indeed, it is an open question whether consonant and vocalic sounds would differ on the sensory pathway they more precisely rely on. We could also consider using our Comparison-based architecture for describing how low-level sensory acuity would affect the learning of the target representations, and how different sensory preference during this learning would result in different sizes and separations of targets in each sensory pathway. Finally, such a learning mechanism with individual-specific sensory preference could contribute to the emergence of learned idiosyncrasies.</p>
<p>Furthermore, to put our approach in a wider theoretical context, we observe that the Comparison-based approach has a structure that could be cast into the general predictive coding framework, as popularized recently by the free-energy principle proposal (Friston and Kiebel, <xref ref-type="bibr" rid="B18">2009</xref>; Feldman and Friston, <xref ref-type="bibr" rid="B14">2010</xref>; Friston, <xref ref-type="bibr" rid="B17">2010</xref>). Indeed, even though our model does not represent time or time-delays specifically, it nevertheless features the idea that &#x0201C;predictions&#x0201D; from internal models would be compared with sensory targets. We note that this is not exactly the same situation as for a comparison between forward predictions and sensory feedback, as would be used for instance in models of trajectory monitoring; nevertheless, the architecture is similar. In the Comparison-based approach, we have proposed a mathematically specific expression of the &#x0201C;comparison&#x0201D; operator, using probabilistic coherence variables and match measures. Whether this would be a plausible, or at least useful mathematical implementation of probabilistic comparison in predictive coding or free-energy architectures is an open question.</p>
</sec>
<sec sec-type="data-availability" id="s5">
<title>Data Availability Statement</title>
<p>The datasets generated for this study are available on request to the corresponding author.</p>
</sec>
<sec id="s6">
<title>Author Contributions</title>
<p>J-FP, JD, and PP contributed conception and design of the study, and revised the manuscript. J-FP implemented the model and performed simulations, and wrote the first draft of the manuscript. All authors contributed to manuscript revision, read, and approved the submitted version.</p>
<sec>
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack><p>Authors wish to thank Jean-Luc Schwartz, Pierre Bessi&#x000E8;re, and Jacques Droulez for inspiring discussions and support.</p>
</ack>
<sec sec-type="supplementary-material" id="s8">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fpsyg.2019.02339/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fpsyg.2019.02339/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alais</surname> <given-names>D.</given-names></name> <name><surname>Burr</surname> <given-names>D.</given-names></name></person-group> (<year>2004</year>). <article-title>The ventriloquist effect results from near-optimal bimodal integration</article-title>. <source>Curr. Biol.</source> <volume>14</volume>, <fpage>257</fpage>&#x02013;<lpage>262</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2004.01.029</pub-id><pub-id pub-id-type="pmid">14761661</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bessi&#x000E8;re</surname> <given-names>P.</given-names></name> <name><surname>Mazer</surname> <given-names>E.</given-names></name> <name><surname>Ahuactzin</surname> <given-names>J. M.</given-names></name> <name><surname>Mekhnacha</surname> <given-names>K.</given-names></name></person-group> (<year>2013</year>). <source>Bayesian Programming</source>. <publisher-loc>Boca Raton, FL</publisher-loc>: <publisher-name>CRC Press</publisher-name>. <pub-id pub-id-type="doi">10.1201/b16111</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Blumstein</surname> <given-names>S. E.</given-names></name> <name><surname>Stevens</surname> <given-names>K. N.</given-names></name></person-group> (<year>1979</year>). <article-title>Acoustic invariance in speech production: evidence from measurements of the spectral characteristics of stop consonants</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>66</volume>, <fpage>1001</fpage>&#x02013;<lpage>1017</lpage>. <pub-id pub-id-type="doi">10.1121/1.383319</pub-id><pub-id pub-id-type="pmid">512211</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Browman</surname> <given-names>C. P.</given-names></name> <name><surname>Goldstein</surname> <given-names>L.</given-names></name></person-group> (<year>1989</year>). <article-title>Articulatory gestures as phonological units</article-title>. <source>Phonology</source> <volume>6</volume>, <fpage>201</fpage>&#x02013;<lpage>251</lpage>. <pub-id pub-id-type="doi">10.1017/S0952675700001019</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Browman</surname> <given-names>C. P.</given-names></name> <name><surname>Goldstein</surname> <given-names>L.</given-names></name></person-group> (<year>1992</year>). <article-title>Articulatory phonology: an overview</article-title>. <source>Phonetica</source> <volume>49</volume>, <fpage>155</fpage>&#x02013;<lpage>180</lpage>. <pub-id pub-id-type="doi">10.1159/000261913</pub-id><pub-id pub-id-type="pmid">1488456</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cai</surname> <given-names>S.</given-names></name> <name><surname>Ghosh</surname> <given-names>S. S.</given-names></name> <name><surname>Guenther</surname> <given-names>F. H.</given-names></name> <name><surname>Perkell</surname> <given-names>J. S.</given-names></name></person-group> (<year>2010</year>). <article-title>Adaptive auditory feedback control of the production of formant trajectories in the Mandarin triphthong /iau/ and its pattern of generalization</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>128</volume>, <fpage>2033</fpage>&#x02013;<lpage>2048</lpage>. <pub-id pub-id-type="doi">10.1121/1.3479539</pub-id><pub-id pub-id-type="pmid">20968374</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="book"><person-group person-group-type="author"><collab>Calliope</collab></person-group> (<year>1984</year>). <source>La Parole et Son Traitement Automatique</source>. <publisher-loc>Paris</publisher-loc>: <publisher-name>Masson</publisher-name>.</citation></ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Chomsky</surname> <given-names>N.</given-names></name> <name><surname>Halle</surname> <given-names>M.</given-names></name></person-group> (<year>1968</year>). <source>The Sound Pattern of English</source>. <publisher-loc>New-York, NY</publisher-loc>: <publisher-name>Harper &#x00026; Raw</publisher-name>.</citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Clements</surname> <given-names>G. N.</given-names></name></person-group> (<year>1985</year>). <article-title>The geometry of phonological features</article-title>. <source>Phonology</source> <volume>2</volume>, <fpage>225</fpage>&#x02013;<lpage>252</lpage>. <pub-id pub-id-type="doi">10.1017/S0952675700000440</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coker</surname> <given-names>C. H.</given-names></name></person-group> (<year>1976</year>). <article-title>A model of articulatory dynamics and control</article-title>. <source>Proc. IEEE</source> <volume>64</volume> <fpage>452</fpage>&#x02013;<lpage>460</lpage>. <pub-id pub-id-type="doi">10.1109/PROC.1976.10154</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Diehl</surname> <given-names>R. L.</given-names></name> <name><surname>Kluender</surname> <given-names>K. R.</given-names></name></person-group> (<year>1989</year>). <article-title>On the objects of speech perception</article-title>. <source>Ecol. Psychol.</source> <volume>1</volume>, <fpage>121</fpage>&#x02013;<lpage>144</lpage>. <pub-id pub-id-type="doi">10.1207/s15326969eco0102_2</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ernst</surname> <given-names>M. O.</given-names></name> <name><surname>Banks</surname> <given-names>M. S.</given-names></name></person-group> (<year>2002</year>). <article-title>Humans integrate visual and haptic information in a statistically optimal fashion</article-title>. <source>Nature</source> <volume>415</volume>, <fpage>429</fpage>&#x02013;<lpage>433</lpage>. <pub-id pub-id-type="doi">10.1038/415429a</pub-id><pub-id pub-id-type="pmid">11807554</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feldman</surname> <given-names>A. G.</given-names></name></person-group> (<year>1986</year>). <article-title>Once more on the equilibrium-point hypothesis (&#x003BB; model) for motor control</article-title>. <source>J. Mot. Behav.</source> <volume>18</volume>, <fpage>17</fpage>&#x02013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1080/00222895.1986.10735369</pub-id><pub-id pub-id-type="pmid">15136283</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feldman</surname> <given-names>H.</given-names></name> <name><surname>Friston</surname> <given-names>K. J.</given-names></name></person-group> (<year>2010</year>). <article-title>Attention, uncertainty, and free-energy</article-title>. <source>Front. Hum. Neurosci.</source> <volume>4</volume>:<fpage>215</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2010.00215</pub-id><pub-id pub-id-type="pmid">21160551</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname> <given-names>Y.</given-names></name> <name><surname>Gracco</surname> <given-names>V. L.</given-names></name> <name><surname>Max</surname> <given-names>L.</given-names></name></person-group> (<year>2011</year>). <article-title>Integration of auditory and somatosensory error signals in the neural control of speech movements</article-title>. <source>J. Neurophysiol.</source> <volume>106</volume>, <fpage>667</fpage>&#x02013;<lpage>679</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00638.2010</pub-id><pub-id pub-id-type="pmid">21562187</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Fowler</surname> <given-names>C. A.</given-names></name></person-group> (<year>1986</year>). <article-title>&#x0201C;An event approach to the study of speech perception from a direct-realist perspective,&#x0201D;</article-title> in <source>Status Report on Speech Research</source>, eds I. G. Mattingly and N. O&#x00027;Brien (<publisher-loc>New Haven, CT</publisher-loc>: <publisher-name>Haskins Laboratories</publisher-name>), <fpage>139</fpage>&#x02013;<lpage>169</lpage>.</citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friston</surname> <given-names>K.</given-names></name></person-group> (<year>2010</year>). <article-title>The free-energy principle: a unified brain theory?</article-title> <source>Nat. Rev. Neurosci.</source> <volume>11</volume>, <fpage>127</fpage>&#x02013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2787</pub-id><pub-id pub-id-type="pmid">20068583</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Friston</surname> <given-names>K.</given-names></name> <name><surname>Kiebel</surname> <given-names>S.</given-names></name></person-group> (<year>2009</year>). <article-title>Predictive coding under the free-energy principle</article-title>. <source>Philos. Trans. R. Soc. B</source> <volume>364</volume>, <fpage>1211</fpage>&#x02013;<lpage>1221</lpage>. <pub-id pub-id-type="doi">10.1098/rstb.2008.0300</pub-id><pub-id pub-id-type="pmid">19528002</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fujimura</surname> <given-names>O.</given-names></name></person-group> (<year>2000</year>). <article-title>The c/d model and prosodic control of articulatory behavior</article-title>. <source>Phonetica</source> <volume>57</volume>, <fpage>128</fpage>&#x02013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1159/000028467</pub-id><pub-id pub-id-type="pmid">10992134</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gay</surname> <given-names>T.</given-names></name> <name><surname>Lindblom</surname> <given-names>B.</given-names></name> <name><surname>Lubker</surname> <given-names>J.</given-names></name></person-group> (<year>1981</year>). <article-title>Production of bite-block vowels: acoustic equivalence by selective compensation</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>69</volume>, <fpage>802</fpage>&#x02013;<lpage>810</lpage>. <pub-id pub-id-type="doi">10.1121/1.385591</pub-id><pub-id pub-id-type="pmid">7240561</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gibson</surname> <given-names>J.</given-names></name></person-group> (<year>1979</year>). <source>The Ecological Approach to Visual Perception</source>. <publisher-loc>Boston, MA</publisher-loc>: <publisher-name>Houghton Mifflin</publisher-name>.</citation></ref>
<ref id="B22">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Goldstein</surname> <given-names>L.</given-names></name> <name><surname>Fowler</surname> <given-names>C. A.</given-names></name></person-group> (<year>2003</year>). <article-title>&#x0201C;Articulatory phonology: a phonology for public language use,&#x0201D;</article-title> in <source>Phonetics and Phonology in Language Comprehension and Production: Differences and Similarities</source>, eds N. O. Schiller and A. Meyer (<publisher-loc>Berlin</publisher-loc>: <publisher-name>Mouton de Gruyter</publisher-name>), <fpage>159</fpage>&#x02013;<lpage>207</lpage>.</citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guenther</surname> <given-names>F. H.</given-names></name></person-group> (<year>1995</year>). <article-title>Speech sound acquisition, coarticulation, and rate effects in a neural network model of speech production</article-title>. <source>Psychol. Rev.</source> <volume>102</volume>, <fpage>594</fpage>&#x02013;<lpage>621</lpage>. <pub-id pub-id-type="doi">10.1037/0033-295X.102.3.594</pub-id><pub-id pub-id-type="pmid">7624456</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guenther</surname> <given-names>F. H.</given-names></name> <name><surname>Ghosh</surname> <given-names>S. S.</given-names></name> <name><surname>Tourville</surname> <given-names>J. A.</given-names></name></person-group> (<year>2006</year>). <article-title>Neural modeling and imaging of the cortical interactions underlying syllable production</article-title>. <source>Brain Lang.</source> <volume>96</volume> <fpage>280</fpage>&#x02013;<lpage>301</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2005.06.001</pub-id><pub-id pub-id-type="pmid">16040108</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guenther</surname> <given-names>F. H.</given-names></name> <name><surname>Hampson</surname> <given-names>M.</given-names></name> <name><surname>Johnson</surname> <given-names>D.</given-names></name></person-group> (<year>1998</year>). <article-title>A theoretical investigation of reference frames for the planning of speech movements</article-title>. <source>Psychol. Rev.</source> <volume>105</volume>:<fpage>611</fpage>. <pub-id pub-id-type="doi">10.1037//0033-295X.105.4.611-633</pub-id><pub-id pub-id-type="pmid">9830375</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Haruno</surname> <given-names>M.</given-names></name> <name><surname>Wolpert</surname> <given-names>D. M.</given-names></name> <name><surname>Kawato</surname> <given-names>M.</given-names></name></person-group> (<year>1999</year>). <article-title>&#x0201C;Multiple paired forward-inverse models for human motor learning and control&#x0201D;</article-title> in <source>Advances in Neural Information Processing Systems 11</source>, eds M. Kearns, S. Solla, and D. Cohn (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>), <fpage>31</fpage>&#x02013;<lpage>37</lpage>.</citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hickok</surname> <given-names>G.</given-names></name></person-group> (<year>2012</year>). <article-title>Computational neuroanatomy of speech production</article-title>. <source>Nat. Rev. Neurosci.</source> <volume>13</volume>, <fpage>135</fpage>&#x02013;<lpage>145</lpage>. <pub-id pub-id-type="doi">10.1038/nrn3158</pub-id><pub-id pub-id-type="pmid">22218206</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houde</surname> <given-names>J. F.</given-names></name> <name><surname>Jordan</surname> <given-names>M. I.</given-names></name></person-group> (<year>1998</year>). <article-title>Sensorimotor adaptation in speech production</article-title>. <source>Science (New York, N.Y.)</source> <volume>1</volume>, <fpage>1213</fpage>&#x02013;<lpage>1216</lpage>. <pub-id pub-id-type="doi">10.1126/science.279.5354.1213</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houde</surname> <given-names>J. F.</given-names></name> <name><surname>Jordan</surname> <given-names>M. I.</given-names></name></person-group> (<year>2002</year>). <article-title>Sensorimotor adaptation of speech I: compensation and adaptation</article-title>. <source>J. Speech Lang. Hear. Res.</source> <volume>45</volume>, <fpage>295</fpage>&#x02013;<lpage>310</lpage>. <pub-id pub-id-type="doi">10.1044/1092-4388(2002/023)</pub-id><pub-id pub-id-type="pmid">12003512</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jordan</surname> <given-names>M. I.</given-names></name> <name><surname>Rumelhart</surname> <given-names>D. E.</given-names></name></person-group> (<year>1992</year>). <article-title>Forward models: supervised learning with a distal teacher</article-title>. <source>Cogn. Sci.</source> <volume>16</volume>, <fpage>307</fpage>&#x02013;<lpage>354</lpage>. <pub-id pub-id-type="doi">10.1207/s15516709cog1603_1</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Katseff</surname> <given-names>S.</given-names></name> <name><surname>Houde</surname> <given-names>J.</given-names></name> <name><surname>Johnson</surname> <given-names>K.</given-names></name></person-group> (<year>2012</year>). <article-title>Partial compensation for altered auditory feedback: a tradeoff with somatosensory feedback?</article-title> <source>Lang. Speech</source> <volume>55</volume>, <fpage>295</fpage>&#x02013;<lpage>308</lpage>. <pub-id pub-id-type="doi">10.1177/0023830911417802</pub-id><pub-id pub-id-type="pmid">22783636</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kawato</surname> <given-names>M.</given-names></name> <name><surname>Maeda</surname> <given-names>Y.</given-names></name> <name><surname>Uno</surname> <given-names>Y.</given-names></name> <name><surname>Suzuki</surname> <given-names>R.</given-names></name></person-group> (<year>1990</year>). <article-title>Trajectory formation of arm movement by cascade neural network model based on minimum torque-change criterion</article-title>. <source>Biol. Cybern.</source> <volume>62</volume>, <fpage>275</fpage>&#x02013;<lpage>288</lpage>. <pub-id pub-id-type="doi">10.1007/BF00201442</pub-id><pub-id pub-id-type="pmid">2310782</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kelso</surname> <given-names>J. A.</given-names></name> <name><surname>Saltzman</surname> <given-names>E. L.</given-names></name> <name><surname>Tuller</surname> <given-names>B.</given-names></name></person-group> (<year>1986</year>). <article-title>The dynamical perspective on speech production: data and theory</article-title>. <source>J. Phonet.</source> <volume>14</volume>, <fpage>29</fpage>&#x02013;<lpage>59</lpage>.</citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kersten</surname> <given-names>D.</given-names></name> <name><surname>Mamassian</surname> <given-names>P.</given-names></name> <name><surname>Yuille</surname> <given-names>A.</given-names></name></person-group> (<year>2004</year>). <article-title>Object perception as Bayesian inference</article-title>. <source>Annu. Rev. Psychol.</source> <volume>55</volume>, <fpage>271</fpage>&#x02013;<lpage>304</lpage>. <pub-id pub-id-type="doi">10.1146/annurev.psych.55.090902.142005</pub-id><pub-id pub-id-type="pmid">14744217</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Keyser</surname> <given-names>S. J.</given-names></name> <name><surname>Stevens</surname> <given-names>K. N.</given-names></name></person-group> (<year>1994</year>). <article-title>Feature geometry and the vocal tract</article-title>. <source>Phonology</source> <volume>11</volume>, <fpage>207</fpage>&#x02013;<lpage>236</lpage>. <pub-id pub-id-type="doi">10.1017/S0952675700001950</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>K&#x000F6;rding</surname> <given-names>K. P.</given-names></name> <name><surname>Wolpert</surname> <given-names>D. M.</given-names></name></person-group> (<year>2004</year>). <article-title>Bayesian integration in sensorimotor learning</article-title>. <source>Nature</source> <volume>427</volume>, <fpage>244</fpage>&#x02013;<lpage>247</lpage>. <pub-id pub-id-type="doi">10.1038/nature02169</pub-id><pub-id pub-id-type="pmid">14724638</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kr&#x000F6;ger</surname> <given-names>B.</given-names></name> <name><surname>Schr&#x000F6;der</surname> <given-names>G.</given-names></name> <name><surname>Opgen?Rhein</surname> <given-names>C.</given-names></name></person-group> (<year>1995</year>). <article-title>A gesture?based dynamic model describing articulatory movement data</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>98</volume>, <fpage>1878</fpage>&#x02013;<lpage>1889</lpage>. <pub-id pub-id-type="doi">10.1121/1.413374</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kr&#x000F6;ger</surname> <given-names>B. J.</given-names></name> <name><surname>Kannampuzha</surname> <given-names>J.</given-names></name> <name><surname>Neuschaefer-Rube</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>Towards a neurocomputational model of speech production and perception</article-title>. <source>Speech Commun.</source> <volume>51</volume>, <fpage>793</fpage>&#x02013;<lpage>809</lpage>. <pub-id pub-id-type="doi">10.1016/j.specom.2008.08.002</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lametti</surname> <given-names>D. R.</given-names></name> <name><surname>Nasir</surname> <given-names>S. M.</given-names></name> <name><surname>Ostry</surname> <given-names>D. J.</given-names></name></person-group> (<year>2012</year>). <article-title>Sensory preference in speech production revealed by simultaneous alteration of auditory and somatosensory feedback</article-title>. <source>J. Neurosci.</source> <volume>32</volume>, <fpage>9351</fpage>&#x02013;<lpage>9358</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0404-12.2012</pub-id><pub-id pub-id-type="pmid">22764242</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lametti</surname> <given-names>D. R.</given-names></name> <name><surname>Rochet-Capellan</surname> <given-names>A.</given-names></name> <name><surname>Neufeld</surname> <given-names>E.</given-names></name> <name><surname>Shiller</surname> <given-names>D. M.</given-names></name> <name><surname>Ostry</surname> <given-names>D. J.</given-names></name></person-group> (<year>2014</year>). <article-title>Plasticity in the human speech motor system drives changes in speech perception</article-title>. <source>J. Neurosci.</source> <volume>34</volume>, <fpage>10339</fpage>&#x02013;<lpage>10346</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0108-14.2014</pub-id></citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Laurent</surname> <given-names>R.</given-names></name> <name><surname>Barnaud</surname> <given-names>M.-L.</given-names></name> <name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Bessi&#x000E8;re</surname> <given-names>P.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name></person-group> (<year>2017</year>). <article-title>The complementary roles of auditory and motor information evaluated in a Bayesian perceptuo-motor model of speech perception</article-title>. <source>Psychol. Rev</source>. <volume>124</volume>, <fpage>572</fpage>&#x02013;<lpage>602</lpage>. <pub-id pub-id-type="doi">10.1037/rev0000069</pub-id><pub-id pub-id-type="pmid">28471206</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liberman</surname> <given-names>A. M.</given-names></name> <name><surname>Cooper</surname> <given-names>F. S.</given-names></name> <name><surname>Shankweiler</surname> <given-names>D. P.</given-names></name> <name><surname>Studdert-Kennedy</surname> <given-names>M.</given-names></name></person-group> (<year>1967</year>). <article-title>Perception of the speech code</article-title>. <source>Psychol. Rev.</source> <volume>74</volume>, <fpage>431</fpage>&#x02013;<lpage>461</lpage>. <pub-id pub-id-type="doi">10.1037/h0020279</pub-id><pub-id pub-id-type="pmid">4170865</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lindblom</surname> <given-names>B.</given-names></name></person-group> (<year>1990</year>). <article-title>&#x0201C;Explaining phonetic variation: a sketch of the h&#x00026;h theory,&#x0201D;</article-title> in <source>Speech Production and Speech Modelling</source>, eds W. Hardcastle and A. Marchal (<publisher-loc>Dordrecht</publisher-loc>: <publisher-name>Kluwer Academic Publishers</publisher-name>), <fpage>403</fpage>&#x02013;<lpage>439</lpage>.</citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lindblom</surname> <given-names>B.</given-names></name></person-group> (<year>1996</year>). <article-title>Role of articulation in speech perception: clues from production</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>99</volume>, <fpage>1683</fpage>&#x02013;<lpage>1692</lpage>. <pub-id pub-id-type="doi">10.1121/1.414691</pub-id><pub-id pub-id-type="pmid">8819859</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>L.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Dang</surname> <given-names>J.</given-names></name></person-group> (<year>2015</year>). <article-title>Strength of syllabic influences on articulation in mandarin chinese and french: insights from a motor control approach</article-title>. <source>J. Phonet.</source> <volume>53</volume>, <fpage>101</fpage>&#x02013;<lpage>124</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2015.09.005</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="thesis"><person-group person-group-type="author"><name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name></person-group> (<year>2002</year>). <source>Production et perception des voyelles au cours de la croissance du conduit vocal : variabilit&#x000E9;, invariance et normalisation</source>. (Unpublished Ph.D. thesis). <publisher-loc>Grenoble</publisher-loc>: <publisher-name>Universit&#x000E9; Stendhal de Grenoble</publisher-name>.</citation></ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Moulin-Frier</surname> <given-names>C.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name> <name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Bessi&#x000E8;re</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>COSMO (&#x0201C;Communicating about Objects using Sensory-Motor Operations&#x0201D;): a Bayesian modeling framework for studying speech communication and the emergence of phonological systems</article-title>. <source>J. Phonet.</source> <volume>53</volume>, <fpage>5</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2015.06.001</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Parrell</surname> <given-names>B.</given-names></name> <name><surname>Ramanarayanan</surname> <given-names>V.</given-names></name> <name><surname>Nagarajan</surname> <given-names>S.</given-names></name> <name><surname>Houde</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;FACTS: a hierarchical task-based control model of speech incorporating sensory feedback,&#x0201D;</article-title> in <source>Proceedings of Interspeech 2018</source> (<publisher-loc>Hyderabad</publisher-loc>), <fpage>1497</fpage>&#x02013;<lpage>1501</lpage>.</citation></ref>
<ref id="B49">
<citation citation-type="thesis"><person-group person-group-type="author"><name><surname>Patri</surname> <given-names>J.-F.</given-names></name></person-group> (<year>2018</year>). <source>Bayesian modeling of speech motor planning: variability, multisensory goals and perceptuo-motor interactions</source>. (Unpublished Ph.D. thesis). <publisher-loc>Grenoble</publisher-loc>: <publisher-name>Universit&#x000E9; Grenoble Alpes</publisher-name>.</citation></ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patri</surname> <given-names>J.-F.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>Optimal speech motor control and token-to-token variability: a Bayesian modeling approach</article-title>. <source>Biol. Cybernet.</source> <volume>109</volume>, <fpage>611</fpage>&#x02013;<lpage>626</lpage>. <pub-id pub-id-type="doi">10.1007/s00422-015-0664-4</pub-id><pub-id pub-id-type="pmid">26497359</pub-id></citation></ref>
<ref id="B51">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Patri</surname> <given-names>J.-F.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Bayesian modeling in speech motor control: a principled structure for the integration of various constraints,&#x0201D;</article-title> in <source>Interspeech 2016</source> (<publisher-loc>San Francisco, CA</publisher-loc>), <fpage>3588</fpage>&#x02013;<lpage>3592</lpage>.</citation></ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Patri</surname> <given-names>J.-F.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Diard</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>What drives the perceptual change resulting from speech motor adaptation? Evaluation of hypotheses in a bayesian modeling framework</article-title>. <source>PLoS Comput. Biol.</source> <volume>14</volume>:<fpage>e1005942</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1005942</pub-id><pub-id pub-id-type="pmid">29357357</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Perkell</surname> <given-names>J. S.</given-names></name> <name><surname>Lane</surname> <given-names>H.</given-names></name> <name><surname>Ghosh</surname> <given-names>S.</given-names></name> <name><surname>Matthies</surname> <given-names>M. L.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Mechanisms of vowel production: auditory goals and speaker acuity,&#x0201D;</article-title> in <source>Proceedings of the 8th International Seminar on Speech Production</source> (<publisher-loc>Strasbourg</publisher-loc>), <fpage>29</fpage>&#x02013;<lpage>32</lpage>.</citation></ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perrier</surname> <given-names>P.</given-names></name></person-group> (<year>2005</year>). <article-title>Control and representations in speech production</article-title>. <source>ZAS Papers Lingust.</source> <volume>40</volume>, <fpage>109</fpage>&#x02013;<lpage>132</lpage>.</citation></ref>
<ref id="B55">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Ma</surname> <given-names>L.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Speech planning for V1CV2 sequences: influence of the planned sequence,&#x0201D;</article-title> in <source>Proceedings of the 8th International Seminar on Speech Production (ISSP 2008)</source>, ed ISSP-2008 (<publisher-loc>Strasbourg</publisher-loc>: <publisher-name>Universit&#x000E9; de Strasbourg</publisher-name>), <fpage>69</fpage>&#x02013;<lpage>72</lpage>.</citation></ref>
<ref id="B56">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Ma</surname> <given-names>L.</given-names></name> <name><surname>Payan</surname> <given-names>Y.</given-names></name></person-group> (<year>2005</year>). <article-title>&#x0201C;Modeling the production of VCV sequences via the inversion of a biomechanical model of the tongue,&#x0201D;</article-title> in <source>Proceedings of Interspeech 2005</source> (<publisher-loc>Lisbon</publisher-loc>), 1041&#x02014;1044.</citation></ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Payan</surname> <given-names>Y.</given-names></name> <name><surname>Buchaillard</surname> <given-names>S.</given-names></name> <name><surname>Nazari</surname> <given-names>M. A.</given-names></name> <name><surname>Chabanas</surname> <given-names>M.</given-names></name></person-group> (<year>2011</year>). <article-title>Biomechanical models to study speech</article-title>. <source>Faits de Langues</source> <volume>37</volume>, <fpage>155</fpage>&#x02013;<lpage>171</lpage>.</citation></ref>
<ref id="B58">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Poggio</surname> <given-names>T.</given-names></name> <name><surname>Girosi</surname> <given-names>F.</given-names></name></person-group> (<year>1989</year>). <source>A Theory of Networks for Approximation and Learning</source>. Technical report, Artificial Intelligence Laboratory &#x00026; Center for Biological Information Processing (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT</publisher-name>).</citation></ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Purcell</surname> <given-names>D. W.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>2006</year>). <article-title>Compensation following real-time manipulation of formants in isolated vowels</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>119</volume>, <fpage>2288</fpage>&#x02013;<lpage>2297</lpage>. <pub-id pub-id-type="doi">10.1121/1.2173514</pub-id><pub-id pub-id-type="pmid">16642842</pub-id></citation></ref>
<ref id="B60">
<citation citation-type="thesis"><person-group person-group-type="author"><name><surname>Robert-Ribes</surname> <given-names>J.</given-names></name></person-group> (<year>1995</year>). <source>Mod&#x000E8;les d&#x00027;int&#x000E9;gration audiovisuelle de signaux linguistiques : de la perception humaine a la reconnaissance automatique des voyelles</source>. (Unpublished Ph.D. thesis). <publisher-loc>Grenoble</publisher-loc>: <publisher-name>Institut National Polytechnique de Grenoble</publisher-name>.</citation></ref>
<ref id="B61">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Saltzman</surname> <given-names>E. L.</given-names></name></person-group> (<year>1986</year>). <article-title>&#x0201C;Task dynamic coordination of the speech articulators: a preliminary model,&#x0201D;</article-title> in <source>Generation and Modulation of Action Patterns</source>, number 15, Experimental Brain Research Series Edition, eds H. Heuer and C. Fromm (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Springer-Verlag</publisher-name>), <fpage>129</fpage>&#x02013;<lpage>144</lpage>.</citation></ref>
<ref id="B62">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Saltzman</surname> <given-names>E. L.</given-names></name> <name><surname>Munhall</surname> <given-names>K. G.</given-names></name></person-group> (<year>1989</year>). <article-title>A dynamical approach to gestural patterning in speech production</article-title>. <source>Ecol. Psychol.</source> <volume>1</volume>, <fpage>333</fpage>&#x02013;<lpage>382</lpage>. <pub-id pub-id-type="doi">10.1207/s15326969eco0104_2</pub-id></citation></ref>
<ref id="B63">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Savariaux</surname> <given-names>C.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name> <name><surname>Orliaguet</surname> <given-names>J.-P.</given-names></name></person-group> (<year>1995</year>). <article-title>Compensation strategies for the perturbation of the rounded vowel [u] using a lip tube: a study of the control space in speech production</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>98</volume>, <fpage>2428</fpage>&#x02013;<lpage>2442</lpage>. <pub-id pub-id-type="doi">10.1121/1.413277</pub-id></citation></ref>
<ref id="B64">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schwartz</surname> <given-names>J.-L.</given-names></name> <name><surname>Basirat</surname> <given-names>A.</given-names></name> <name><surname>M&#x000E9;nard</surname> <given-names>L.</given-names></name> <name><surname>Sato</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>The perception-for-action-control theory (PACT): a perceptuo-motor theory of speech perception</article-title>. <source>J. Neurolinguist.</source> <volume>25</volume>, <fpage>336</fpage>&#x02013;<lpage>354</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneuroling.2009.12.004</pub-id></citation></ref>
<ref id="B65">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shadmehr</surname> <given-names>R.</given-names></name> <name><surname>Mussa-Ivaldi</surname> <given-names>F. A.</given-names></name></person-group> (<year>1994</year>). <article-title>Adaptive representation of dynamics during learning of a motor task</article-title>. <source>J. Neurosci.</source> <volume>14</volume>, <fpage>3208</fpage>&#x02013;<lpage>3224</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.14-05-03208.1994</pub-id><pub-id pub-id-type="pmid">8182467</pub-id></citation></ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shiller</surname> <given-names>D. M.</given-names></name> <name><surname>Sato</surname> <given-names>M.</given-names></name> <name><surname>Gracco</surname> <given-names>V. L.</given-names></name> <name><surname>Baum</surname> <given-names>S. R.</given-names></name></person-group> (<year>2009</year>). <article-title>Perceptual recalibration of speech sounds following speech motor learning</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>125</volume>, <fpage>1103</fpage>&#x02013;<lpage>1113</lpage>. <pub-id pub-id-type="doi">10.1121/1.3058638</pub-id><pub-id pub-id-type="pmid">19206885</pub-id></citation></ref>
<ref id="B67">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Stevens</surname> <given-names>K. N.</given-names></name></person-group> (<year>1972</year>). <article-title>&#x0201C;The quantal nature of speech: evidence from articulatory-acoustic data,&#x0201D;</article-title> in <source>Human Communication: A Unified View</source>, eds E. David and P. Denes (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>McGraw-Hill</publisher-name>), <fpage>51</fpage>&#x02013;<lpage>66</lpage>.</citation></ref>
<ref id="B68">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stevens</surname> <given-names>K. N.</given-names></name></person-group> (<year>1996</year>). <article-title>Critique: articulatory-acoustic relations and their role in speech perception</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>99</volume>, <fpage>1693</fpage>&#x02013;<lpage>1694</lpage>. <pub-id pub-id-type="doi">10.1121/1.414692</pub-id><pub-id pub-id-type="pmid">8964929</pub-id></citation></ref>
<ref id="B69">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stevens</surname> <given-names>K. N.</given-names></name> <name><surname>Blumstein</surname> <given-names>S. E.</given-names></name></person-group> (<year>1978</year>). <article-title>Invariant cues for place of articulation in stop consonants</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>64</volume>, <fpage>1358</fpage>&#x02013;<lpage>1368</lpage>. <pub-id pub-id-type="doi">10.1121/1.382102</pub-id><pub-id pub-id-type="pmid">744836</pub-id></citation></ref>
<ref id="B70">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sussman</surname> <given-names>H. M.</given-names></name> <name><surname>McCaffrey</surname> <given-names>H. A.</given-names></name> <name><surname>Matthews</surname> <given-names>S. A.</given-names></name></person-group> (<year>1991</year>). <article-title>An investigation of locus equations as a source of relational invariance for stop place categorization</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>90</volume>, <fpage>1309</fpage>&#x02013;<lpage>1325</lpage>. <pub-id pub-id-type="doi">10.1121/1.401923</pub-id></citation></ref>
<ref id="B71">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tian</surname> <given-names>X.</given-names></name> <name><surname>Poeppel</surname> <given-names>D.</given-names></name></person-group> (<year>2010</year>). <article-title>Mental imagery of speech and movement implicates the dynamics of internal forward models</article-title>. <source>Front. Psychol.</source> <volume>1</volume>:<fpage>166</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2010.00166</pub-id><pub-id pub-id-type="pmid">21897822</pub-id></citation></ref>
<ref id="B72">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tremblay</surname> <given-names>S.</given-names></name> <name><surname>Shiller</surname> <given-names>D. M.</given-names></name> <name><surname>Ostry</surname> <given-names>D. J.</given-names></name></person-group> (<year>2003</year>). <article-title>Somatosensory basis of speech production</article-title>. <source>Nature</source> <volume>423</volume>, <fpage>866</fpage>&#x02013;<lpage>869</lpage>. <pub-id pub-id-type="doi">10.1038/nature01710</pub-id><pub-id pub-id-type="pmid">12815431</pub-id></citation></ref>
<ref id="B73">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Villacorta</surname> <given-names>V. M.</given-names></name> <name><surname>Perkell</surname> <given-names>J. S.</given-names></name> <name><surname>Guenther</surname> <given-names>F. H.</given-names></name></person-group> (<year>2007</year>). <article-title>Sensorimotor adaptation to perturbations of vowel acoustics and its relation to perception</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>122</volume>, <fpage>2306</fpage>&#x02013;<lpage>2319</lpage>. <pub-id pub-id-type="doi">10.1121/1.2773966</pub-id><pub-id pub-id-type="pmid">17902866</pub-id></citation></ref>
<ref id="B74">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Winkler</surname> <given-names>R.</given-names></name> <name><surname>Ma</surname> <given-names>L.</given-names></name> <name><surname>Perrier</surname> <given-names>P.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;A model of optimal speech production planning integrating dynamical constraints to achieve appropriate articulatory timing&#x0201D;</article-title> in <source>Proceedings of the 9th International Seminar on Speech Production</source> (<publisher-loc>Montr&#x000E9;al, QC</publisher-loc>) <fpage>235</fpage>&#x02013;<lpage>236</lpage>.</citation></ref>
<ref id="B75">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Yan</surname> <given-names>H.</given-names></name> <name><surname>Dang</surname> <given-names>J.</given-names></name> <name><surname>Cao</surname> <given-names>M.</given-names></name> <name><surname>Kr&#x000F6;ger</surname> <given-names>B. J.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;A new framework of neurocomputational model for speech production,&#x0201D;</article-title> in <source>Proceedings of the 9th International Symposium on Chinese Spoken Language Processing</source> (<publisher-loc>Singapore</publisher-loc>), <fpage>294</fpage>&#x02013;<lpage>298</lpage>.</citation></ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> The research leading to these results has received funding from the European Research Council under the European Community&#x00027;s Seventh Framework Programme (FP7/2007-2 013 Grant Agreement no. 339152, Speech Unit(e)s, PI: Jean-Luc-Schwartz), and from the European Union&#x00027;s Horizon 2020 research and innovation programme under the Marie Sk&#x00142;odowska-Curie grant agreement No 754490 (MINDED Program). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.</p></fn>
</fn-group>
</back>
</article>