Event Abstract

Sensorimotor Modeling of Speech Production, Speech Perception, and Speech Acquisition

  • 1 Neurophonetics Group at DPPCD, RWTH Aachen Univ., Germany; School of Computer Science, Tianjin Univ., China , Germany
  • 2 Neurophonetics Group at DPPCD, RWTH Aachen University, Germany, Germany
  • 3 Department of Phoniatrics, Pedaudiology, and Communication Disorders (DPPCD), RWTH Aachen University, Germany, Germany

"Our model of speech production, speech perception, and speech acquisition has been implemented and tested by simulating early phases of speech acquisition (i.e. babbling phase and imitation phase) and by performing production and perception tests after learning (Kröger et al. 2009). The detailed structure of the model is given in Fig. 1. A characteristic feature of our approach is that we assume a self-organizing phonetic map which is associated with working memory state maps (distributed neural representations), representing the motor plan, the somatosensory activation pattern (tactile and proprioceptive), and the auditory activation pattern of syllables. Speech acquisition is simulated in our approach by applying a huge amount of training items to the model. These training items represent stimuli, which are exposed to a newborn and later on to a toddler during the first two years of lifetime. Acquisition starts with "babbling", i.e. a training phase which is mainly language independent. Here the model generates random motor patterns (motor plan states) and produces appropriate auditory and somatosensory patterns (auditory and somatosensory states). Motor plan and sensory states are exposed to the model nearly simultaneously and thus allow associative learning, i.e. an association of specific motor plan states with corresponding sensory states (Kröger et al. 2009). This learning leads to an adjustment of synaptic weights between neurons of state maps and neurons of the self-organizing phonetic map. Neurons within the phonetic map represent specific sensorimotor states and these states are ordered with respect to phonetic features within this map. This initial sensorimotor babbling training later on allows "imitation training", because now the model is able to generate motor patterns, if external auditory stimuli are given by an external speaker ("mother"). Imitation training leads to a further ordering of states within the phonetic map and to language-specific speaking skills. After babbling and imitation training (imitaton of Standard German), the current version of our model has associated motor plan and sensory representations of the 200 most frequent syllables of Standard German and is capable of reproducing and perceiving (identifying) these syllables. References: Kröger BJ, Kannampuzha J, Neuschaefer-Rube C (2009) Towards a neurocomputational model of speech production and perception. Speech Communication 51, 793-809"

Figure 1

Keywords: computational neuroscience, Sensorimotor Modeling, Speech Perception, speech acquisition, simulation

Conference: 5th INCF Congress of Neuroinformatics, Munich, Germany, 10 Sep - 12 Sep, 2012.

Presentation Type: Poster

Topic: Neuroinformatics

Citation: Kroeger B, Eckers C and Neuschaefer-Rube C (2014). Sensorimotor Modeling of Speech Production, Speech Perception, and Speech Acquisition. Front. Neuroinform. Conference Abstract: 5th INCF Congress of Neuroinformatics. doi: 10.3389/conf.fninf.2014.08.00021

Copyright: The abstracts in this collection have not been subject to any Frontiers peer review or checks, and are not endorsed by Frontiers. They are made available through the Frontiers publishing platform as a service to conference organizers and presenters.

The copyright in the individual abstracts is owned by the author of each abstract or his/her employer unless otherwise stated.

Each abstract, as well as the collection of abstracts, are published under a Creative Commons CC-BY 4.0 (attribution) licence (https://creativecommons.org/licenses/by/4.0/) and may thus be reproduced, translated, adapted and be the subject of derivative works provided the authors and Frontiers are attributed.

For Frontiers’ terms and conditions please see https://www.frontiersin.org/legal/terms-and-conditions.

Received: 21 Mar 2013; Published Online: 27 Feb 2014.

* Correspondence: Dr. Bernd J. Kroeger, Neurophonetics Group at DPPCD, RWTH Aachen Univ., Germany; School of Computer Science, Tianjin Univ., China, Aachen, Germany, bernd.kroeger@rwth-aachen.de