<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="editorial">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Robot. AI</journal-id>
<journal-title>Frontiers in Robotics and AI</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Robot. AI</abbrev-journal-title>
<issn pub-type="epub">2296-9144</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frobt.2021.674832</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Robotics and AI</subject>
<subj-group>
<subject>Editorial</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Editorial: Language and Robotics</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Taniguchi</surname> <given-names>Tadahiro</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/379696/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Horii</surname> <given-names>Takato</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/380175/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Hinaut</surname> <given-names>Xavier</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/118163/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Spranger</surname> <given-names>Michael</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/170798/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Mochihashi</surname> <given-names>Daichi</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/508357/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Nagai</surname> <given-names>Takayuki</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/381312/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>College of Information Science and Engineering, Ritsumeikan University</institution>, <addr-line>Kyoto</addr-line>, <country>Japan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Graduate School of Engineering Science, Osaka University</institution>, <addr-line>Suita</addr-line>, <country>Japan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Institut National de Recherche en Informatique et en Automatique (INRIA), Rocquencourt</institution>, <addr-line>&#x000CE;le-de-France</addr-line>, <country>France</country></aff>
<aff id="aff4"><sup>4</sup><institution>Sony Computer Science Laboratories</institution>, <addr-line>Tokyo</addr-line>, <country>Japan</country></aff>
<aff id="aff5"><sup>5</sup><institution>Department of Statistical Inference and Mathematics, The Institute of Statistical Mathematics</institution>, <addr-line>Tokyo</addr-line>, <country>Japan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited and reviewed by: Mikhail Prokopenko, The University of Sydney, Australia</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Tadahiro Taniguchi <email>taniguchi&#x00040;ci.ritsumei.ac.jp</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Computational Intelligence in Robotics, a section of the journal Frontiers in Robotics and AI</p></fn></author-notes>
<pub-date pub-type="epub">
<day>12</day>
<month>04</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>8</volume>
<elocation-id>674832</elocation-id>
<history>
<date date-type="received">
<day>02</day>
<month>03</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>08</day>
<month>03</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2021 Taniguchi, Horii, Hinaut, Spranger, Mochihashi and Nagai.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Taniguchi, Horii, Hinaut, Spranger, Mochihashi and Nagai</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license></permissions>
<related-article id="RA1" related-article-type="commentary-article" xlink:href="https://www.frontiersin.org/research-topics/8861/language-and-robotics" ext-link-type="uri">Editorial on the Research Topic <article-title>Language and Robotics</article-title></related-article> <kwd-group>
<kwd>language acquisition by robots</kwd>
<kwd>multimodal communication</kwd>
<kwd>concept formation</kwd>
<kwd>symbol grounding in robotics</kwd>
<kwd>symbol emergence in robotics</kwd>
<kwd>deep learning for robotics</kwd>
<kwd>emergence of communication</kwd>
</kwd-group>
<counts>
<fig-count count="0"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="12"/>
<page-count count="3"/>
<word-count count="1625"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Language in the real-world environment involves a wide range of challenges in robotics and artificial intelligence (AI). Service robots are required to communicate and collaborate with people using language in the real-world environment. When a robot receives a spoken command from a user in a domestic environment, the robot must understand its meaning in the context of the specific environment. For example, to understand the meaning of &#x0201C;please bring me a pen in Takato&#x00027;s room&#x0201D; the robot needs to know where to find a pen and where Takato&#x00027;s room is. Futhermore, words or expressions (i.e., sounds processed as symbols) can be invented naturally in our daily environment and their meaning can change (Spranger, <xref ref-type="bibr" rid="B7">2016</xref>) over time (i.e., depending on the culture or age of the speaker). Robots thus need to adapt like humans to these versatile aspects of language and demonstrate the ability to learn any language (Hinaut and Twiefel, <xref ref-type="bibr" rid="B3">2019</xref>). In robotics, language understanding inevitably involves multimodal learning, semantic mapping, and behavior learning. To enable a robot to interact orally with people in a long-term manner, we need to develop an AI that makes a robot learn and adapt to language in the real-world environment and in an on-line manner. This topic thus raises several challenges to bridge the gap from low-level sensorimotor interaction (Pagliarini et al., <xref ref-type="bibr" rid="B6">in press</xref>) to high-level compositional symbolic communication. Taking inspiration of how children acquire language can help to design the simplest mechanisms to deal with these challenges. Conversely, robotics can help modeling and test hypotheses about language acquisition and language grounding (Cangelosi and Schlesinger, <xref ref-type="bibr" rid="B1">2015</xref>; Taniguchi et al., <xref ref-type="bibr" rid="B10">2016</xref>, <xref ref-type="bibr" rid="B12">2018</xref>; Hinaut and Spranger, <xref ref-type="bibr" rid="B2">2019</xref>), in particular through cross-situational experiments (Taniguchi et al., <xref ref-type="bibr" rid="B9">2017</xref>; Juven and Hinaut, <xref ref-type="bibr" rid="B4">2020</xref>).</p>
<p>Following the successfully organized session &#x0201C;Language and Robotics&#x0201D; held in IEEE-IROS 2018<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref>, we organized this Research Topic. We aimed to publish original papers from robotics, natural language processing, machine learning, and cognitive science to share knowledge about the state-of-the-art machine learning methods and perspectives that contribute to modeling language-related capabilities in robotics.</p></sec>
<sec id="s2">
<title>2. About the Research Topic</title>
<p>We are pleased to present five research articles related to semantic mapping, language understanding, motion segmentation, symbol emergence, and language evolution. In this section, we briefly introduce each paper.</p>
<p>First, three papers focused on language-related cognitive capabilities integrating real-world sensor information full of uncertainty and high-dimensional. Each method involves deep learning methods dealing with high-dimensional uncertain real data in robotics. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frobt.2019.00115">Nagano et al.</ext-link> proposed a new machine learning method called a hierarchical Dirichlet process-variational autoencoder-Gaussian process-hidden semi-Markov model (HVGH). The method extended a hierarchical Dirichlet process-Gaussian process-hidden semi-Markov model (HDP-GP-HSMM) that can automatically segment time series data. HVGH integrated variational autoencoder and the HDP-GP-HSMM and achieved automatic motion segmentation along with representation learning. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frobt.2019.00031">Katsumata et al.</ext-link> proposed a statistical semantic mapping method called SpCoMapping, which means spatial concept formation and semantic mapping. The proposed model employed Markov random field into a pre-existing spatial concept formation method and became able to learn the arbitrary shape of a place on a map. The method integrated multimodal information, e.g., language, vision, and position, to find semantic information of places. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frobt.2019.00144">Tada et al.</ext-link> proposed a robust language understanding method by introducing noise injection into the sequence-to-sequence network. Recently, semantic parsing that enables a robot to understand the meaning of human user commands is developed based on deep learning methods. However, semantic parsing in natural language processing does not assume the existence of speech recognition errors. This paper showed the conventional idea of noise injection to sequence-to-sequence network semantic parsing can improve the robustness of a robot&#x00027;s language understanding.</p>
<p>Second, two papers focused on the emergence, or evolution, of symbols. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frobt.2020.00012">Cambier et al.</ext-link> described the perspectives of language evolution in swarm robotics. They advocated an approach based on language games for the further development of emergent communication in swarm robots. They suggested that swarm robotics can be an ideal testbed to advance research on the emergence of language-like communication. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frobt.2019.00134">Hagiwara et al.</ext-link> proposed a new computational model representing symbol emergence. The model proposed in this paper regarded symbol emergence as a multiagent multimodal categorization problem. The convergence of the algorithm was guaranteed based on the theory of Markov chain Monte Carlo. This symbol emergence model involved sharing signs among agents and making each agent form internal representations based on its sensorimotor information.</p></sec>
<sec id="s3">
<title>3. Next Step</title>
<p>With the great success of this Research Topic, we organized related workshops and a tutorial<xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>. A survey paper related to this topic has already been published (Tangiuchi et al., <xref ref-type="bibr" rid="B8">2019</xref>). We believe that integrating low-level and high-level cognitive capabilities (Nakamura et al., <xref ref-type="bibr" rid="B5">2018</xref>; Taniguchi et al., <xref ref-type="bibr" rid="B11">2020</xref>) in conjunction with language learning in the real-world environment is crucial to creating an artificial cognitive system, i.e., a robot, which can conduct lifelong learning in the real-world environment and achieves long-term human-robot interaction to support daily human activities. The intersection of language and robotics is a crucial Research Topic for further advancement in robotics and AI. We hope that this special issue will accelerate the cutting-edge studies in robotics and AI that aim to create human-level embodied AI that can communicate and collaborate with people in the real-world environment.</p></sec>
<sec id="s4">
<title>Author Contributions</title>
<p>All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.</p></sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
</body>
<back>
<ack><p>The authors gratefully acknowledge the contributions of participants in this special issue.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Cangelosi</surname> <given-names>A.</given-names></name> <name><surname>Schlesinger</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <source>Developmental Robotics: From Babies to Robots</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.<pub-id pub-id-type="pmid">31866848</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hinaut</surname> <given-names>X.</given-names></name> <name><surname>Spranger</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>Learning to parse grounded language using reservoir computing</article-title>, in <source>2019 Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)</source> (<publisher-loc>Oslo</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>266</fpage>&#x02013;<lpage>271</lpage>.</citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hinaut</surname> <given-names>X.</given-names></name> <name><surname>Twiefel</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>Teach your robot your language! trainable neural parser for modeling human sentence processing: Examples for 15 languages</article-title>. <source>IEEE Trans. Cogn. Dev. Syst.</source> <volume>12</volume>, <fpage>179</fpage>&#x02013;<lpage>188</lpage>. <pub-id pub-id-type="doi">10.1109/TCDS.2019.2957006</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Juven</surname> <given-names>A.</given-names></name> <name><surname>Hinaut</surname> <given-names>X.</given-names></name></person-group> (<year>2020</year>). <article-title>Cross-situational learning with reservoir computing for language acquisition modelling</article-title>, in <source>2020 International Joint Conference on Neural Networks (IJCNN)</source> (<publisher-loc>Glasgow</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>8</lpage>.</citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nakamura</surname> <given-names>T.</given-names></name> <name><surname>Nagai</surname> <given-names>T.</given-names></name> <name><surname>Taniguchi</surname> <given-names>T.</given-names></name></person-group> (<year>2018</year>). <article-title>Serket: an architecture for connecting stochastic models to realize a large-scale cognitive model</article-title>. <source>Front. Neurorobot.</source> <volume>12</volume>:<fpage>25</fpage>. <pub-id pub-id-type="doi">10.3389/fnbot.2018.00025</pub-id><pub-id pub-id-type="pmid">29997493</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pagliarini</surname> <given-names>S.</given-names></name> <name><surname>Leblois</surname> <given-names>A.</given-names></name> <name><surname>Hinaut</surname> <given-names>X.</given-names></name></person-group> (in press). <article-title>Vocal imitation in sensorimotor learning models: a comparative review</article-title>. <source>IEEE Trans. Cogn. Dev. Syst.</source> <pub-id pub-id-type="doi">10.1109/TCDS.2020.3041179</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Spranger</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <source>The Evolution of Grounded Spatial Language</source>. <publisher-name>Language Science Press</publisher-name> (<publisher-loc>Berlin</publisher-loc>).</citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tangiuchi</surname> <given-names>T.</given-names></name> <name><surname>Mochihashi</surname> <given-names>D.</given-names></name> <name><surname>Nagai</surname> <given-names>T.</given-names></name> <name><surname>Uchida</surname> <given-names>S.</given-names></name> <name><surname>Inoue</surname> <given-names>N.</given-names></name> <name><surname>Kobayashi</surname> <given-names>I.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Survey on frontiers of language and robotics</article-title>. <source>Adv. Robot.</source> <volume>33</volume>, <fpage>700</fpage>&#x02013;<lpage>730</lpage>. <pub-id pub-id-type="doi">10.1080/01691864.2019.1632223</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taniguchi</surname> <given-names>A.</given-names></name> <name><surname>Taniguchi</surname> <given-names>T.</given-names></name> <name><surname>Cangelosi</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>Cross-situational learning with bayesian generative models for multimodal category and word learning in robots</article-title>. <source>Front. Neurorobot.</source> <volume>11</volume>:<fpage>66</fpage>. <pub-id pub-id-type="doi">10.3389/fnbot.2017.00066</pub-id><pub-id pub-id-type="pmid">29311888</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taniguchi</surname> <given-names>T.</given-names></name> <name><surname>Nagai</surname> <given-names>T.</given-names></name> <name><surname>Nakamura</surname> <given-names>T.</given-names></name> <name><surname>Iwahashi</surname> <given-names>N.</given-names></name> <name><surname>Ogata</surname> <given-names>T.</given-names></name> <name><surname>Asoh</surname> <given-names>H.</given-names></name></person-group> (<year>2016</year>). <article-title>Symbol emergence in robotics: a survey</article-title>. <source>Adv. Robot.</source> <volume>30</volume>, <fpage>706</fpage>&#x02013;<lpage>728</lpage>. <pub-id pub-id-type="doi">10.1080/01691864.2016.1164622</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Taniguchi</surname> <given-names>T.</given-names></name> <name><surname>Nakamura</surname> <given-names>T.</given-names></name> <name><surname>Suzuki</surname> <given-names>M.</given-names></name> <name><surname>Kuniyasu</surname> <given-names>R.</given-names></name> <name><surname>Hayashi</surname> <given-names>K.</given-names></name> <name><surname>Taniguchi</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Neuro-serket: development of integrative cognitive system through the composition of deep probabilistic generative models</article-title>, in <source>New Generation Computing</source> (<publisher-loc>Cham</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>26</lpage>.</citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taniguchi</surname> <given-names>T.</given-names></name> <name><surname>Ugur</surname> <given-names>E.</given-names></name> <name><surname>Hoffmann</surname> <given-names>M.</given-names></name> <name><surname>Jamone</surname> <given-names>L.</given-names></name> <name><surname>Nagai</surname> <given-names>T.</given-names></name> <name><surname>Rosman</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Symbol emergence in cognitive developmental systems: a survey</article-title>. <source>IEEE Trans. Cogn. Dev. Syst.</source> <volume>11</volume>, <fpage>494</fpage>&#x02013;<lpage>516</lpage>. <pub-id pub-id-type="doi">10.1109/TCDS.2018.2867772</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>The Workshop on Language and Robotics: <ext-link ext-link-type="uri" xlink:href="http://iros2018.emergent-symbol.systems/home">http://iros2018.emergent-symbol.systems/home</ext-link>.</p></fn>
<fn id="fn0002"><p><sup>2</sup>A first workshop was on Deep Probabilistic Generative Models for Cognitive Architecture in Robotics: <ext-link ext-link-type="uri" xlink:href="https://sites.google.com/site/dpgmcar2019/">https://sites.google.com/site/dpgmcar2019/</ext-link>. A second workshop was on Sensorimotor Interaction, Language and Embodiment of Symbols (SMILES): <ext-link ext-link-type="uri" xlink:href="https://sites.google.com/view/smiles-workshop/">https://sites.google.com/view/smiles-workshop/</ext-link>. The tutorial was on Deep Probabilistic Generative Models for Robotics: <ext-link ext-link-type="uri" xlink:href="https://sites.google.com/view/dpgmfr/home">https://sites.google.com/view/dpgmfr/home</ext-link>.</p></fn>
</fn-group>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This Research Topic was partially supported by a Grant-in-Aid for Scientific Research 18H03308 and 16H06569, funded by the Ministry of Education, Culture, Sports, and Science, and Technology, Japan, by CREST (JPMJCR15E3).</p>
</fn>
</fn-group>
</back>
</article>