Editorial: Language and Robotics
- 1College of Information Science and Engineering, Ritsumeikan University, Kyoto, Japan
- 2Graduate School of Engineering Science, Osaka University, Suita, Japan
- 3Institut National de Recherche en Informatique et en Automatique (INRIA), Rocquencourt, Île-de-France, France
- 4Sony Computer Science Laboratories, Tokyo, Japan
- 5Department of Statistical Inference and Mathematics, The Institute of Statistical Mathematics, Tokyo, Japan
Editorial on the Research Topic
Language and Robotics
Language in the real-world environment involves a wide range of challenges in robotics and artificial intelligence (AI). Service robots are required to communicate and collaborate with people using language in the real-world environment. When a robot receives a spoken command from a user in a domestic environment, the robot must understand its meaning in the context of the specific environment. For example, to understand the meaning of “please bring me a pen in Takato's room” the robot needs to know where to find a pen and where Takato's room is. Futhermore, words or expressions (i.e., sounds processed as symbols) can be invented naturally in our daily environment and their meaning can change (Spranger, 2016) over time (i.e., depending on the culture or age of the speaker). Robots thus need to adapt like humans to these versatile aspects of language and demonstrate the ability to learn any language (Hinaut and Twiefel, 2019). In robotics, language understanding inevitably involves multimodal learning, semantic mapping, and behavior learning. To enable a robot to interact orally with people in a long-term manner, we need to develop an AI that makes a robot learn and adapt to language in the real-world environment and in an on-line manner. This topic thus raises several challenges to bridge the gap from low-level sensorimotor interaction (Pagliarini et al., in press) to high-level compositional symbolic communication. Taking inspiration of how children acquire language can help to design the simplest mechanisms to deal with these challenges. Conversely, robotics can help modeling and test hypotheses about language acquisition and language grounding (Cangelosi and Schlesinger, 2015; Taniguchi et al., 2016, 2018; Hinaut and Spranger, 2019), in particular through cross-situational experiments (Taniguchi et al., 2017; Juven and Hinaut, 2020).
Following the successfully organized session “Language and Robotics” held in IEEE-IROS 20181, we organized this Research Topic. We aimed to publish original papers from robotics, natural language processing, machine learning, and cognitive science to share knowledge about the state-of-the-art machine learning methods and perspectives that contribute to modeling language-related capabilities in robotics.
2. About the Research Topic
We are pleased to present five research articles related to semantic mapping, language understanding, motion segmentation, symbol emergence, and language evolution. In this section, we briefly introduce each paper.
First, three papers focused on language-related cognitive capabilities integrating real-world sensor information full of uncertainty and high-dimensional. Each method involves deep learning methods dealing with high-dimensional uncertain real data in robotics. Nagano et al. proposed a new machine learning method called a hierarchical Dirichlet process-variational autoencoder-Gaussian process-hidden semi-Markov model (HVGH). The method extended a hierarchical Dirichlet process-Gaussian process-hidden semi-Markov model (HDP-GP-HSMM) that can automatically segment time series data. HVGH integrated variational autoencoder and the HDP-GP-HSMM and achieved automatic motion segmentation along with representation learning. Katsumata et al. proposed a statistical semantic mapping method called SpCoMapping, which means spatial concept formation and semantic mapping. The proposed model employed Markov random field into a pre-existing spatial concept formation method and became able to learn the arbitrary shape of a place on a map. The method integrated multimodal information, e.g., language, vision, and position, to find semantic information of places. Tada et al. proposed a robust language understanding method by introducing noise injection into the sequence-to-sequence network. Recently, semantic parsing that enables a robot to understand the meaning of human user commands is developed based on deep learning methods. However, semantic parsing in natural language processing does not assume the existence of speech recognition errors. This paper showed the conventional idea of noise injection to sequence-to-sequence network semantic parsing can improve the robustness of a robot's language understanding.
Second, two papers focused on the emergence, or evolution, of symbols. Cambier et al. described the perspectives of language evolution in swarm robotics. They advocated an approach based on language games for the further development of emergent communication in swarm robots. They suggested that swarm robotics can be an ideal testbed to advance research on the emergence of language-like communication. Hagiwara et al. proposed a new computational model representing symbol emergence. The model proposed in this paper regarded symbol emergence as a multiagent multimodal categorization problem. The convergence of the algorithm was guaranteed based on the theory of Markov chain Monte Carlo. This symbol emergence model involved sharing signs among agents and making each agent form internal representations based on its sensorimotor information.
3. Next Step
With the great success of this Research Topic, we organized related workshops and a tutorial2. A survey paper related to this topic has already been published (Tangiuchi et al., 2019). We believe that integrating low-level and high-level cognitive capabilities (Nakamura et al., 2018; Taniguchi et al., 2020) in conjunction with language learning in the real-world environment is crucial to creating an artificial cognitive system, i.e., a robot, which can conduct lifelong learning in the real-world environment and achieves long-term human-robot interaction to support daily human activities. The intersection of language and robotics is a crucial Research Topic for further advancement in robotics and AI. We hope that this special issue will accelerate the cutting-edge studies in robotics and AI that aim to create human-level embodied AI that can communicate and collaborate with people in the real-world environment.
All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.
This Research Topic was partially supported by a Grant-in-Aid for Scientific Research 18H03308 and 16H06569, funded by the Ministry of Education, Culture, Sports, and Science, and Technology, Japan, by CREST (JPMJCR15E3).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The authors gratefully acknowledge the contributions of participants in this special issue.
2. ^A first workshop was on Deep Probabilistic Generative Models for Cognitive Architecture in Robotics: https://sites.google.com/site/dpgmcar2019/. A second workshop was on Sensorimotor Interaction, Language and Embodiment of Symbols (SMILES): https://sites.google.com/view/smiles-workshop/. The tutorial was on Deep Probabilistic Generative Models for Robotics: https://sites.google.com/view/dpgmfr/home.
Hinaut, X., and Spranger, M. (2019). “Learning to parse grounded language using reservoir computing,” in 2019 Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob) (Oslo: IEEE), 266–271.
Hinaut, X., and Twiefel, J. (2019). Teach your robot your language! trainable neural parser for modeling human sentence processing: Examples for 15 languages. IEEE Trans. Cogn. Dev. Syst. 12, 179–188. doi: 10.1109/TCDS.2019.2957006
Juven, A., and Hinaut, X. (2020). “Cross-situational learning with reservoir computing for language acquisition modelling,” in 2020 International Joint Conference on Neural Networks (IJCNN) (Glasgow: IEEE), 1–8.
Nakamura, T., Nagai, T., and Taniguchi, T. (2018). Serket: an architecture for connecting stochastic models to realize a large-scale cognitive model. Front. Neurorobot. 12:25. doi: 10.3389/fnbot.2018.00025
Tangiuchi, T., Mochihashi, D., Nagai, T., Uchida, S., Inoue, N., Kobayashi, I., et al. (2019). Survey on frontiers of language and robotics. Adv. Robot. 33, 700–730. doi: 10.1080/01691864.2019.1632223
Taniguchi, A., Taniguchi, T., and Cangelosi, A. (2017). Cross-situational learning with bayesian generative models for multimodal category and word learning in robots. Front. Neurorobot. 11:66. doi: 10.3389/fnbot.2017.00066
Taniguchi, T., Nakamura, T., Suzuki, M., Kuniyasu, R., Hayashi, K., Taniguchi, A., et al. (2020). “Neuro-serket: development of integrative cognitive system through the composition of deep probabilistic generative models,” in New Generation Computing (Cham), 1–26.
Keywords: language acquisition by robots, multimodal communication, concept formation, symbol grounding in robotics, symbol emergence in robotics, deep learning for robotics, emergence of communication
Citation: Taniguchi T, Horii T, Hinaut X, Spranger M, Mochihashi D and Nagai T (2021) Editorial: Language and Robotics. Front. Robot. AI 8:674832. doi: 10.3389/frobt.2021.674832
Received: 02 March 2021; Accepted: 08 March 2021;
Published: 12 April 2021.
Edited and reviewed by: Mikhail Prokopenko, The University of Sydney, Australia
Copyright © 2021 Taniguchi, Horii, Hinaut, Spranger, Mochihashi and Nagai. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Tadahiro Taniguchi, email@example.com