<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Robot. AI</journal-id>
<journal-title>Frontiers in Robotics and AI</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Robot. AI</abbrev-journal-title>
<issn pub-type="epub">2296-9144</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">679893</article-id>
<article-id pub-id-type="doi">10.3389/frobt.2021.679893</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Robotics and AI</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>When Even a Robot Tutor Zooms: A Study of Embodiment, Attitudes, and Impressions</article-title>
<alt-title alt-title-type="left-running-head">Kanero et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Robot Tutor on Zoom</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Kanero</surname>
<given-names>Junko</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1073646/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tunal&#x131;</surname>
<given-names>Elif Tutku</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1357126/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Oran&#xe7;</surname>
<given-names>Cansu</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/503718/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>G&#xf6;ksun</surname>
<given-names>Tilbe</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/244199/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>K&#xfc;ntay</surname>
<given-names>Aylin C.</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/919026/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Faculty of Arts and Social Sciences, Sabanc&#x131; University, <addr-line>Istanbul</addr-line>, <country>Turkey</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>MPRG iSearch, Max Planck Institute for Human Development, <addr-line>Berlin</addr-line>, <country>Germany</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Department of Psychology, Ko&#xe7; University, <addr-line>Istanbul</addr-line>, <country>Turkey</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/382549/overview">Wafa Johal</ext-link>, University of New South Wales, Australia</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1269751/overview">Pauline Chevalier</ext-link>, Italian Institute of Technology (IIT), Italy</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/408305/overview">Hatice Kose</ext-link>, Istanbul Technical University, Turkey</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Junko Kanero, <email>jkanero@sabanciuniv.edu</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Human-Robot Interaction, a section of the journal Frontiers in Robotics and&#x20;AI</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>30</day>
<month>06</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>8</volume>
<elocation-id>679893</elocation-id>
<history>
<date date-type="received">
<day>12</day>
<month>03</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>25</day>
<month>05</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Kanero, Tunal&#x131;, Oran&#xe7;, G&#xf6;ksun and K&#xfc;ntay.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Kanero, Tunal&#x131;, Oran&#xe7;, G&#xf6;ksun and K&#xfc;ntay</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>This study used an online second language (L2) vocabulary lesson to evaluate whether the physical body (i.e.,&#x20;embodiment) of a robot tutor has an impact on how the learner learns from the robot. In addition, we tested how individual differences in attitudes toward robots, first impressions of the robot, anxiety in learning L2, and personality traits may be related to L2 vocabulary learning. One hundred Turkish-speaking young adults were taught eight English words in a one-on-one Zoom session either with a NAO robot tutor (<italic>N</italic>&#x20;&#x3d; 50) or with a voice-only tutor (<italic>N</italic>&#x20;&#x3d; 50). The findings showed that participants learned the vocabulary equally well from the robot and voice tutors, indicating that the physical embodiment of the robot did not change learning gains in a short vocabulary lesson. Further, negative attitudes toward robots had negative effects on learning for participants in the robot tutor condition, but first impressions did not predict vocabulary learning in either of the two conditions. L2 anxiety, on the other hand, negatively predicted learning outcomes in both conditions. We also report that attitudes toward robots and the impressions of the robot tutor remained unchanged before and after the lesson. As one of the first to examine the effectiveness of robots as an online lecturer, this study presents an example of comparable learning outcomes regardless of physical embodiment.</p>
</abstract>
<kwd-group>
<kwd>human-robot interaction</kwd>
<kwd>second language learning (L2 learning)</kwd>
<kwd>embodiment</kwd>
<kwd>attitudes</kwd>
<kwd>impressions</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>
<italic>Social robots</italic>, robots that interact and communicate with humans by following the behavioral norms of human-human interactions (e.g., <xref ref-type="bibr" rid="B6">Bartneck and Forlizzi, 2004</xref>; <xref ref-type="bibr" rid="B17">Kanero et&#x20;al., 2018</xref>), are becoming abundant across a variety of settings such as homes, hospitals, and schools. A particularly interesting application of social robots is language education because of the significance of the topic as well as the unique characteristics of social robots. Language education is critical for people of all ages. For children, language abilities are known to predict future academic achievement and social skills (<xref ref-type="bibr" rid="B14">Hoff, 2013</xref>; <xref ref-type="bibr" rid="B42">Milligan et al. 2007</xref>); for adults, language skills can broaden social and occupational opportunities (e.g., <xref ref-type="bibr" rid="B28">Paolo and Tansel, 2015</xref>). Learning another language can also contribute to the development of cognitive skills in children (<xref ref-type="bibr" rid="B21">Kov&#xe1;cs and Mehler, 2009</xref>), and the attainment of them in older adults (<xref ref-type="bibr" rid="B7">Bialystok et&#x20;al., 2004</xref>). Importantly, a wealth of research in psychology and education suggests that learning both first (L1) and second language (L2) requires <italic>interactions</italic> (<xref ref-type="bibr" rid="B35">Verga and Kotz, 2013</xref>; <xref ref-type="bibr" rid="B20">Konishi et&#x20;al., 2014</xref>; <xref ref-type="bibr" rid="B25">Lytle and Kuhl, 2017</xref>). As a social agent with a physical body, a robot can play the role of a tutor through vocal, gestural, and facial expressions to provide an interactive learning experience (<xref ref-type="bibr" rid="B13">Han&#x20;et&#x20;al., 2008</xref>; <xref ref-type="bibr" rid="B19">Kennedy et&#x20;al., 2015</xref>; <xref ref-type="bibr" rid="B17">Kanero et&#x20;al., 2018</xref>). The current study focuses on <italic>embodiment</italic> and examines whether and how important it is for L2 learners to interact with a robot tutor with a physical&#x20;body.</p>
<p>The general bodily affordances of social robots were suggested to improve the learning experience as they can engage with the learners&#x2019; physical world and elicit social behaviors from them (<xref ref-type="bibr" rid="B38">Belpaeme et&#x20;al., 2018</xref>). For instance, when teaching a new word, robots can perform gestures with their hands to depict the target object or direct the learner&#x2019;s attention to the object with their eyes, both of which are an integral part of interacting and learning with robots (<xref ref-type="bibr" rid="B1">Admoni and Scassellati, 2017</xref>; <xref ref-type="bibr" rid="B17">Kanero et&#x20;al., 2018</xref>). Some studies indicate that interacting with a robot in person or through a screen may not have much of a difference in terms of learning (e.g., <xref ref-type="bibr" rid="B19">Kennedy et&#x20;al., 2015</xref>), and studies on language learning with intelligent virtual agents provide support to this (<xref ref-type="bibr" rid="B26">Macedonia et&#x20;al., 2014</xref>). In fact, a study on second language learning found participants performing worse after interacting with a physically present robot as opposed to its virtual version or a voice-only tutor, speculatively because it was too novel and interesting, hence distracting, for the participants (<xref ref-type="bibr" rid="B31">Rosenthal-von der P&#xfc;tten et&#x20;al., 2016</xref>). On the other hand, there is also research suggesting that interacting with a physically present robot may yield better outcomes. For instance, one study found that adults performed better in solving logic puzzles when they were partnered off with a physically present robot as opposed to a disembodied voice or a video of a robot (<xref ref-type="bibr" rid="B22">Leyzberg et&#x20;al., 2012</xref>), though solving a logic puzzle is inherently different from learning a language.</p>
<p>
<italic>Embodiment</italic> has been defined in many different ways partially because the term is used in various disciplines including philosophy, psychology, computer science, and robotics (see <xref ref-type="bibr" rid="B10">Deng et&#x20;al., 2019</xref>). One of the clear definitions provided by roboticists is that of <xref ref-type="bibr" rid="B29">Pfeifer and Scheier (1999)</xref>: &#x201c;In artificial systems, the term refers to the fact that a particular agent is realized as a physical robot or as a simulated agent&#x201d; (p. 649). Focusing on the social aspect, <xref ref-type="bibr" rid="B5">Barsalou et&#x20;al. (2003)</xref> states that embodiment is the &#x201c;states of the body, such as postures, arm movements, and facial expressions, [which] arise during social interaction and play central roles in social information processing&#x201d; (p. 43). In human-robot interaction, <xref ref-type="bibr" rid="B23">Li (2015)</xref> made a distinction between what he calls <italic>physical presence</italic> and <italic>physical embodiment</italic> to systematically evaluate the different bodily affordances of robots. According to Li (2015), <italic>physical presence</italic> differentiates a robot in the same room with the user and a robot appearing on the screen. On the other hand, <italic>physical embodiment</italic> differentiates a (co-present or telepresent) materialized robot and a virtual agent (e.g., a computer-generated image of a robot).</p>
<p>The review by <xref ref-type="bibr" rid="B23">Li (2015)</xref> concluded that the physical presence of the robot, but not its embodiment, has a positive influence on social interactions. Critically, however, the conclusion was drawn based on four studies from three publications only. Overall, while previous research provides valuable insights into how different dimensions of physicality influence human-robot interaction, they fall short in revealing the difference between having and not having a body and face on learning outcomes. Although their appearance can simulate different animate agents such as a human or an animal, all social robots have a body and face. How does this influence people&#x2019;s learning, as opposed to not having either? Following the distinctions drawn by <xref ref-type="bibr" rid="B23">Li (2015)</xref>, we compare a robot tutor (embodied but not physically present) with a voice-only tutor (not embodied nor physically present) in an online lesson to understand the effects of physical embodiment.</p>
<p>Research also suggests that embodiment may have different implications for different people, as in individuals with Autism Spectrum Disorder struggling with understanding the emotions of a virtual agent than a real agent, whether it is a robot or a human, in contrast to typically developed individuals (<xref ref-type="bibr" rid="B8">Chevalier et&#x20;al., 2017</xref>). People&#x2019;s varying attitudes toward robots may also influence their preference for a physical or virtual robot (<xref ref-type="bibr" rid="B24">Ligthart and Truong, 2015</xref>). Another study with children also suggests that age and experience may diminish the effect of physical presence, as it found that younger children with hearing impairments learned more words in sign language when they interacted with a physically present robot than a video of it, whereas older children with more experience in sign language equally benefited from both (<xref ref-type="bibr" rid="B40">K&#xf6;se et&#x20;al., 2015</xref>). Therefore, the current study further explores interrelations among individual differences (specifically attitudes toward robots, first impressions of the robot tutor, anxiety about learning a second language, and personality traits) and learning outcomes across different degrees of embodiment.</p>
<p>Although not much is known specifically about the effects of individual differences in learning with robots, some studies have explored how attitudes and personality are related to the ways in which a person interacts with a robot. For example, the patterns of speech and eye gaze were observed while adults built an object with a humanoid robot (<xref ref-type="bibr" rid="B16">Ivaldi et&#x20;al., 2017</xref>). The study found that individuals with negative attitudes toward robots tended to look less at the robot&#x2019;s face and more at the robot&#x2019;s hands. In another study, when approached by a robot, individuals with high levels of negative attitudes toward robots and the personality trait of neuroticism kept a larger personal space between the robot and themselves (<xref ref-type="bibr" rid="B34">Takayama and Pantofaru, 2009</xref>).</p>
<p>In the case of language learning, <xref ref-type="bibr" rid="B18">Kanero et&#x20;al. (2021)</xref> were first to examine how attitudes toward robots, anxiety about learning L2, and personality may predict the learning outcomes of a robot-led L2 vocabulary lesson. The study found that negative attitudes measured through the Negative Attitudes toward Robots Scale (NARS; <xref ref-type="bibr" rid="B27">Nomura et&#x20;al., 2006</xref>) as well as anxiety about learning L2 measured through the Foreign Language Classroom Anxiety Scale (FLCAS; <xref ref-type="bibr" rid="B15">Horwitz et&#x20;al., 1986</xref>) predicted the number of words participants learned in an in-person vocabulary lesson with a robot tutor. The results also showed that the robot was an effective language tutor, akin to a human tutor. However, it is unclear whether the tutor robot is as effective when it is not physically present, and whether individual differences such as attitudes toward robots and L2 anxiety predict the learning outcomes for a telepresent robot&#x20;tutor.</p>
<p>In addition to the individual difference measures used in the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), the current study also assesses the learners&#x2019; impressions of robots, which are expected to affect their engagement in the long run. Previous studies in human-human interaction suggest that the first impression is formed very quickly after just seeing a picture of an individual and might remain unchanged even after meeting and interacting with the same individual in person (e.g., <xref ref-type="bibr" rid="B12">Gunaydin et&#x20;al., 2017</xref>; see also <xref ref-type="bibr" rid="B36">Willis and Todorov, 2006</xref>). However, it is unclear whether the same principle applies to commercial social robots (e.g., NAO), which are inanimate objects with a homogeneous appearance shared across individuals. Therefore, in the current study, we included an additional measure to examine if the impressions of the robot have a role in robot-led learning. Further, we evaluate whether the impressions of the robot tutor as well as attitudes toward robots change before and after interacting with the robot&#x20;tutor.</p>
<p>In summary, this study explores the impact of having a body in robot-led language lessons by comparing a robot tutor and a voice-only tutor in terms of learning outcomes as well as the influence of attitudes, impressions, L2 anxiety, personality. We also report the details of the learner&#x2019;s general attitudes toward roborts, impressions of the robot tutor, and preferences to the specific type of tutor (robot vs. voice vs. human). In the Discussion, we also compare our data to the data of the previous study (<xref ref-type="bibr" rid="B18">Kanero et al., 2021</xref>) to address whether the physical presence in robot-led language lessons would affect these factors.</p>
</sec>
<sec sec-type="methods" id="s2">
<title>Methods</title>
<sec id="s2-1">
<title>Participants</title>
<p>The dataset consisted of 100 native Turkish-speaking young adults: 50 in the robot tutor condition (<italic>age range</italic> &#x3d; 18&#x2013;32&#xa0;years; <italic>M</italic>
<sub>
<italic>ag</italic>e</sub> &#x3d; 23.49&#xa0;years; <italic>SD</italic> &#x3d; 2.53; 33 females, 17 males), and 50 in the voice tutor condition (<italic>age range</italic> &#x3d; 18&#x2013;35&#xa0;years; <italic>M</italic>
<sub>
<italic>age</italic>
</sub> &#x3d; 24.15&#xa0;years; <italic>SD</italic> &#x3d; 3.62; 33 females, 16 males, 1 other). We relied on a convenience sample, and participants were recruited through advertisements on social media as well as word of mouth. Before the lesson, the average English test score of participants (<xref ref-type="bibr" rid="B30">Quick Placement Test</xref>; <xref ref-type="bibr" rid="B30">University of Cambridge Local Examinations Syndicate [UCLES], 2001)</xref> was 39.68 out of 60 in the robot tutor condition (<italic>score range</italic> &#x3d; 16&#x2013;58; <italic>SD</italic> &#x3d; 9.07) and 37.64 in the voice tutor condition (<italic>score range</italic> &#x3d; 20&#x2013;55; <italic>SD</italic> &#x3d; 9.25). Participants had no known vision or hearing impairments. One participant in the robot tutor condition did not show up for the second session, and thus the two delayed language tests and the post-lesson survey were not administered to this participant. In addition, one participant in the robot tutor condition was not taught one of the eight vocabulary words due to a technical error, and thus the test data for that word were not used. Participants were given a gift card for their participation.</p>
</sec>
<sec id="s2-2">
<title>Materials and Procedures</title>
<p>The experiment was completed via the online video call software <italic>Zoom</italic> (<ext-link ext-link-type="uri" xlink:href="https://zoom.us/">https://zoom.us</ext-link>) in two sessions. In the first session, participants first filled out a demographic form. They then completed a short English language test (Quick Placement Test; <xref ref-type="bibr" rid="B30">UCLES, 2001</xref>), and a questionnaire assessing their attitudes toward robots, L2 anxiety, personality traits, and their impression of the robot or voice tutor. The test and questionnaires were administered using the online survey platform Qualtrics (<ext-link ext-link-type="uri" xlink:href="https://www.qualtrics.com">https://www.qualtrics.com</ext-link>). Then, participants received a one-on-one English lesson either from the robot or the voice tutor. For the lesson, participants were sent to a breakout room<xref ref-type="fn" rid="FN1">
<sup>1</sup>
</xref>, and participants were alone with the tutor. Immediately after the lesson, participants in both conditions completed two measures of learning (i.e.,&#x20;immediate production and receptive tests). The second session took place one week later, and participants connected via Zoom again and completed the same vocabulary tests (i.e.,&#x20;delayed production and receptive tests). The same set of tests and surveys were administered in the robot tutor and voice tutor conditions, but in the voice tutor condition, the term &#x201c;voice assistant&#x201d; was used in place of &#x201c;robot&#x201d; for the surveys on the attitudes, impressions, and preference (see <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> for a schematic representation of the procedure, and <xref ref-type="fig" rid="F2">Figure&#x20;2</xref> for the appearance of the robot and voice-only tutors).</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The procedure of the lesson from the participant&#x2019;s perspective. In the voice-only tutor condition, the voice sound spectrum appeared instead of the robot (see <xref ref-type="fig" rid="F2">Figure&#x20;2</xref> and <xref ref-type="sec" rid="s10">Supplementary Material</xref>). Step 4 (Production and Receptive Tests) was repeated one week later.</p>
</caption>
<graphic xlink:href="frobt-08-679893-g001.tif"/>
</fig>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The appearance of the robot tutor <bold>(A)</bold> and the voice tutor <bold>(B)</bold>. See <xref ref-type="sec" rid="s10">Supplementary Material</xref> for the videos of the robot and voice tutors).</p>
</caption>
<graphic xlink:href="frobt-08-679893-g002.tif"/>
</fig>
<sec id="s2-2-1">
<title>Negative Attitudes Toward Robots</title>
<p>
<italic>Negative Attitudes toward Robots Scale</italic> (NARS; <xref ref-type="bibr" rid="B27">Nomura et&#x20;al., 2006</xref>) was used to assess attitudes toward robots. The NARS consists of 14 questions divided into three subordinate scales: negative attitude toward interacting with robots (S1), negative attitude toward the social influence of robots (S2), and negative attitude toward emotions involved in the interaction with robots (S3). The Turkish version of the NARS (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>) was used. Participants rated how well each of the statements represented their attitudes toward robots on a scale of 1&#x2013;5 (1: I strongly disagree/Kesinlikle kat&#x131;lm&#x131;yorum, 2: I disagree/Kat&#x131;lm&#x131;yorum, 3: Undecided/Karars&#x131;z&#x131;m, 4: I agree/Kat&#x131;l&#x131;yorum, 5: I strongly agree/Kesinlikle kat&#x131;l&#x131;yorum). In the voice tutor condition, the word &#x201c;robot&#x201d; on the NARS scale was replaced by &#x201c;voice assistant.&#x201d;</p>
</sec>
<sec id="s2-2-2">
<title>Impressions of the Robot Tutor</title>
<p>To assess participants&#x2019; impressions of the robot/voice tutor, we administered an impression survey with 17 questions used by Gunaydin and her colleagues (2017; available publicly at <ext-link ext-link-type="uri" xlink:href="https://osf.io/nhmtw/?view_only=9f6efafeba4b48dc9b6a73b6a3d145ee">https://osf.io/nhmtw/?view_only&#x3d;9f6efafeba4b48dc9b6a73b6a3d145ee</ext-link>). The survey shows a photograph of the robot or voice tutor, depending on the condition, and consists of two parts: The first eight questions ask participants to rate their willingness to engage and interact with the target in the future (e.g., This robot/voice assistant seems like a robot/voice assistant I would like to get to know/Tan&#x131;mak istedi&#x11f;im bir robot/sesli asistan gibi g&#xf6;z&#xfc;k&#xfc;yor) on a scale of 1&#x2013;7 (1: I fully disagree/Hi&#xe7; kat&#x131;lm&#x131;yorum, 2: I disagree/Kat&#x131;lm&#x131;yorum, 3: I somewhat disagree/K&#x131;smen kat&#x131;lm&#x131;yorum, 4: I neither agree nor disagree/Ne kat&#x131;l&#x131;yorum ne de kat&#x131;lm&#x131;yorum, 5: I somewhat agree/K&#x131;smen kat&#x131;l&#x131;yorum, 6: I agree/kat&#x131;l&#x131;yorum, 7: I fully agree/Tamamen kat&#x131;l&#x131;yorum). The next nine questions ask participants to rate how their interaction will be with the robot/voice assistant (e.g., How much do you think you will like this robot/voice assistant?/Bu robotu/sesli asistan&#x131; ne kadar sevece&#x11f;inizi d&#xfc;&#x15f;&#xfc;n&#xfc;yorsunuz?) on a scale of 1&#x2013;7 (1: Not at all/Hi&#xe7;, 7: Very much/&#xc7;ok fazla). After the lesson, participants rated the same items but were told to rate the statements based on their interactions with their tutor. The original survey in English was translated into Turkish by the second author and research assistants who are native speakers of Turkish. To adapt to our study, the word &#x201c;person&#x201d; was replaced with &#x201c;robot&#x201d; for the robot tutor condition and &#x201c;voice assistant&#x201d; for the voice tutor condition.</p>
</sec>
<sec id="s2-2-3">
<title>L2 Anxiety</title>
<p>The Turkish version of the Foreign Language Classroom Anxiety Scale (FLCAS; <xref ref-type="bibr" rid="B15">Horwitz et&#x20;al., 1986</xref>) translated by <xref ref-type="bibr" rid="B4">Ayd&#x131;n et&#x20;al. (2016)</xref> was administered. The FLCAS consists of 33 statements (e.g., I never feel quite sure of myself when I am speaking in my foreign language class/Yabanc&#x131; dil derslerinde konu&#x15f;urken kendimden asla emin olam&#x131;yorum.) to be rated on a scale of 1&#x2013;5 (1: I fully disagree/Hi&#xe7; kat&#x131;lm&#x131;yorum, 2: I disagree/Kat&#x131;lm&#x131;yorum, 3: I neither agree nor disagree/Ne kat&#x131;l&#x131;yorum ne de kat&#x131;lm&#x131;yorum, 4: I agree/Kat&#x131;l&#x131;yorum, 5: I fully agree/Tamamen kat&#x131;l&#x131;yorum).</p>
</sec>
<sec id="s2-2-4">
<title>Personality Traits</title>
<p>The Turkish version of a personality inventory was used to test the five personality traits &#x2013; openness to experience, conscientiousness, extroversion, agreeableness, and neuroticism (<xref ref-type="bibr" rid="B9">Demir and Kumkale, 2013</xref>). This survey included 44 questions addressing each of the five traits &#x2013; 7 items for conscientiousness (e.g., I stick to my plans/Yapt&#x131;&#x11f;&#x131;m planlara sad&#x131;k kal&#x131;r&#x131;m); 10 items for neuroticism (e.g., I am depressed/Depresifimdir); 9 items for each of openness to experience (e.g., My interests are very diverse/&#x130;lgi alanlar&#x131;m &#xe7;ok &#xe7;e&#x15f;itlidir), extroversion (e.g., I am talkative/Konu&#x15f;kan&#x131;md&#x131;r), and agreeableness (e.g., I am helpful/Yard&#x131;mseverimdir). Participants rated how well each of the statements represented their personality on a scale of 1&#x2013;5 (1: I strongly disagree/Kesinlikle kat&#x131;lm&#x131;yorum, 2: I disagree/Kat&#x131;lm&#x131;yorum, 3: I neither agree nor disagree/Ne kat&#x131;l&#x131;yorum, ne de kat&#x131;lm&#x131;yorum, 4: I agree/Kat&#x131;l&#x131;yorum, 5: I strongly agree/Kesinlikle kat&#x131;l&#x131;yorum).</p>
</sec>
<sec id="s2-2-5">
<title>Post-Lesson Vocabulary Tests</title>
<p>Immediately after the lesson, we first administered the production vocabulary test (hereafter the immediate production test), and then the receptive vocabulary test (hereafter the immediate receptive test). To assess to what extent vocabulary was retained over time, participants completed the same measures again after a delay of one week (delayed post-lesson tests). The definitions of the target words used in the production test were the same as the definitions used in the lesson. In the receptive test, the pictures from the Peabody Picture Vocabulary Test, Fourth Edition (PPVT-4; <xref ref-type="bibr" rid="B11">Dunn and Dunn, 2007</xref>), which correspond to the target words, were used. In the production test, the experimenter provided the definitions of the learned English words one by one in a randomized order, and the participant was asked to say the corresponding English word. In the receptive test, the participant heard the learned English word and was asked to choose a picture that matched the word from four options. The delayed post-lesson tests were conducted via Zoom seven days after the lesson. Due to schedule conflicts, however, three participants in the robot tutor condition and two participants in the voice tutor condition completed these tests after six days, while four participants in each condition completed the tests after eight days. Also, three participants in the voice tutor condition completed the test after nine days.</p>
</sec>
<sec id="s2-2-6">
<title>Tutor Preference</title>
<p>After the delayed post-lesson tests, we also asked participants to rate how much they want to learn English from a human, a robot, and a voice assistant. A scale of 1&#x2013;5 was used (1: I certainly do not want/Kesinlikle istemem, 2: I do not want/&#x130;stemem, 3: I neither want nor not want/Ne isterim ne istemem, 4: I want/&#x130;sterim, 5: I certainly want/Kesinlikle isterim).</p>
</sec>
<sec id="s2-2-7">
<title>English Lesson With the Robot or Voice Tutor</title>
<p>Following the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), participants were taught eight English nouns &#x2013; upholstery, barb, angler, caster, dromedary, cairn, derrick, and cupola (see <xref ref-type="table" rid="T1">Table&#x20;1</xref>; see <xref ref-type="bibr" rid="B18">Kanero et&#x20;al. (2021)</xref> for the details of the word selection process).</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>The target words and their definitions used in the&#x20;study.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Word</th>
<th align="center">Definition</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Upholstery</td>
<td align="left">
<italic>Bu kelime d&#xf6;&#x15f;emelik kuma&#x15f; anlam&#x131;na gelir</italic> (This word means fabric used to make a soft covering)</td>
</tr>
<tr>
<td align="left">barb</td>
<td align="left">
<italic>Bu kelime &#xe7;engel ya da kanca anlam&#x131;na gelir</italic> (This word means the tip of an arrow or fishhook)</td>
</tr>
<tr>
<td align="left">Angler</td>
<td align="left">
<italic>Bu kelime olta ile bal&#x131;k tutan kimse anlam&#x131;na gelir</italic> (This word means a person who fishes with hook and line)</td>
</tr>
<tr>
<td align="left">Caster</td>
<td align="left">
<italic>Bu kelime bir &#x15f;eye tak&#x131;lan k&#xfc;&#xe7;&#xfc;k tekerlek anlam&#x131;na gelir</italic> (This word means a little wheel attached to something)</td>
</tr>
<tr>
<td align="left">Dromedary</td>
<td align="left">
<italic>Bu kelime tek h&#xf6;rg&#xfc;&#xe7;l&#xfc; deve anlam&#x131;na gelir</italic> (This word means a one-humped camel)</td>
</tr>
<tr>
<td align="left">Cairn</td>
<td align="left">
<italic>Bu kelime ta&#x15f; y&#x131;&#x11f;&#x131;n&#x131; anlam&#x131;na gelir</italic> (This word means a mound of stones)</td>
</tr>
<tr>
<td align="left">Derrick</td>
<td align="left">
<italic>Bu kelime petrol kuyusu &#xfc;zerindeki kule anlam&#x131;na gelir</italic> (This word means a tower over an oil well)</td>
</tr>
<tr>
<td align="left">Cupola</td>
<td align="left">
<italic>Bu kelime bir &#xe7;at&#x131; &#xfc;st&#xfc;ne in&#x15f;a edilen k&#xfc;&#xe7;&#xfc;k kubbe benzeri yap&#x131; anlam&#x131;na gelir</italic> (This word means a rounded vault-like structure built on top of a roof)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In both tutor conditions, the robot or voice tutor briefly chatted with the participant and explained the structure of the lesson first, and then introduced the words one by one. Each target word was taught in four steps:<list list-type="simple">
<list-item>
<p>1) The tutor introduced the target L2 (English) word and asked the participant whether the participant already knew the word (Note that none of the participants knew any of the words).</p>
</list-item>
<list-item>
<p>2) The tutor introduced the definition of the target word in L1 (Turkish, see <xref ref-type="table" rid="T1">Table&#x20;1</xref>).</p>
</list-item>
<list-item>
<p>3) The tutor asked the participant to utter the target word following the tutor, three&#x20;times.</p>
</list-item>
<list-item>
<p>4) The tutor again defined the word and asked the participant to repeat the definition.</p>
</list-item>
</list>
</p>
<p>After learning every two target words, the participant was given a mini quiz in which the tutor provided the definitions of the target words and asked the participant for the corresponding word. The lesson lasted for about 15&#xa0;min. At the end of the lesson, the robot or the voice tutor asked the participant to return to the previous room and find the experimenter they met prior to the lesson. The human experimenter administered the immediate production and receptive vocabulary&#x20;tests.</p>
<p>To use the same voice in English and Turkish speech, we recorded the speech of a female bilingual experimenter and added sound effects to make the speech sound robot-like. The same set of speech sounds were used for both the robot and voice tutors. The visuals of both tutors were presented as a series of seamlessly transitioning videos on Zoom. The movements of the robot tutor were filmed (see <xref ref-type="fig" rid="F2">Figure&#x20;2A</xref>), whereas the soundwaves of the voice tutor were created using Adobe After Effect (<ext-link ext-link-type="uri" xlink:href="https://www.adobe.com/products/aftereffects.html">https://www.adobe.com/products/aftereffects.html</ext-link>; <xref ref-type="fig" rid="F2">Figure&#x20;2B</xref>).</p>
<p>The robot tutor provided no facial expressions but moved its head and arms during the lesson to keep the participant engaged. Most actions were chosen from the Animated Speech library of SoftBank Robotics (<ext-link ext-link-type="uri" xlink:href="http://doc.aldebaran.com/2-1/naoqi/audio/alanimatedspeech_advanced.html">http://doc.aldebaran.com/2-1/naoqi/audio/alanimatedspeech_advanced.html</ext-link>), although some were created by the first author to better suit the lesson.<xref ref-type="fn" rid="FN2">
<sup>2</sup>
</xref> While pronouncing the target L2 word and its definition, the robot stood still without any movements to avoid the motor sound of the robot hindering the hearing. There were unavoidable behavioral differences between the two tutors (e.g., the motor sound of the robot), but otherwise, the differences between the two tutors were kept minimal.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>Robot vs. Voice Tutor</title>
<p>We first examined if participants in the robot tutor and voice tutor conditions differed in their post-lesson test scores. We compared the two tutor conditions across all four learning outcome measures: immediate production test, immediate receptive test, delayed production test, and delayed receptive test. We conducted simple Generalized Linear Mixed Models (GLMMs) on each post-lesson test with Tutor Type (robot vs. voice) as a fixed effect and Word as random intercepts.<xref ref-type="fn" rid="FN3">
<sup>3</sup>
</xref> In this model, we also added the pre-lesson English test scores as an additional fixed effect to control for the difference in English proficiency between the conditions. As shown in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>, participants did not differ in terms of learning outcomes across conditions (immediate production test, <italic>B</italic>&#x20;&#x3d; 0.01, <italic>SE</italic> &#x3d; 0.16, <italic>Z</italic>&#x20;&#x3d; 0.04, <italic>p</italic>&#x20;&#x3d; 0.968; delayed production test, <italic>B</italic>&#x20;&#x3d; -0.29, <italic>SE</italic> &#x3d; 0.20, <italic>Z</italic>&#x20;&#x3d; -1.44, <italic>p</italic>&#x20;&#x3d; 0.149; immediate receptive test, <italic>B</italic>&#x20;&#x3d; 0.04, <italic>SE</italic> &#x3d; 0.18, <italic>Z</italic>&#x20;&#x3d; 0.20, <italic>p</italic>&#x20;&#x3d; 0.845; or the delayed receptive test, <italic>B</italic>&#x20;&#x3d; -0.26, <italic>SE</italic> &#x3d; 0.16, <italic>Z</italic>&#x20;&#x3d; -1.64, <italic>p</italic>&#x20;&#x3d; 0.101).</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Mean number of correct answers in the robot tutor and voice tutor conditions in the four post-lesson tests. <italic>N</italic>&#x20;&#x3d; 100 for the immediate production and receptive tests; <italic>N</italic>&#x20;&#x3d; 99 for the delayed production and receptive tests. The highest possible score for each test was eight. The error bars indicate the standard errors.</p>
</caption>
<graphic xlink:href="frobt-08-679893-g003.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>Predicting the Learning Outcomes With Individual Difference Factors</title>
<p>Next, we examined whether some participants learned better or worse from robots depending on their attitudes toward robots, the first impression of the robot or voice tutor, anxiety in L2 learning, and personality traits. As indicated by Cronbach&#x2019;s alphas in <xref ref-type="table" rid="T2">Table&#x20;2</xref>, each of these variables was measured reliably. Therefore, items measuring each construct were averaged to create relevant indices. For NARS, L2 anxiety, and personality, values ranged between 1 and 5. Higher values for NARS indicated having more negative attitudes toward robots; similarly, higher values for L2 Anxiety indicated having greater anxiety. For the impression survey, the values ranged between 1 and 7 and higher values indicated a more positive first impression.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Descriptive statistics for the individual difference measures.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="2" align="center">Robot tutor</th>
<th colspan="2" align="center">Voice tutor</th>
</tr>
<tr>
<th colspan="1" align="left"/>
<th align="center">
<italic>&#x3b1;</italic>
</th>
<th align="center">
<italic>Mean</italic>
</th>
<th align="center">
<italic>SD</italic>
</th>
<th align="center">
<italic>Mean</italic>
</th>
<th align="center">
<italic>SD</italic>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">NARS (14)</td>
<td align="center">0.88</td>
<td align="center">2.71</td>
<td align="center">0.71</td>
<td align="center">2.55</td>
<td align="center">0.63</td>
</tr>
<tr>
<td colspan="1" align="left">L2 anxiety (33)</td>
<td align="center">0.95</td>
<td align="center">2.57</td>
<td align="center">0.79</td>
<td align="center">2.48</td>
<td align="center">0.69</td>
</tr>
<tr>
<td colspan="1" align="left">Personality (44)</td>
<td align="left"/>
<td align="left"/>
<td align="left"/>
<td align="left"/>
<td align="left"/>
</tr>
<tr>
<td colspan="1" align="left">&#x2003;Openness (9)</td>
<td align="center">0.76</td>
<td align="center">4.13</td>
<td align="center">0.55</td>
<td align="center">4.15</td>
<td align="center">0.47</td>
</tr>
<tr>
<td colspan="1" align="left">&#x2003;Conscientiousness (7)</td>
<td align="center">0.81</td>
<td align="center">3.20</td>
<td align="center">0.74</td>
<td align="center">3.22</td>
<td align="center">0.59</td>
</tr>
<tr>
<td colspan="1" align="left">&#x2003;Extroversion (9)</td>
<td align="center">0.92</td>
<td align="center">3.64</td>
<td align="center">0.87</td>
<td align="center">3.86</td>
<td align="center">0.82</td>
</tr>
<tr>
<td colspan="1" align="left">&#x2003;Agreeableness (9)</td>
<td align="center">0.77</td>
<td align="center">3.94</td>
<td align="center">0.52</td>
<td align="center">3.85</td>
<td align="center">0.58</td>
</tr>
<tr>
<td colspan="1" align="left">&#x2003;Neuroticism (10)</td>
<td align="center">0.86</td>
<td align="center">3.37</td>
<td align="center">0.75</td>
<td align="center">3.35</td>
<td align="center">0.71</td>
</tr>
<tr>
<td colspan="1" align="left">Impression (17)</td>
<td align="center">0.92</td>
<td align="center">4.19</td>
<td align="center">1.27</td>
<td align="center">3.51</td>
<td align="center">1.02</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>N&#x20;&#x3d; 100. The number in parenthesis indicates the number of items in the&#x20;scale.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3-3">
<title>Negative Attitudes Toward Robots</title>
<p>We built four separate GLMMs, one for each post-lesson test (immediate production, immediate receptive, delayed production, and delayed receptive), with Word as a random intercept to examine whether negative attitudes toward robots and voice assistants predicted the number of words participants learned. As shown in <xref ref-type="table" rid="T3">Table&#x20;3</xref>, in line with the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), negative attitudes toward robots predicted the learning outcomes in a robot-led vocabulary lesson, though only in the delayed tests. Negative attitudes did not predict learning in the voice tutor condition.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>GLMMs with NARS as the sole predictor for the four post-lesson scores.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="4" align="center">Robot tutor</th>
<th colspan="4" align="center">Voice tutor</th>
</tr>
<tr>
<td colspan="1" align="left"/>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="1" align="left">Immediate production</td>
<td align="center">-0.08</td>
<td align="center">0.15</td>
<td align="center">-0.51</td>
<td align="center">0.610</td>
<td align="center">0.03</td>
<td align="center">0.18</td>
<td align="center">0.20</td>
<td align="center">0.843</td>
</tr>
<tr>
<td colspan="1" align="left">Immediate receptive</td>
<td align="center">-0.20</td>
<td align="center">0.18</td>
<td align="center">-1.14</td>
<td align="center">0.253</td>
<td align="center">0.01</td>
<td align="center">0.20</td>
<td align="center">0.07</td>
<td align="center">0.947</td>
</tr>
<tr>
<td colspan="1" align="left">Delayed production</td>
<td align="center">-0.45</td>
<td align="center">0.22</td>
<td align="center">-2.01</td>
<td align="center">0.045</td>
<td align="center">-0.03</td>
<td align="center">0.21</td>
<td align="center">-0.13</td>
<td align="center">0.895</td>
</tr>
<tr>
<td align="left">Delayed receptive</td>
<td align="center">-0.31</td>
<td align="center">0.15</td>
<td align="center">-2.08</td>
<td align="center">0.038</td>
<td align="center">0.05</td>
<td align="center">0.18</td>
<td align="center">0.29</td>
<td align="center">0.771</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>For the immediate tests, <italic>N</italic>&#x20;&#x3d; 50 in both conditions; For the delayed tests, <italic>N</italic>&#x20;&#x3d; 49 in the robot tutor condition and <italic>N</italic>&#x20;&#x3d; 50 in the voice tutor condition.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3-4">
<title>First Impressions of the Robot</title>
<p>To evaluate the relation between the first impressions of the tutor and the learning outcomes, we followed the same steps and built GLMMs separately for the two tutor conditions. As shown in <xref ref-type="table" rid="T4">Table&#x20;4</xref>, there was no significant relation between learning outcomes and first impression in either condition.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>GLMMs with the first impression as the sole predictor for the four post-lesson scores.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="4" align="center">Robot tutor</th>
<th colspan="4" align="center">Voice tutor</th>
</tr>
<tr>
<td colspan="1" align="left"/>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="1" align="left">Immediate production</td>
<td align="center">-0.06</td>
<td align="center">0.09</td>
<td align="center">-0.68</td>
<td align="center">0.500</td>
<td align="center">0.04</td>
<td align="center">0.11</td>
<td align="center">0.37</td>
<td align="center">0.715</td>
</tr>
<tr>
<td colspan="1" align="left">Immediate receptive</td>
<td align="center">0.04</td>
<td align="center">0.10</td>
<td align="center">0.42</td>
<td align="center">0.678</td>
<td align="center">0.05</td>
<td align="center">0.13</td>
<td align="center">0.43</td>
<td align="center">0.664</td>
</tr>
<tr>
<td colspan="1" align="left">Delayed production</td>
<td align="center">-0.03</td>
<td align="center">0.11</td>
<td align="center">-0.26</td>
<td align="center">0.797</td>
<td align="center">0.01</td>
<td align="center">0.13</td>
<td align="center">0.05</td>
<td align="center">0.959</td>
</tr>
<tr>
<td colspan="1" align="left">Delayed receptive</td>
<td align="center">0.02</td>
<td align="center">0.08</td>
<td align="center">0.24</td>
<td align="center">0.811</td>
<td align="center">-0.04</td>
<td align="center">0.11</td>
<td align="center">-0.33</td>
<td align="center">0.738</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>For the immediate tests, <italic>N</italic>&#x20;&#x3d; 50 in both conditions; For the delayed tests, <italic>N</italic>&#x20;&#x3d; 49 in the robot tutor condition and <italic>N</italic>&#x20;&#x3d; 50 in the voice tutor condition.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3-5">
<title>L2 Anxiety and Personality Traits</title>
<p>The influence of L2 learning anxiety was similarly examined by building a GLMM for each post-lesson test for the robot tutor and voice tutor conditions with Word as a random intercept. In the robot tutor condition, L2 Anxiety predicted the scores of most tests except the immediate receptive test; in the voice tutor condition, the significance was found in the delayed production and receptive tests (<xref ref-type="table" rid="T5">Table&#x20;5</xref>).</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>GLMMs with L2 Anxiety as the sole predictor for the four post-lesson scores.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="4" align="center">Robot tutor condition</th>
<th colspan="4" align="center">Voice tutor condition</th>
</tr>
<tr>
<td colspan="1" align="left"/>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
<td align="center">
<italic>B</italic>
</td>
<td align="center">
<italic>SE</italic>
</td>
<td align="center">
<italic>Z</italic>
</td>
<td align="center">
<italic>p</italic>
</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="1" align="left">Immediate production</td>
<td align="center">-0.43</td>
<td align="center">0.14</td>
<td align="center">-2.97</td>
<td align="center">0.003</td>
<td align="center">-0.16</td>
<td align="center">0.16</td>
<td align="center">-1.00</td>
<td align="center">0.315</td>
</tr>
<tr>
<td colspan="1" align="left">Immediate receptive</td>
<td align="center">-0.22</td>
<td align="center">0.16</td>
<td align="center">-1.35</td>
<td align="center">0.176</td>
<td align="center">-0.15</td>
<td align="center">0.18</td>
<td align="center">-0.83</td>
<td align="center">0.404</td>
</tr>
<tr>
<td colspan="1" align="left">Delayed production</td>
<td align="center">-0.42</td>
<td align="center">0.20</td>
<td align="center">-2.08</td>
<td align="center">0.037</td>
<td align="center">-1.36</td>
<td align="center">0.27</td>
<td align="center">-5.04</td>
<td align="center">&#x3c;0.001</td>
</tr>
<tr>
<td colspan="1" align="left">Delayed receptive</td>
<td align="center">-0.31</td>
<td align="center">0.14</td>
<td align="center">-2.25</td>
<td align="center">0.025</td>
<td align="center">-0.41</td>
<td align="center">0.16</td>
<td align="center">-2.53</td>
<td align="center">0.012</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>For the immediate tests, <italic>N</italic>&#x20;&#x3d; 50 in both conditions; For the delayed tests, <italic>N</italic>&#x20;&#x3d; 49 in the robot tutor condition and <italic>N</italic>&#x20;&#x3d; 50 in the voice tutor condition.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>We also built four GLMMs for each post-lesson test to evaluate the relevance of personality traits. In concert with the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), the personality traits were not reliable predictors of the learning outcomes of the robot-led L2 lesson. In the robot tutor condition, extroversion was positively correlated with the immediate receptive test scores (<italic>B</italic>&#x20;&#x3d; 0.41, <italic>SE</italic> &#x3d; 0.17, <italic>Z</italic>&#x20;&#x3d; 2.35, <italic>p</italic>&#x20;&#x3d; 0.019), and agreeableness was positively correlated with the delayed receptive scores (<italic>B</italic>&#x20;&#x3d; 0.65, <italic>SE</italic> &#x3d; 0.23, <italic>Z</italic>&#x20;&#x3d; 2.77, <italic>p</italic>&#x20;&#x3d; 0.006).</p>
</sec>
<sec id="s3-6">
<title>Attitudes, Impressions, and Preferences</title>
<sec id="s3-6-1">
<title>Attitudes Toward Robots</title>
<p>With the purpose of assessing the change in attitudes after the interaction with the robot or voice tutor, the normality assumption of the data was first examined. In comparing the attitude scores between the two tutor conditions or between the pre- and post-lesson surveys, we performed a Shapiro-Wilk&#x2019;s test of normality. We then used <italic>t</italic>-tests when the two compared data are both normally distributed, and Wilcoxon Signed-Ranks Tests when the normality assumption was violated. The&#x20;difference between the tutor conditions was not significant in either before (<italic>Z</italic>&#x20;&#x3d; 0.97, <italic>p</italic>&#x20;&#x3d; 0.334) nor after the lesson [<italic>t</italic> (97) &#x3d; 1.17, <italic>p</italic>&#x20;&#x3d; 0.244]. Negative attitudes toward robots/voice assistants did not change before and after the lesson in the robot tutor condition (<italic>Z</italic>&#x20;&#x3d; 1.10, <italic>p</italic>&#x20;&#x3d; 0.267), nor the voice tutor condition [<italic>t</italic> (49) &#x3d; 1.65, <italic>p</italic>&#x20;&#x3d; 0.105]. In other words, interacting with the tutor did not improve learners&#x2019; attitudes toward the specific tutor (<xref ref-type="fig" rid="F4">Figure&#x20;4</xref>).</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Mean of the NARS ratings in the robot tutor and voice tutor conditions before and after the lesson. <italic>N</italic>&#x20;&#x3d; 100 for the pre-lesson NARS; <italic>N</italic>&#x20;&#x3d; 99 for the post-lesson NARS. The highest possible score for each test was 5. The error bars indicate the standard errors.</p>
</caption>
<graphic xlink:href="frobt-08-679893-g004.tif"/>
</fig>
</sec>
<sec id="s3-6-2">
<title>Impressions of the Robot Tutor</title>
<p>A paired sample <italic>t</italic>-test on the impression survey indicated that, in the robot condition, participants&#x2019; impressions of the robot before and after the lesson did not significantly change [<italic>t</italic> (48) &#x3d; -0.22, <italic>p</italic>&#x20;&#x3d; 0.407]. In the voice tutor condition, on the other hand, the ratings were significantly higher after than before the lesson [<italic>t</italic> (49) &#x3d; -3.78, <italic>p</italic>&#x20;&#x3c; 0.001]. In addition, independent paired <italic>t</italic>-tests demonstrated, that whereas the difference in the pre-lesson impression scores between the two tutor conditions was significant [<italic>t</italic> (98) &#x3d; 2.89, <italic>p</italic>&#x20;&#x3d; 0.005], the two did not differ significantly in the post-lesson impression scores [<italic>t</italic> (97) &#x3d; -0.06, <italic>p</italic>&#x20;&#x3d; 0.954]. These results indicate that, although the expectation was different for the two tutors, the impressions became comparable after having an actual interaction (see <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>).</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Mean ratings of the impression survey in the robot tutor and voice tutor conditions before and after the lesson. <italic>N</italic>&#x20;&#x3d; 100 for pre-lesson and <italic>N</italic>&#x20;&#x3d; 99 for post-lesson. The highest possible score for each test was 7. The error bars indicate the standard errors.</p>
</caption>
<graphic xlink:href="frobt-08-679893-g005.tif"/>
</fig>
</sec>
</sec>
<sec id="s3-7">
<title>Preference of Tutors</title>
<p>Wilcoxon Signed-Ranks tests suggest that participants in the robot tutor condition preferred a human tutor to a robot (<italic>Z</italic>&#x20;&#x3d; 5.30, <italic>p</italic>&#x20;&#x3c; 0.001) or a voice tutor (<italic>Z</italic>&#x20;&#x3d; 5.52, <italic>p</italic>&#x20;&#x3c; 0.001), but did not differ in their preference for a robot tutor and a voice tutor (<italic>Z</italic>&#x20;&#x3d; 1.07, <italic>p</italic>&#x20;&#x3d; 0.286; see <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>). In the voice tutor condition, participants also preferred a human tutor to a robot tutor (<italic>Z</italic>&#x20;&#x3d; 5.59, <italic>p</italic>&#x20;&#x3c; 0.001), and to a voice tutor (<italic>Z</italic>&#x20;&#x3d; 5.39, <italic>p</italic>&#x20;&#x3c; 0.001); and they also preferred a robot tutor to a voice tutor (<italic>Z</italic>&#x20;&#x3d; 2.15, <italic>p</italic>&#x20;&#x3d; 0.031). Participants in both tutor conditions did not significantly differ in their preference for human tutor (<italic>Z</italic>&#x20;&#x3d; -1.27, <italic>p</italic>&#x20;&#x3d; 0.206), robot tutor (<italic>Z</italic>&#x20;&#x3d; -0.85, <italic>p</italic>&#x20;&#x3d; 0.397) or voice tutor (<italic>Z</italic>&#x20;&#x3d; -0.28, <italic>p</italic>&#x20;&#x3d; 0.778).</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Preference ratings for each tutor after the lesson. <italic>N</italic>&#x20;&#x3d; 49 in the robot tutor condition; <italic>N</italic>&#x20;&#x3d; 50 in the voice tutor condition. The highest possible score for each test was 5. The error bars indicate the standard errors.</p>
</caption>
<graphic xlink:href="frobt-08-679893-g006.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>As the presence of social robots in our lives is becoming more and more prominent, it is critical to understand when and for whom robots can provide the most benefit. The present study examined the physical embodiment of robots and individual differences among learners to evaluate the effectiveness of robot tutors in an online L2 vocabulary lesson. To further understand the circumstances in which the robot tutor is effective, we also assessed how learners&#x2019; individual differences in attitudes toward robots, impressions of the robot tutor, anxiety about L2 learning, and personality traits were related to their learning outcomes. Through a stringent evaluation using two different outcome measures at two time points, we found that embodiment did not affect learning in our lesson, and individuals with negative attitudes toward robots and L2 learning anxiety learned fewer words in the robot-led lesson.</p>
<sec id="s4-1">
<title>Embodiment of the Robot Tutor</title>
<p>The learning outcomes were comparable on all four measures between the robot tutor and voice tutor conditions, and thus we did not see an advantage of the robot tutor having a body. Our results are in concert with previous research that did not find benefits of physical embodiment in learning (e.g., <xref ref-type="bibr" rid="B19">Kennedy et&#x20;al., 2015</xref>). To further confirm the conclusion, we also conducted an exploratory analysis comparing the current data with the data from the in-person robot lesson in our previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>). We built four GLMMs for each post-lesson test examining the main effects of 1) embodiment (in-person and Zoom robot tutors vs. Zoom voice tutor), and 2) physical presence (in-person robot tutor vs. Zoom robot and voice tutors). Neither embodiment nor physical presence was identified as a significant predictor (all <italic>p</italic>&#x2019;s &#x3e; 0.080). Therefore, we found no evidence of the robot&#x2019;s embodiment or physical presence affecting the learning outcomes of the simple L2 vocabulary lesson. As discussed further in <italic>Changes in Attitudes, Impressions, and Preferences From Before to After the Lesson</italic> section, we did not find an impact of physical embodiment on the learning outcomes or impressions of the robot tutor after the lesson either. The context of our paradigm must be taken into consideration in interpreting these results, as our vocabulary learning task was solely conversational and did not require the robot to interact with the physical world. Embodiment may not be a factor in such non-physical settings (<xref ref-type="bibr" rid="B24">Ligthart and Truong, 2015</xref>), hence learning environments with physical materials may yield different results.</p>
</sec>
<sec id="s4-2">
<title>Individual Differences and Learning Outcomes</title>
<p>In concert with the previous study concerning in-person lessons (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), we found that negative attitudes toward robots as well as anxiety about learning L2 are related to L2 vocabulary learning with a robot, though the relations were less pronounced. In addition to the measures used previously, we tested the effect of the learner&#x2019;s first impression of the robot tutor. The current study was among the first to test whether 1) the first impressions of the robot affect the learning outcomes, and 2) the impressions of the robot change before and after the interaction. Contrary to our expectation, the first impression ratings did not predict the number of words participants learned from the lesson. Therefore, we found that the NARS, which assessed participants&#x2019; general attitudes toward robots, was a better predictor of the learning outcomes than the impression of the specific robot&#x20;tutor.</p>
<p>The inclusion of the impression survey is also relevant for the discussion of the construct validity of the questionnaires used in HRI studies. Many studies used the NARS (<xref ref-type="bibr" rid="B27">Nomura et&#x20;al., 2006</xref>) to measure the attitudes of participants and to predict participants&#x2019; behaviors (<xref ref-type="bibr" rid="B27">Nomura et&#x20;al., 2006</xref>; <xref ref-type="bibr" rid="B34">Takayama and Pantofaru, 2009</xref>; <xref ref-type="bibr" rid="B16">Ivaldi et&#x20;al., 2017</xref>). In both the current study and the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), although the NARS predicted the number of words participants learned, the correlation was weak to moderate. One possibility was that the difference in generality between the independent variables (i.e.,&#x20;general attitudes toward all robots) and dependent variables (i.e.,&#x20;the number of words learned from a specific robot) led to the relatively weak correlations. Importantly, the impression survey in the current study was a less general measure, but we did not find a correlation between the impression and learning outcomes.</p>
</sec>
<sec id="s4-3">
<title>Changes in Attitudes, Impressions, and Preferences From Before to After the Lesson</title>
<p>On average, participants&#x2019; attitudes toward robots (and voice assistants) became more positive after they interacted with the specific tutor, but the change was not statistically significant. It should be noted that our lesson was very short and the interaction was minimal, and we may expect a greater change when the lesson is longer and more interactive. The NARS was also tested in the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>), and thus we can compare the data of the current study with the data from the in-person lesson. We found that the learner&#x2019;s negative attitudes toward robots did not significantly change before and after the in-person lesson either [<italic>t</italic> (49) &#x3d; -1.02, <italic>p</italic>&#x20;&#x3d; 0.31]. As per participants&#x2019; impressions of the tutor, the first impressions were better for the robot tutor than the voice tutor, but the impressions became comparable between the two conditions after the actual interaction. The results may indicate that, although the impression before the lesson can be affected by embodiment, the short Zoom session was enough for the learners to override the first impressions and assess the agent based on the actual interactive and communicative capabilities. With regard to the learner&#x2019;s preference, we observed a clear preference for a human tutor over both of the machine tutors, and some preference for the robot tutor over the voice tutor. These results also emphasize the importance of choosing different scales depending on what the researcher plans to evaluate.</p>
</sec>
<sec id="s4-4">
<title>Limitations and Future Directions</title>
<p>In the current study, embodiment did not facilitate vocabulary learning, and the learner&#x2019;s attitudes toward robots and anxiety about learning L2 consistently predicted learning outcomes. In terms of physical presence, however, we could only compare the current study with the previous study (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>) to anecdotally discuss its lack of impact. Therefore, a direct comparison between in-person and virtual lessons should be made before drawing a conclusion. It would also be critical to further test the unique features of robots (e.g., the ability to perform gestures) and to consider other aspects of language such as grammar and speaking (<xref ref-type="bibr" rid="B18">Kanero et&#x20;al., 2021</xref>). Similarly, the lesson scenarios, the demographic characteristics of participants (e.g., education, familiarity with robots) and the morphology of robots (e.g., Pepper, Kismet, Leonardo) might affect learning outcomes. Future research should not only investigate the influence of these factors on learning outcomes, but also analyze the detailed nature of human-robot interaction (e.g., the learner&#x2019;s behaviors during the lesson).</p>
<p>Perhaps most importantly, in the current study, the human-robot interaction was limited to one session lasting only about 15&#xa0;min. Needless to say, more research is needed to examine whether the physical body of a robot affects learning outcomes in other settings such as a lesson on another subject, or in a longer and more interactive lesson. The effects of embodiment may be more pronounced when multiple lessons are provided over a longer period of time. Further, some researchers suggest that robot tutors may reduce the L2 anxiety of child learners in the long run (<xref ref-type="bibr" rid="B3">Alemi et&#x20;al., 2015</xref>), and thus future research may focus on the long-term effects of robot language lessons on the anxiety levels of children and adults. Another recent study also found that children between 5 and 6&#xa0;years old do not interact with voice assistants as much as they interact with humans (<xref ref-type="bibr" rid="B2">Aeschlimann et&#x20;al., 2020</xref>). To our knowledge, no child study has compared robots and voice assistants. Overall, developmental research should adopt an experimental design similar to our study and examine whether the current findings can be replicated with a younger population.</p>
<p>Our data in the voice tutor condition also provide insights into the effectiveness of voice assistants such as Amazon Alexa and Apple Siri. Research with children suggests that voice assistants are perceived as a source of information that can answer questions about a wide range of subjects, including language such as definitions, spellings, and translations (<xref ref-type="bibr" rid="B41">Lovato et&#x20;al., 2019</xref>). Our results show that adults can learn a second language from voice assistants as well, at least to the same extent they do with social robots. It should also be noted that one reason why we did not find a link between negative attitudes toward voice assistants and learning outcomes might be that we adapted a questionnaire about robots, simply by changing the word &#x201c;robot&#x201d; to &#x201c;voice assistant.&#x201d; While this manipulation made the two conditions as comparable as possible, the validity of the voice assistant questionnaire should be carefully considered. Future research may use our findings as a base to explore how and for whom voice tutors are beneficial.</p>
<p>Finally, we should also point out that the current study was conducted amid the COVID-19 pandemic. We believe that our findings are generalizable, and if anything, the pandemic might have provided a better setting to evaluate the impact of (dis)embodiment. Online education has become abundant, and people may be less hesitant to engage in virtual interactions, hence the difference between in-person and online interactions should be less driven by the unfamiliarity of online interactions in the current climate. Nevertheless, more studies should be conducted to critically assess the generalizability of the findings.</p>
</sec>
<sec id="s4-5">
<title>Conclusion</title>
<p>This study was the first to empirically investigate the influence of the robot&#x2019;s physical embodiment on second language learning. The study presents an example of embodiment not affecting the learning outcomes although the results should be interpreted cautiously until the results are replicated for different language learning tasks and using various scenarios and interaction designs. Evaluating the influences of individual differences in robot-led Zoom lessons, we also found that the learner&#x2019;s general attitudes toward robots predict learning outcomes. Our findings provide some hope for the difficult situation during the COVID-19 pandemic because participants successfully learned vocabulary in a short Zoom lesson. The current results also encourage more researchers to be engaged in studying the influence of the user&#x2019;s individual differences in human-robot interaction and policymakers and educators to carefully consider how social robots and other technological devices should be incorporated in educational settings.</p>
</sec>
</sec>
</body>
<back>
<sec id="s5">
<title>Data Availability Statement</title>
<p>The datasets generated and analyzed for this study will be available from the corresponding author on reasonable request.</p>
</sec>
<sec id="s6">
<title>Ethics Statement</title>
<p>The studies involving human participants were reviewed and approved by Sabanc&#x131; University. The participants completed an online consent form to participate in this study.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>JK conceived the study in consultation with ET and CO. JK and ET were in charge of collecting the data. ET and JK analyzed the data in consultation with CO, TG, and AK. JK, ET, and CO drafted the manuscript, and all authors critically edited it. All authors contributed to the project and approved the final submitted version of the manuscript.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>The publication fee of the manuscript was paid from the institutional research funds of Sabanc&#x131; University and Ko&#xe7; University.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ack>
<p>We would like to thank the members of Sabanc&#x131; University Mind, Language, &#x0026; Technology Lab &#x2013; Serap &#xd6;zl&#xfc; for creating study materials and conducting the experiment, and &#x130;rem G&#xf6;zelekli, Ceren Boynuk, and Ay&#x15f;enaz Akbay for conducting the experiment.</p>
</ack>
<sec id="s10">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frobt.2021.679893/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frobt.2021.679893/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="Video1.MP4" id="SM1" mimetype="application/MP4" xmlns:xlink="http://www.w3.org/1999/xlink">
<label>Video 1.MP4</label>
<caption>
<p>An excerpt from the robot tutor lesson.</p>
</caption>
</supplementary-material>
<supplementary-material xlink:href="Video2.MP4" id="SM2" mimetype="application/MP4" xmlns:xlink="http://www.w3.org/1999/xlink">
<label>Video 2.MP4</label>
<caption>
<p>An excerpt from the voice tutor lesson.</p>
</caption>
</supplementary-material>
</sec>
<fn-group>
<fn id="FN1">
<label>1</label>
<p>Breakout room is a feature in Zoom that allows the host to split one Zoom session into multiple separate subsessions whereby participants in separate breakout rooms do not see each other. We put the participant into a separate breakout room away from the Experimenter and the tutor so that participants do not need to feel watched or pressured.</p>
</fn>
<fn id="FN2">
<label>2</label>
<p>The gestures used in the lesson were mostly generic except that, when the participant repeated the target word following the robot tutor, the robot made the &#x201c;pinched fingers&#x201d; gesture where all fingers were put together with the palm side up and the hand was moved up and down. This conventional gesture means &#x201c;very good&#x201d; in the Turkish culture.</p>
</fn>
<fn id="FN3">
<label>3</label>
<p>We used GLMMs in these analyses, because our data are not normally distributed, and because they allow us to analyze the responses of participants without averaging across trials (<xref ref-type="bibr" rid="B39">Jaeger, 2008</xref>). As the outcome (the scores of the four post-lesson tests) was a binary variable (correct vs. incorrect), logit (log-odds) was used as the link function. GLMMs were generated in R (<xref ref-type="bibr" rid="B43">R Development Core Team, 2016</xref>) using the <italic>lme4.glmer</italic> function (<xref ref-type="bibr" rid="B37">Bates, 2015</xref>). In all models, we included the random effect of item (e.g., L2 words) as some L2 vocabulary words may be inherently more difficult to learn than others. All models were fit by maximum likelihood using adaptive Gauss-Hermite quadrature (nAGQ &#x3d;&#x20;1).</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Admoni</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Scassellati</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Social Eye Gaze in Human-Robot Interaction: A Review</article-title>. <source>J.&#x20;Hum.-Robot Interact.</source> <volume>6</volume>, <fpage>25</fpage>. <pub-id pub-id-type="doi">10.5898/JHRI.6.1.Admoni</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aeschlimann</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bleiker</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wechner</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Gampe</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Communicative and Social Consequences of Interactions with Voice Assistants</article-title>. <source>Comput. Hum. Behav.</source> <volume>112</volume>, <fpage>106466</fpage>. <pub-id pub-id-type="doi">10.1016/j.chb.2020.106466</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Alemi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Meghdari</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ghazisaedy</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>The Impact of Social Robotics on L2 Learners&#x2019; Anxiety and Attitude in English Vocabulary Acquisition</article-title>. <source>Int. J.&#x20;Soc. Robotics</source> <volume>7</volume>, <fpage>523</fpage>&#x2013;<lpage>535</lpage>. <pub-id pub-id-type="doi">10.1007/s12369-015-0286-y</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ayd&#x131;n</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Harputlu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>G&#xfc;zel</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>&#x00C7;elik</surname>
<given-names>S. S.</given-names>
</name>
<name>
<surname>U&#x015F;tuk</surname>
<given-names>&#x00D6;.</given-names>
</name>
<name>
<surname>Gen&#xe7;</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A Turkish Version of Foreign Language Anxiety Scale: Reliability and Validity</article-title>. <source>Proced. - Soc. Behav. Sci.</source> <volume>232</volume>, <fpage>250</fpage>&#x2013;<lpage>256</lpage>. <pub-id pub-id-type="doi">10.1016/j.sbspro.2016.10.011</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Barsalou</surname>
<given-names>L. W.</given-names>
</name>
<name>
<surname>Niedenthal</surname>
<given-names>P. M.</given-names>
</name>
<name>
<surname>Barbey</surname>
<given-names>A. K.</given-names>
</name>
<name>
<surname>Ruppert</surname>
<given-names>J.&#x20;A.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>The Psychology of Learning and Motivation: Advances in Research and Theory</article-title>, <source>Elsevier Science</source>, <fpage>43</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1016/s0079-7421(03)01011-9</pub-id> <comment>Social Embodiment</comment>. </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Belpaeme</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kennedy</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ramachandran</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Scassellati</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Tanaka</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Social Robots for Education: A Review</article-title>. <source>Science Rob.</source>, <fpage>3</fpage>. <pub-id pub-id-type="doi">10.1126/scirobotics.aat5954</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Bartneck</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Forlizzi</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2004</year>). &#x201c;<article-title>A Design-Centred Framework for Social Human-Robot Interaction</article-title>,&#x201d; in <source>RO-MAN 2004: 13th IEEE International Workshop on Robot and Human Interactive Communication</source>. (<publisher-loc>Kurashiki, Japan</publisher-loc>: <publisher-name>Institute of Electrical and Electronics Engineers</publisher-name>), <fpage>591</fpage>&#x2013;<lpage>594</lpage>. <pub-id pub-id-type="doi">10.1109/ROMAN.2004.1374827</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Bates</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>M&#x00E4;chler</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bolker</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Walker</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Fitting Linear Mixed-Effects Models Using lme4</article-title>. <source>J. Statistical Software</source> <volume>67</volume>, <fpage>1</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v067.i01</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bialystok</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Craik</surname>
<given-names>F. I. M.</given-names>
</name>
<name>
<surname>Klein</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Viswanathan</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Bilingualism, Aging, and Cognitive Control: Evidence from the Simon Task</article-title>. <source>Psychol. Aging</source> <volume>19</volume>, <fpage>290</fpage>&#x2013;<lpage>303</lpage>. <pub-id pub-id-type="doi">10.1037/0882-7974.19.2.290</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chevalier</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Martin</surname>
<given-names>J.-C.</given-names>
</name>
<name>
<surname>Isableu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bazile</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Tapus</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Impact of Sensory Preferences of Individuals with Autism on the Recognition of Emotions Expressed by Two Robots, an Avatar, and a Human</article-title>. <source>Auton. Robot</source> <volume>41</volume>, <fpage>613</fpage>&#x2013;<lpage>635</lpage>. <pub-id pub-id-type="doi">10.1007/s10514-016-9575-z</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Demir</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Kumkale</surname>
<given-names>G. T.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Individual Differences in Willingness to Become an Organ Donor: A Decision Tree Approach to Reasoned Action</article-title>. <source>Personal. Individual Differences</source> <volume>55</volume>, <fpage>63</fpage>&#x2013;<lpage>69</lpage>. <pub-id pub-id-type="doi">10.1016/j.paid.2013.02.002</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Mutlu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Mataric</surname>
<given-names>M. J.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Embodiment in Socially Interactive Robots</article-title>. <source>FNT in Robotics</source> <volume>7</volume> (<issue>4</issue>), <fpage>251</fpage>&#x2013;<lpage>356</lpage>. <pub-id pub-id-type="doi">10.1561/2300000056</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dunn</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Dunn</surname>
<given-names>D. M.</given-names>
</name>
</person-group> (<year>2007</year>). <source>Peabody Picture Vocabulary Test(PPVT)</source>. <edition>Fourth edition</edition> (<publisher-loc>Minneapolis, MN</publisher-loc>: <publisher-name>NCS Pearson</publisher-name>).</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gunaydin</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Selcuk</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Zayas</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Impressions Based on a Portrait Predict, 1-Month Later, Impressions Following a Live Interaction</article-title>. <source>Soc. Psychol. Personal. Sci.</source> <volume>8</volume>, <fpage>36</fpage>&#x2013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1177/1948550616662123</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jo</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Jo</surname>
<given-names>J. H.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>Comparative Study on the Educational Use of home Robots for Children</article-title>. <source>J.&#x20;Inf. Process. Syst.</source> <volume>4</volume>, <fpage>159</fpage>&#x2013;<lpage>168</lpage>. <pub-id pub-id-type="doi">10.3745/JIPS.2008.4.4.159</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hoff</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Interpreting the Early Language Trajectories of Children from Low-SES and Language Minority Homes: Implications for Closing Achievement Gaps</article-title>. <source>Dev. Psychol.</source> <volume>49</volume>, <fpage>4</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1037/a0027238</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Horwitz</surname>
<given-names>E. K.</given-names>
</name>
<name>
<surname>Horwitz</surname>
<given-names>M. B.</given-names>
</name>
<name>
<surname>Cope</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>1986</year>). <article-title>Foreign Language Classroom Anxiety</article-title>. <source>Mod. Lang. J.</source> <volume>70</volume>, <fpage>125</fpage>&#x2013;<lpage>132</lpage>. <pub-id pub-id-type="doi">10.2307/32731710.1111/j.1540-4781.1986.tb05256.x</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ivaldi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lefort</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Peters</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chetouani</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Provasi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zibetti</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Towards Engagement Models that Consider Individual Factors in HRI: On the Relation of Extroversion and Negative Attitude towards Robots to Gaze and Speech during a Human-Robot Assembly Task</article-title>. <source>Int. J.&#x20;Soc. Robotics</source> <volume>9</volume>, <fpage>63</fpage>&#x2013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1007/s12369-016-0357-8</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jaeger</surname>
<given-names>T. F.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>Categorical Data Analysis: Away from ANOVAs (transformation or not) and Towards Logit Mixed Models</article-title>. <source>J. Mem. Lang.</source> <volume>59</volume>, <fpage>434</fpage>&#x2013;<lpage>446</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2007.11.007</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kanero</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ge&#xe7;kin</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Oran&#xe7;</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Mamus</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>K&#xfc;ntay</surname>
<given-names>A. C.</given-names>
</name>
<name>
<surname>G&#xf6;ksun</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Social Robots for Early Language Learning: Current Evidence and Future Directions</article-title>. <source>Child. Dev. Perspect.</source> <volume>12</volume>, <fpage>146</fpage>&#x2013;<lpage>151</lpage>. <pub-id pub-id-type="doi">10.1111/cdep.12277</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kanero</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Oran&#xe7;</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ko&#x15f;kulu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kumkale</surname>
<given-names>G. T.</given-names>
</name>
<name>
<surname>G&#xf6;ksun</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>K&#xfc;ntay</surname>
<given-names>A. C.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Are Tutor Robots for Everyone? the Influence of Attitudes, Anxiety, and Personality on Robot-Led Language Learning</article-title>. <source>Int. J.&#x20;Soc. Robotics</source>. <pub-id pub-id-type="doi">10.1007/s12369-021-00789-3</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kennedy</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Baxter</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Belpaeme</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Comparing Robot Embodiments in a Guided Discovery Learning Interaction with Children</article-title>. <source>Int. J.&#x20;Soc. Robotics</source> <volume>7</volume>, <fpage>293</fpage>&#x2013;<lpage>308</lpage>. <pub-id pub-id-type="doi">10.1007/s12369-014-0277-4</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Konishi</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kanero</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Freeman</surname>
<given-names>M. R.</given-names>
</name>
<name>
<surname>Golinkoff</surname>
<given-names>R. M.</given-names>
</name>
<name>
<surname>Hirsh-Pasek</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Six Principles of Language Development: Implications for Second Language Learners</article-title>. <source>Dev. Neuropsychol.</source> <volume>39</volume>, <fpage>404</fpage>&#x2013;<lpage>420</lpage>. <pub-id pub-id-type="doi">10.1080/87565641.2014.931961</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kov&#xe1;cs</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Mehler</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Cognitive Gains in 7-Month-Old Bilingual Infants</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>106</volume>, <fpage>6556</fpage>&#x2013;<lpage>6560</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.0811323106</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>K&#x00F6;se</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Uluer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Akalin</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Yorganci</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>&#x00D6;zkul</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ince</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>The Effect of Embodiment in Sign Language Tutoring with Assistive Humanoid Robots</article-title>. <source>Int J. Soc. Rob.</source> <volume>7</volume>, <fpage>537</fpage>&#x2013;<lpage>548</lpage>. <pub-id pub-id-type="doi">10.1007/s12369-015-0311-1</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lovato</surname>
<given-names>S. B.</given-names>
</name>
<name>
<surname>Piper</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Wartella</surname>
<given-names>E. A.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Hey Google, Do Unicorns Exist?: Conversational Agents as a Path to Answers to Children&#x2019;s Questions</article-title>,&#x201d; in <source>Proceedings of the 18th ACM International Conference on Interaction Design and Children</source> (<publisher-loc>Boise, ID USA</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>301</fpage>&#x2013;<lpage>313</lpage>. <pub-id pub-id-type="doi">10.1145/3311927.3323150</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Leyzberg</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Spaulding</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Toneva</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Scassellati</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>The Physical Presence of a Robot Tutor Increases Cognitive Learning Gains</article-title>. <source>Proc. Annu. Meet. Cogn. Sci. Soc.</source> <volume>6</volume>. </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>The Benefit of Being Physically Present: A Survey of Experimental Works Comparing Copresent Robots, Telepresent Robots and Virtual Agents</article-title>. <source>Int. J.&#x20;Human-Computer Stud.</source> <volume>77</volume>, <fpage>23</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijhcs.2015.01.001</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ligthart</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Truong</surname>
<given-names>K. P.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Selecting the Right Robot: Influence of User Attitude, Robot Sociability and Embodiment on User Preferences</article-title>,&#x201d; in <source>24th IEEE International Symposium on Robot and Human Interactive Communication</source>. (<publisher-loc>Kobe, Japan</publisher-loc>: <publisher-name>ROMAN: IEEE</publisher-name>), <fpage>682</fpage>&#x2013;<lpage>687</lpage>. <pub-id pub-id-type="doi">10.1109/ROMAN.2015.7333598</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lytle</surname>
<given-names>S. R.</given-names>
</name>
<name>
<surname>Kuhl</surname>
<given-names>P. K.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Social Interaction and Language Acquisition: Toward a Neurobiological View</article-title>,&#x201d; in <source>The Handbook of Psycholinguistics</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Fern&#xe1;ndez</surname>
<given-names>E. M.</given-names>
</name>
<name>
<surname>Cairns</surname>
<given-names>H. S.</given-names>
</name>
</person-group> (<publisher-loc>Hoboken, NJ</publisher-loc>: <publisher-name>Wiley Blackwell</publisher-name>), <fpage>615</fpage>&#x2013;<lpage>634</lpage>. <pub-id pub-id-type="doi">10.1002/9781118829516.ch27</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Macedonia</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Groher</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Roithmayr</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Intelligent Virtual Agents as Language Trainers Facilitate Multilingualism</article-title>. <source>Front. Psychol.</source> <volume>5</volume>. <pub-id pub-id-type="doi">10.3389/fpsyg.2014.00295</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Milligan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Astington</surname>
<given-names>J. W.</given-names>
</name>
<name>
<surname>Dack</surname>
<given-names>L. A.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Language and Theory of Mind: Meta-Analysis of the Relation Between Language Ability and False-belief Understanding</article-title>. <source>Child Develop.</source> <volume>78</volume>, <fpage>622</fpage>&#x2013;<lpage>646</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-8624.2007.01018.x</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nomura</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kanda</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Suzuki</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Experimental Investigation into Influence of Negative Attitudes toward Robots on Human-Robot Interaction</article-title>. <source>AI Soc.</source> <volume>20</volume>, <fpage>138</fpage>&#x2013;<lpage>150</lpage>. <pub-id pub-id-type="doi">10.1007/s00146-005-0012-7</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Paolo</surname>
<given-names>A. D.</given-names>
</name>
<name>
<surname>Tansel</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Returns to Foreign Language Skills in a Developing Country: The Case of Turkey</article-title>. <source>J.&#x20;Dev. Stud.</source> <volume>51</volume>, <fpage>407</fpage>&#x2013;<lpage>421</lpage>. <pub-id pub-id-type="doi">10.1080/00220388.2015.1019482</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Pfeifer</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Scheier</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>1999</year>). <source>Understanding Intelligence</source>. <publisher-name>MIT Press</publisher-name>.</citation>
</ref>
<ref id="B43">
<citation citation-type="book">
<collab>R Development Core Team</collab> (<year>2020</year>). <source>R: A Language and Environment for Statistical Computing</source>. Available at: <ext-link ext-link-type="uri" xlink:href="http://www.r-project.org/">http://www.r-project.org/</ext-link>.</citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Rosenthal-von der P&#xfc;tten</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Stra&#xdf;mann</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Kr&#xe4;mer</surname>
<given-names>N. C.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Robots or Agents - Neither Helps You More or Less during Second Language Acquisition</article-title>,&#x201d; in <source>Intelligent Virtual Agents Lecture Notes in Computer Science.</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Traum</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Swartout</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Khooshabeh</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Kopp</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Scherer</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Leuski</surname>
<given-names>A.</given-names>
</name>
</person-group> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>256</fpage>&#x2013;<lpage>268</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-47665-0_23</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Takayama</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Pantofaru</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Influences on Proxemic Behaviors in Human-Robot Interaction</article-title>. <source>IEEE/RSJ&#x20;Int. Conf. Intell. Robots Syst.</source> <fpage>5495</fpage>&#x2013;<lpage>5502</lpage>. <pub-id pub-id-type="doi">10.1109/IROS.2009.5354145</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="book">
<collab>University of Cambridge Local Examinations Syndicate</collab> (<year>2001</year>). <source>Quick Placement Test</source>. <publisher-loc>Oxford, UK</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Verga</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kotz</surname>
<given-names>S. A.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>How Relevant Is Social Interaction in Second Language Learning?</article-title>. <source>Front. Hum. Neurosci.</source> <volume>7</volume>. <pub-id pub-id-type="doi">10.3389/fnhum.2013.00550</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Willis</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Todorov</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>First Impressions</article-title>. <source>Psychol. Sci.</source> <volume>17</volume>, <fpage>592</fpage>&#x2013;<lpage>598</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-9280.2006.01750.x</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>
