Original Research ARTICLE
Biologically Inspired Emotional Expressions for Artificial Agents
- 1Department of Mechatronics, Optics and Engineering Informatics, Budapest University of Technology and Economics, Budapest, Hungary
- 2Department of Ethology, Eötvös Loránd University, Budapest, Hungary
- 3Institute for Computer Science and Control, Hungarian Academy of Sciences, Budapest, Hungary
- 4Department of Precision Mechanics, Chuo University, Tokyo, Japan
- 5MTA-ELTE Comparative Ethology Research Group, Budapest, Hungary
- 6Department of Telecommunications and Media Informatics, Budapest University of Technology and Economics, Budapest, Hungary
A special area of human-machine interaction, the expression of emotions gains importance with the continuous development of artificial agents such as social robots or interactive mobile applications. We developed a prototype version of an abstract emotion visualization agent to express five basic emotions and a neutral state. In contrast to well-known symbolic characters (e.g., smileys) these displays follow general biological and ethological rules. We conducted a multiple questionnaire study on the assessment of the displays with Hungarian and Japanese subjects. In most cases participants were successful in recognizing the displayed emotions. Fear and sadness were most easily confused with each other while both the Hungarian and Japanese participants recognized the anger display most correctly. We suggest that the implemented biological approach can be a viable complement to the emotion expressions of some artificial agents, for example mobile devices.
New technologies such as Information and Communication Technologies (ICT) have been on the rise in the past decades alongside with the new emerging field of social robotics.
Artificial agents used in ICT devices could be viewed by users as somewhat living beings, similarly to how children regard social robots differently than strictly inanimate objects (Kahn et al., 2006). A biological approach in the design of these agents' embodiment and behavior could enhance the interaction between humans and the artificial agents.
Emotion expression is an important aspect, which could be used as a means of communication for alerts and notifications (e.g., low battery) but it could also be used to make the device or robot more acceptable to humans and easier to interact with (Hudlicka, 2003). In the field of social robotics emotion expression has notable research interests (Breazeal, 2003). Most studies focus on mimicking human facial expressions (e.g., androids, Becker-Asano and Ishiguro, 2011; virtual faces, Gockley et al., 2004) which can prove difficult at the current level of technology, and could incite the Uncanny-Valley effect (Mori, 1970). Some studies use emoticons displayed on the agent's screen. Emoticons are static pictures or looped short animations used as an addition to textual communication between people (Walther and D'Addario, 2001). There is a great variety of emoticons that are based on the major features of human facial expressions associated with certain emotions (Dresner and Herring, 2010), most frequently the six basic emotions (fear, happiness, sadness, anger, disgust, surprise) described by Ekman and Oster (1979). Although above chance recognition of the basic facial expressions of these emotions has been shown across cultures (Ekman et al., 1987) the success of recognition is far from perfect (Russell, 1994). Research has shown that some aspects of emotion recognition are learned, as the cultural differences in recognition decrease with learning via cultural exposure (Elfenbein and Ambady, 2003) and also children up to 10–12 years of age are less successful in recognition of emotions from facial gestures than adults (Durand et al., 2007). The cross-cultural differences of emotion recognition and expression are of special interest in case of East Asian and Western cultures, due to the different display rules of emotion expressions in collectivistic and individualistic cultures (respectively) (Matsumoto et al., 2008; Safdar et al., 2009). Emoticons derived from facial expressions also have cultural dependence (Park et al., 2013), and they require learning as their recognition becomes better with regular exposure (McDougald et al., 2011). Emoticons are highly human-specific and by using facial expressions as their basis, they are not easily variable or individualized. The use of broader attributes of emotion expression could help to facilitate personalization, which is favorable among users (Blom and Monk, 2003).
If we consider less narrow, that is, less human face specific expressions of inner states, we must address the general rules of emotion expression behaviors in animals, which can be observed across multiple taxa (Darwin, 1872; Plutchik, 2001). Defining these rules can give us a set of guidelines that can be used to implement emotion expression systems that are dynamic, can be used across multiple modalities and do not rely on human facial expressions, thus their recognition might be consistent across cultures. In social robotics certain behaviors of social animals are considered as good models for designing functionally analogous behaviors for robots (Miklósi and Gácsi, 2012). Gácsi et al. (2016) used the emotion expression of dogs (Canis familiaris) as a basis for the emotionally expressive behavior of a mechanoid-looking robot. Their results show that participants tended to interpret the robot's actions as expressions of emotions and could identify the emotions significantly above chance level.
We propose a similar approach for creating generalized simple expressive behaviors that are based on common biological rules of emotion expressions of multiple animal species.
Animals express numerous behavioral elements in certain situations that are assumed to reflect their inner (emotional) states. Ethological research has revealed some common behavior patterns that are recognizable across wide range of taxa. Next, we summarize some of these observable behaviors and the derived rules that we use for creating our generalized model for displaying emotional states (Table 1). In this study we focused only on the visual displays of animals, but the inclusion of other modalities (e.g., acoustic signals) into the emotion expression displays is also possible.
We chose the basic emotions that have an ecological function and can be relevant in social robots or mobile agents: anger, happiness, fear, surprise and sadness. These emotions can be interpreted as coordinates in a dimensional model, e.g., the circumplex model of affect by Russell (1980) in which emotions are positioned based on their valence (positive or negative) and intensity (high or low). In the following summary the described emotions are listed based on their intensity (Russell, 1980).
It is assumed that the reaction to sudden changes in the environment is caused by a mismatch between the observed and the expected environment or event, and is accompanied by a change in the emotional state referred to as surprise (Barto et al., 2013). It is a relatively short emotional display and can involve momentary freezing (staying motionless) (Plutchik, 1980), orientation at the source of stimulus (Sokolov, 1963) or a combination of them. Novel stimuli elicit alertness (Scherer, 2001), which can be accompanied by the widening of eyelid apertures (Descovich et al., 2017), by increased visibility of the white sclera (Kim et al., 2004), and by raised eyebrows (in humans: Lucey et al., 2010). Surprise in itself cannot be categorized as having a positive or negative valence as it is based on the circumstances (e.g., if the stimulus is a suddenly appearing conspecific or a predator) therefore it is difficult to generalize a rule for the avoidance or approach behavior. The startle reaction usually contains stereotypic behaviors, for example in certain tadpoles the sudden stimulus causes them to quickly bend their tails and move forward in a different direction than before (Yamashita et al., 2000). Thus, for the creation of the surprise emotion the virtual agent displayed freezing, change of orientation, and an alert state with a slight size increase and brightened color.
Behavioral reactions associated with fear can be observed in animals e.g., during predator-avoidance and retreating, frequently characterized by the flight initiation distance (Stankowich and Blumstein, 2005). The animals might have the potential for active avoidance which results in fast locomotion and fleeing from the approaching predator (Bonenfant and Kramer, 1996). The goal of the fleeing is usually to gain distance from the predator and to reach refuge (Dill and Houtman, 1989) where they can hide (Rhoades and Blumstein, 2007) which is often helped by a crouching posture (Caro, 2005). As fear is often the result of a stressful stimulus, pale colors can also occur (Conte, 2004; pallor in humans: Garfinkel and Critchley, 2014) due to the redirection of blood from the skin and periphery to the muscles (Vianna and Carrive, 2005). In the fear display the agent flees quickly to the corner of the screen, its size decreases and it displays paler colors.
The majority of threat displays follow general rules, as many evolved from movements and behaviors that occurred before or during fighting (Andersson, 1980). Agonistic behaviors contain threat displays, the display of weaponry, size exaggeration (Nishida et al., 1999; de Boer et al., 2003), vivid colors e.g., via blood rush (Drummond and Quah, 2001), or showing vividly pigmented areas during aggressive displays (Evans and Norris, 1996; Chen and Fernald, 2011), and typical approach patterns (Scott and Fredericson, 1951). The approach patterns can be part of ritualistic behaviors intended to deal with the conflict without an actual fight, or could be the general dynamics of an aggressive approach (starting with orientation and a slow approach, followed by a fast movement to reach the target) (Dill, 1978). Thus, for the creation of the anger display the virtual agent displayed vivid color, accelerating approach, increase in size, and angular forms (display of “weaponry,” similarly to teeth or talons).
This emotion was based on the play and greeting behavior of animals which are both connected to positive emotional states (Panksepp, 2004; Trezza et al., 2010; Konok et al., 2015). During play animals exhibit a wide range of behavior elements that can originate from other contexts, e.g., from fighting or sexual behaviors (Palagi et al., 2016) which we did not want to include, but less specific behaviors, e.g., contact seeking, locomotor-rotational and jumping movements are found in both animals and humans (Pellis and Pellis, 1983; children: Pellegrini and Smith, 1998; Cordoni, 2009; Held and Špinka, 2011) while increased blood flow due to physical activity can result in facial flushing in humans (Izikson et al., 2006). The greeting behavior of dogs is interpreted as happy or joyful by humans (Konok et al., 2015), however such behavior can be also observed in socialized wolves (Topál et al., 2005), and macaques (De Marco et al., 2014). Many aspects of this behavior can be considered an opposition to agonistic behaviors, although fast movements are characteristic in both instances. Thus, for the creation of the happy display the virtual agent displayed fast and energetic movements, jumping, approach to contact and less angular forms (hiding “weaponry”).
In animals behavioral changes caused by stress which the animal cannot cope with may evoke anhedonia, and a passive, withdrawn behavioral state (Herrera-Pérez et al., 2008; Nestler and Hyman, 2010). In some species there are characteristic changes in coloration due to elevated stress, as the color of the animal fades e.g., in fish (Conte, 2004) and in case of chronic stress, reptiles (Fitze et al., 2009). Depression and chronic sadness is usually associated with slow movements and contraction of the body/hunched posture (Panksepp et al., 1991; Michalak et al., 2009; humans: Buyukdura et al., 2011; Konok et al., 2015), resulting in smaller perceived body size. Thus, for the creation of the sadness emotion expression the virtual agent displayed decreased size, fading coloration and slow movements.
We tested this new approach by developing emotion displays of basic emotions (happiness, anger, fear, surprise, sadness) and a neutral state for an artificial agent by dynamically changing its visual attributes. The attributed changes follow general biological rules that are based on animal behavior and communication of inner states. A set of published studies that use abstract forms for complex emotion expression in ICT devices describe the eMOTO (Fagerberg et al., 2003, 2004; Sundström et al., 2005), which is designed to enrich text messages with affective content. The system expresses emotions by using dynamic color and shape background themes of the text message application via changing its color, texture and the movement of forms constituting the texture. The creation of the expressions seems to draw from an artistic approach, as the reasoning behind the expression design is not discussed in detail and not well-referenced. Color changes are also used in other studies to express emotions in social robots e.g., with specific light patterns evoking rain or fire (Feldmaier et al., 2016), dynamic changes of luminosity and hue (Terada et al., 2012), or in a multimodal system together with vibration and sound signals (Song and Yamada, 2017), but the designated features of the emotion expressions are not based on biological rules or are set by the participants as part of the experiment.
In our study we used a complex, abstract form as the structure of our agent as we wanted to emphasize the being-like complexity of the agent while avoiding similarities with any existing organisms or symbols. Our aim was to create an agent that can display emotions in short, few seconds long animations. As this agent provides only one possible model for the proposed approach, we did not offer several variations of it or search for the very best displays, but only wanted to demonstrate that a more diverse set of emotion displays can be developed using dynamic changes in the agent's “body” movements, appearance and “approach-avoidance” behaviors. We used the visual modality as a basis, but other modalities (e.g., auditory, tactile) could be integrated in the future.
The artificial agent (originally named EDA: Emotion Display Agent) was developed by György Persa and Péter Baranyi using Ogre3D graphics rendering engine. The display animations were recorded with Fraps®. The user interface of the agent enabled the researchers to change multiple parameters of its embodiment, movement pattern and dynamic changes (Baranyi et al., 2011; Persa et al., 2012). Table 2 contains the description of changeable parameters.
The agent is comprised of a sphere and a labyrinth-like, angular part, “grid” (Figure 1). This structure was chosen to accommodate some visual properties that could make the agent look more similar to a living being. The combined movement and relative size change of the sphere and the grid enables the agent to be seen as rounder or more angular, depending on the displayed emotion.
Figure 1. The figure shows the dynamic changes of the agent during the emotion displays. The starting state is the same for all displays (happiness, fear, anger, surprise, sadness, neutral).
The agent is shown on a screen and can visualize basic emotions in 2D, but due to its form and movements, it is generally perceived as a 3D object/being. The emotions expressed by the agent were designed to be displayed in a given time limit between 4 and 15 s. The differences in the display durations reflect some of the potential variability found in the temporal dynamics of emotions and emotionally expressive behaviors (facial expressions: Ekman, 1984; Frijda et al., 1991; Verduyn et al., 2015), and are also the result of specific behaviors (e.g., hiding increases the length of the fear display).
1. We expected that participants would recognize the generalized simple expressive behaviors giving the highest score to the respective emotions in multiple choice questions.
2. We also expected lower recognition rates for simple emotional states that share more behavioral attributes.
3. We also hypothesized that our displays present universal biological features, thus there would be no major differences in the assessment of the Hungarian and Japanese participants.
Both the Hungarian and Japanese subjects were unpaid volunteers recruited from university students. The Hungarian sample consisted of 114 participants, 78 women and 34 men (age 21 ± SD 1.7 years). The Japanese sample consisted of 22 participants, 21 men and one woman (age 22 ± SD 1.1 years). The Hungarian sample, due to the higher number of participants, was tested in six groups. For the cultural comparison a balanced Hungarian sample was used.
The study was carried out in accordance with the recommendations of the Institutional Review Board of Institute of Biology, Eötvös Loránd University, Budapest, Hungary. The protocol was reviewed and approved by the Institutional Review Board of Institute of Biology, Eötvös Loránd University, Budapest, Hungary. All subjects gave informed oral consent in accordance with the Declaration of Helsinki. The study was conducted in 2011, when written consent was not required for questionnaire studies. Participants took part in the study voluntarily and anonymously.
The testing took place at the Department of Ethology at the Eötvös Loránd University in Budapest, Hungary and in the Human-System Laboratory at Chuo University in Tokyo, Japan.
The emotion displays were defined for each emotion based on the above ethological summary and were recorded as short videos.
We used human fine tuning as a second step to finalize the design of the emotion expressions: we conducted a single questionnaire preliminary study with this first set of animations to get a pre-evaluation of the displayed emotions. The participants of the preliminary study were Hungarian volunteers of the staff and students of the Department of Ethology at the Eötvös Loránd University (N = 13, 8 females and 5 males, age 34.4 ± SD 10.2 years). The experimenter used a Microsoft PowerPoint® presentation to play the 5 videos in a fixed order (anger, fear, happiness, sadness, and surprise), the experimenter changed the slides manually after each video.
Before viewing the videos, the experimenter told the subjects to consider the agent as an unknown being and that the subjects should fill out the closed-ended questionnaire according to what emotions they think the agent expresses, which emotion is more adequate for the displayed animation.
The instructions for the questionnaire were: “Mark for each emotion how much it is relevant for the agent (‘living being’) on the given video (to what extent it describes its actual inner state).” After the subjects watched the videos, the experimenter handed out the questionnaire. The subjects then viewed the animations again one by one and the subjects were instructed to fill out the relevant parts of the questionnaire after each video. The subjects had to mark on a Likert scale (from 0 to 5) how characteristic each emotion (angry, happy, afraid, surprised, sad) was to each video. The questionnaire data of the preliminary study were analyzed with simple sum of scores for each emotion from all the participants.
The results of the preliminary study showed that the summed scores for most emotions were the highest in case of the relevant videos, e.g., when the subjects viewed the video in which the agent displayed sadness, the subjects gave the highest score to the sad emotion. The only exception was the happiness display, in which the participants gave slightly more scores to the anger emotion (happy 36, angry 41 scores).
Based on the results of the preliminary study and on the interviews conducted with the participants, some changes were made in the displays, and we also added a neutral state display in order to have a sort of “baseline.” The finalized attributes of each display are summed up in Table 1, while the dynamic changes of the agent during the emotion displays can be seen in Figure 1. The videos are available in the supplementary material. All displays started the same way, showing the agent in its neutral state (greenish sphere color, visible medium sized grid). Based on the results of the preliminary study, the questionnaire was divided into two parts, in Q1 we used open-ended questions in which the participants could freely express what they thought the displays conveyed. The instruction on Questionnaire 1 was: “Let's suppose that the agent on the video is an unfamiliar living being. What do you think its current thought/mood is, how do you think it is feeling right now? (not sentences, just words!)”. In Questionnaire 2 we used the same closed-ended questions as in the preliminary study. We analyzed the two questionnaires separately.
Due to the two questionnaires the process of the presentation was modified. There were three presentations. In the first presentation the videos were all played in a single succession (watching video1-video6 in random order that differed for the 6 Hungarian groups of participants). Next, the participants received Q1 with the open-ended questions and the videos were presented to them again but this time one by one, in a random order that differed from the order of the first viewing. The participants were instructed to fill out the relevant part of Q1 after each video. Then Q1 forms were collected, and Q2 questionnaires with the closed-ended questions were given to the participants. The subjects then were presented with the videos the third time, in another random order and were instructed to fill out Q2. The random orders for the videos were defined a priori, and the participants were identified through a codename. The test procedure was the same in both countries, but in case of the Japanese group although the open-ended questionnaire was filled out by participants as in the Hungarian group, it was later discarded from analysis due to difficulties in translating the responses into terms that properly correspond to the Hungarian ones. In case of the Japanese participants, the random order of viewing was the same as one of the Hungarian group's.
Open Questions (Q1)
The answers for each video (separately for each viewing) were collected and categorized. If the participant used multiple definitions/adjectives, only the first three were considered. We used a scoring system in a modified form, adapted from another study on emotion expression in robots (Gácsi et al., 2016) to categorize the given answers. Answers with scores 1 or 2 were words that referred to not being-like objects (that could still show behaviors and perform actions, e.g., machines). Answers with scores 3 or 4 related to concepts that could be associated with the inner states or emotions of living beings.
Score 1: formal description of the observed behavior (e.g., agent changes color, spins faster).
Score 2: indicating some contextual behavior that attributes a meaning to the video which cannot be directly observed from the behavior of the agent (e.g., learns something, runs).
Score 3: mentioning a term or phrase, which implicitly indicates some inner state but without naming that concrete emotion (e.g., threatens somebody, tries to escape, reclusive).
Score 4: naming an inner state or emotion (or the lack of emotion) explicitly either as a verb, a noun or an adjective (e.g., happy, it is afraid of something, neutral, tired). We used descriptive analysis to show the ratio of types of answers based on the scoring system. The answers that received scores 3 or 4 were analyzed further. In the next step the answers that received scores 3 or 4 were selected and each expression was counted. We created answer groups based on the expressions that occurred more than five times per video, as expressions with similar meanings were grouped together, resulting in 5–8 categories per display.
Closed Questions (Q2)
In the Q2 questionnaire the participants had to mark on a Likert scale from 0 to 5 for each emotion their relevancy to the emotion displayed in the given video, therefore the participants did not only choose one emotion per video.
We measured success of recognition by giving 1 point if participants gave the biggest score to the correct emotion and 0 if they gave it to another (in case the participant gave the same high scores to other emotions as well as to the correct one, we divided the 1 point by the number of emotions that got the same score). We also measured if the participants gave the maximum 5 score to the correct emotion. Next, we summed the scores of the emotions separately in case of each displays (e.g., summed the scores given by all participants to the happy emotion in the neutral display) to show how all the emotions were scored for each display. Finally, we created confusion matrices to show which emotions were mixed up by the participants, by counting the percentage of answers that gave the biggest score in case of each emotion, similarly to measuring success (the score given for an emotion received 1 point if participants gave the biggest score to that emotion; in case the participant gave the same points to other emotions as well, we divided the 1 point by the number of emotions that got the same score).
Statistical analysis was conducted with IBM® SPSS® Statistics 20. As a first step, we compared the success of recognition in each display to chance level (binomial test with 0.2 as chance level). To obtain dichotomous data we used two derived variables for measuring success to chance level, by creating both a permissive (largest value given for the correct emotion, the same score could also be given to other emotions) and a strict (only the correct emotion received the largest score) success variable, e.g., if the Sadness display received a score 4 for both the sad and angry emotions, it was considered successful by the permissive, and unsuccessful by the strict success variable.
Next, we used Friedman tests and Wilcoxon signed rank tests, with Benjamini-Hochberg corrections to examine whether the scores given to the correct emotions differed significantly in the displays from the scores given to the other emotions. We conducted a hierarchical cluster analysis to examine the closeness of scored emotions (cluster method: Ward's method; variables: scored emotions; counts: chi-squared measure). Finally, we used Generalized Linear Mixed Models (GLMM) with the fixed effects of the participants' gender, the category of the scored emotion (happy, angry, afraid etc. emotions), the type of display (Happiness, Anger etc. videos), and their two-way interactions. The targets were the scores given for all the emotions across all displays and the random effects were the subjects. We used multinomial probability distribution and cumulative logit link function.
Open Questions (Q1)
As the first step of the descriptive analysis we summarized the answers based on the scoring system. In all displays the proportion of scores 3 and 4 were high. The answers that received scores 4 or 3 amounted to at least 77% of all answers (in case of the Surprise display) and at most 91% (in Happiness), meaning that the vast majority of respondents used words that express or imply emotions or inner states that are characteristic of living beings. The Happiness, Anger and Neutral displays prompted answers that were direct expressions of emotions (score 4) in 64–72% of the cases (Figure 2). Answers receiving scores 3 and 4 were further analyzed. Inter-observer reliability for separating answers with scores 1–2 from the ones with scores 3–4 was assessed by two observers on a balanced sample using Cohen's Kappa (k = 0.7; p < 0.001). Cohen's Kappa is a frequently used method for measuring inter-observer reliability, where k = 0.7 is considered a substantial agreement (Viera and Garrett, 2005).
Figure 2. The figure shows the percentage of answers given to the open-ended questionnaire in terms of the types of answers. The red lines indicate the separation of scores 1–2, indicated by dark gray colors (words referring to not being-like agents) and scores 3–4, indicated with light gray (words referring to being-like agents). Answers receiving scores 3 and 4 were analyzed further. Score 4: naming an inner state or emotion (or the lack of emotion) explicitly; score 3: mentioning a term or phrase, which implicitly indicates some inner state but without naming a concrete emotion; score 2: indicating some contextual behavior that attributes a meaning to the video which cannot be directly observed; score 1: formal description of the observed behavior.
Most answers for the Sadness display were distributed between the “sad, lonely” (17%), the “tired, resting” (17%), “shy, reclusive” (16%) and “calm” (13%) categories, while a smaller number of expressions belonged to the “afraid, hides” (8%) and thinking (6%) groups. The “sad, lonely” category had one of the highest percentages, but there were many expressions connected to inner states such as tired, sleepy, thinking. The words used to describe the Surprise display also covered a big array. Most answers were in the correct “surprised, confused” category (16%), while the meaning of some other groups could be also related to surprise, such as “curious, inquiring” (8%), or “nervous, startled” (6%). On the other hand, some answers either depicted the display as “happy, satisfied” (11%), “neutral, indifferent” (10%), “confident” (7%), “thinking” (6%), or “calm” (4%).
The Fear display was correctly recognized in 34% of the cases in which the used words belonged to the “afraid, retreats, shy,” and 3% to the related “nervous” category (Figure 3A). The second largest group was the “lonely, reclusive” (22%), followed by the opposing “calm” (7%), “happy” (6%), “sad” (4%) and finally the “tired” (3%) categories. The Happiness display yielded less categories, with the leading “happy, playful” at 28%, followed by groups that all convey a similarly high arousal level: “energetic, excited” (14%), “nervous, stressed” (9%), curious (3%). A high percentage of answers fall into the “angry, aggressive” group, which has a directly opposite meaning in valence, but the same level in arousal. The clearest division can be seen in the case of the Anger display (Figure 3B), in which 64% of the participants answered with words that are in the correct “angry, aggressive, attacking” category. It is followed by “nervous, startled” (11%), “energetic, excited” (6%) and “rages, hysterical” (4%). The answers conveying emotions with opposite valence but similar arousal level are also present in smaller numbers in the “happy, playful” (3%) group.
Figure 3. The categories of answers given for the Fear (A) and Anger (B) displays. The darker shades signify the correctness of the answers, the white column shows the answers that are neutral in the case of the emotion at hand, while the striped columns indicate somewhat opposing meanings.
Finally, in the Neutral display most of the answers corresponded to the intended inner state. 32% gave answers that fit into the “calm, balanced,” and 19% into the “neutral, indifferent” category. 9% attributed positive valence to the display (“happy, good mood”), while the rest of the answers fell into the “tired, resting” (7%), “bored” (4%) and “thinking” (4%) groups.
Closed Questions (Q2)
We demonstrate the success of recognition in Figure 4, with two different depictions of success: the percentage ratio of the highest scores given to the correct emotion and the percentage of answers that gave the maximum 5 scores to the correct emotion.
Figure 4. The striped columns show the largest score given to the correct emotions (in case the same high scores were given to other emotions as well as to the correct one, the 1 point was divided by the number of emotions that got the same score), the black columns show the maximum 5 score given for the correct emotions, while the gray column depicts the percentage of answers in which the participants gave 5 scores for any emotions in case of the neutral display in the Hungarian group.
In three displays more than half of the participants gave the largest score for the emotions that the agent was meant to display, while in the Surprise display 49%, in the Happiness display 46% of them gave the highest score to the right emotion. Similarly to the open questions, anger proved to be the most evident display, as 79% of the participants gave the highest score to the correct emotion, which was almost always the maximum 5 score. The differences between the answers that gave the largest value to the right emotion, and the ones that gave the maximum 5 score indicate the certainty with which the participants chose the correct emotion. In case of the surprise display, despite of almost half of the participants gave the largest score to the right emotion, it was the maximum 5 score only in 22% of the answers.
The success of the neutral state display is visible by the small number of answers (merely more than 10%) who gave the maximum 5 points to any emotions when viewing the neutral state video, indicating that the participants did not strongly associate the neutral state with any of the five emotions.
The analysis of sum of scores (Figure 5) shows that Anger is once again the most clearly recognized display, and that the neutral state received generally low scores. The closest scores are present at the Sadness display (sad: 389; afraid: 339 scores) and in the case of Happiness, where the participants also gave high scores to anger (happy: 285; angry: 237 scores).
Figure 5. Analysis of sum of scores in the Hungarian group. The X sign shows the correct emotions in each display.
We created a confusion matrix, based on which emotion received the biggest score from the participants in each display. Figure 6 depicts which emotions were mixed up more often by the participants. In the Sadness display the fear emotion was typically mixed with the sad, and in the Happiness display angry was confused with the happy emotion. The neutral display was generally considered to show happiness or sadness.
Figure 6. Confusion matrix of answers in the Hungarian group. The gray cells indicate the percentage at which each emotion received the biggest score. The correct emotions are indicated by dark borders.
The recognition of the displayed emotions was significantly higher than chance level in both the permissive and strict dichotomous success variables (permissive: all p < 0.001; strict: all p < 0.001).
The scores given to the correct emotions were significantly higher than the scores given to all the other emotions in four displays (Fear, Surprise, Sadness and Anger) (Friedman test + Wilcoxon signed rank test; ***p < 0.001, except in the Sadness display, where the difference between the correct answer (sad) and the scores of the afraid emotion was significant at p = 0.042). In the Happiness display the scores of happy and angry were not significantly different (p = 0.116) (Figure 7). The cluster analysis showed which emotions were perceived by participants as occurring together (Figure 8). Based on the results, the afraid and sad emotions had similar score patterns across displays, while the scores for happy, angry and surprise emotions were closer together. In the GLMM analysis the overall corrected model was significant [F(59, 3281) = 18.216; p < 0.001; n = 3345]. The participants' gender had no effect on the scores, and neither did the gender–display type and gender–emotion category interactions. The effects of display type [F(5, 3281) = 20.554; p < 0.001; n = 3345] and emotion category were significant [F(4, 3281) = 9.823; p < 0.001; n = 3345] and so were their interaction [F(20, 3281) = 37.802; p < 0.001; n = 3345]. The three-way interaction of gender, emotion category and display type was also significant [F(20, 3281) = 2.184; p = 0.002; n = 3345].
Figure 7. The figure shows the scores given for the six displays, while the circles show the medians for the scores given to the correct emotions by the Hungarian participants. The * indicates outliers.
Figure 8. Ward's hierarchical cluster analysis, showing the similarities of scores given to each emotion by the Hungarian participants.
Closed Questions (Q2)
The participants gave the biggest score to the correct emotion more than half of the time in the Sadness (77%) and the Anger (85%) displays, and these scores were usually the maximum 5 score which indicates the participants' high certainty when scoring the correct emotions. The percentage of answers were close to 50% for the largest value in case of the Surprise (44%) and the Fear (45%) displays, but in the Happiness display it only reached 28%. Both in the Surprise and Happiness displays the certainty was fairly low, as only 14% of the answers gave the maximum 5 score. The neutral display has rarely received a maximum score for any emotions (5%), reassuring that the participants did not strongly associate it with the five emotions (Figure 9). The analysis of sum of scores is visualized in Figure 10, showing that the Anger display received the most scores, while the neutral state received the less. Together with the confusion matrix, the results show that the sad and afraid emotions were mixed frequently in the Fear display (afraid: 47%, sad: 35%) and in the Surprise display surprised (46%) was sometimes mixed with the happy (32%) emotion. The most mismatched display was Happiness, in which only 28% of the answers gave the biggest score to happy, and most of the participants recognized it as angry (51%). The neutral display was most recognized as sad (40%), followed by afraid and happy, both only at 21%.
Figure 9. The striped columns show the largest score given to the correct emotions, the black columns show the maximum 5 score given for the correct emotions, while the gray column depicts the percentage of answers in which the participants gave 5 scores for any emotions in case of the neutral display in the Japanese group.
Figure 10. Analysis of sum of scores in the Japanese group. The X sign shows the correct emotions in each display.
The success of recognition of the displayed emotions was significantly higher than chance level in most cases in the permissive success (in the sad, afraid, surprised and angry displays p < 0.001 while it was not significant in the happy display p = 0.056). In case of the strict success the recognition was significantly higher than chance level in the sad and angry displays (p < 0.001) but not in the surprised, afraid and happy displays (p = 0.133; p = 0.267; p = 0.457). We compared the scores given to the correct emotions to the scores given to the other emotions with Friedman tests and Wilcoxon signed-rank tests and found significant difference in the Sadness (sad-happy, sad-angry, sad-surprised: p < 0.001 sad-afraid: p = 0.007) and Anger displays (angry-afraid, angry-happy, angry-sad: p < 0.001 angry-surprised: p = 0.001) where the scores of the correct emotion differed from the rest. In case of the Surprise and Fear displays only one emotion did not differ in its scores from the correct one (surprised-happy in the Surprise display and afraid-sad in the Fear display) while the rest received significantly less scores than the correct answer. The scores of the correct emotion did not differ from any other in the Happiness display. Due to only having one female participant we did not include gender in the GLMM analysis. GLMM showed that the overall corrected model was significant [F(29, 626) = 10.505; p < 0.001; n = 660], and so were the effects of display type [F(5, 626) = 10.358; p < 0.001; n = 660], emotion category [F(4, 626) = 10.402; p < 0.001; n = 660], and also their interaction [F(20, 626) = 13.402; p < 0.001; n = 660]. The hierarchical cluster analysis showed that participants usually scored displays similarly for the afraid and sad emotions, indicating that if a display was scored as sad, it was likely also scored as afraid (Figure 11). Angry, surprised and happy were closer together, therefore in the Japanese sample we can see a similar pattern of scoring, as the high arousal (angry, surprised, happy) were more likely to be scored the same in the displays.
Figure 11. Ward's hierarchical cluster analysis showing the similarities of scores given to each emotion by the Japanese participants.
We used matched subject groups to provide an appropriate basis for the intercultural comparison of Hungarian and Japanese subjects. We excluded the only female participant from the Japanese group, and selected from the total Hungarian subject group a subsample matched based on gender and age to the Japanese subjects. The Japanese sample consisted of 21 male participants (age 22.3 ± SD 1.1 years), similarly to the matched Hungarian sample which consisted of 21 participants (age 21.1 ± SD 1.3 years).
Closed-Ended Questionnaire (Q2)
Based on the confusion matrix (Figure 12) and the scores given for the six displays (Figure 13) Japanese participants gave high scores in case of the Sadness display for the sad emotion, while Hungarians gave similar scores to both the sad and afraid emotions. The Surprise display received high scores in both groups with the biggest median for the surprise emotion, but the other emotions also received high scores. In case of Fear, Hungarians gave more scores to afraid, while Japanese participants gave similar scores to the afraid and sad emotions. The Happiness display was scored mainly as happy, angry or surprised in both groups, but in the Japanese group the scores given to surprised and angry exceeded the scores given to the happy emotion. The Anger display was overwhelmingly scored as angry by both groups, but surprised and afraid also received high scores. Neither of the emotions was given high scores in the Neutral display, but Hungarians considered the display to be somewhat happy, while Japanese participants scored it rather as sad.
Figure 12. Confusion matrix for the answers of the matched Japanese and Hungarian participants. The gray cells indicate the percentage at which each emotion received the biggest score. The correct emotions are indicated by dark borders.
Figure 13. The figure shows the scores given for the six displays, while the circles show the medians for the scores given to the correct emotions by the Japanese participants and the participants in the matched Hungarian group. The * indicates outliers.
In the GLMM analysis we did not include gender due to the all-male subject groups, but included nationality as a fixed effect. The overall corrected model was significant [F(59, 1196) = 8.146; p < 0.001; n = 1260]. The effects of display type [F(5, 1196) = 13.576; p < 0.001; n = 1260] and the emotion category [F(4, 1196) = 8.215; p < 0.001; n = 1260] were significant, but not the effect of nationality, or its interaction with the display type. However, the nationality-emotion category interaction was significant [F(4, 1196) = 5.451; p < 0.001; n = 1260], as well as the three-way interaction of display type, emotion category and nationality [F(20, 1196) = 1.998; p = 0.006; n = 1260].
The first hypothesis concerned the correct recognition of emotions. The results show that in most cases the subjects could recognize the emotions that the agent's display was meant to convey in the closed-ended questionnaire. The Hungarian participants recognized the Anger display correctly based on both the descriptive and statistical analysis, and the certainty of their answers was the highest in this display as they gave the maximum score almost 80% of the time. The recognition of anger from human facial expressions is around 92% in adults in forced choice tests and 78% in a free-label test (Russell, 1994) and 88% in 5–7 years old children (Durand et al., 2007). In the eMOTO study by Fagerberg et al. (2004), some aspects of emotional expression showed similarities with our displays, as they also used more angular forms and strong red colors to express anger. These similarities might arise from the cultural use of colors, but that can also be based on the color's biological relevance, for example red is frequently found in warning colors (Stevens and Ruxton, 2011), or during the increase of facial blood flow that frequently co-occurs with anger, which leads to redness in the human face (Drummond and Quah, 2001). The strong interplay of cultural and biological expressions of anger might have contributed to the high recognition of this emotion in our agent. Although the recognition of the other emotions did not reach the usual recognition rates of facial displays of emotions (Russell, 1994; Elfenbein and Ambady, 2002), the displays were usually recognized above chance level. The Neutral display proved to be perceived as truly neutral as it did not receive high scores for any emotions. In studies examining emotions expressed by social robots the recognition rates are similar or lower than that of human facial expressions, e.g., Probo: 100–65% (Saldien et al., 2010); Kismet: 86–57% (Breazeal, 2003); Eddie: 90–20% (Gonsior et al., 2012). Although some features of these robots are animal-like or stylized, their emotion expressions are mostly based on human facial expressions. Emotions expressed via virtual faces are recognized between 97.86% (happiness) and 22.96% (disgust) (Dyck et al., 2008). In a study by Gácsi et al. (2016) the multimodal emotion expressions of a social robot were based on emotionally expressive behaviors of dogs instead of human facial expressions. The percentage of correct answers were slightly higher than in our current study (88–62%), which can be due to the physical embodiment of the robot.
In the open-ended questionnaire the answers depicting the correct emotions differed widely by display, ranging from 16% (Sadness) to 64% (Anger). These percentages are much lower than the success of recognition achieved in the closed-ended questionnaire. Which is to be expected, considering that the participants had no prior experience with similar emotion expression agents, and the open-ended format enabled them to also answer with words depicting different emotions or states than in the closed-ended format.
Following our second hypothesis, the lower rate of recognition can be explained by the similarities of behavioral attributes between some of the displays. In the open-ended questionnaire some inner states, like tired, sleepy or thinking were mentioned in the case of several displays, but their prevalence was the highest in the Sadness display. As sadness is a low arousal emotion (Russell, 1980) it can be difficult to differentiate between inner states like tiredness or sleepiness and the expression of sadness on the behavioral level, even in the case of humans (Nettle, 2009), especially if there is no additional context. The answers were the most divided in case of the Surprise display, which can originate from the lack of clear positive or negative valence, as the reaction to sudden stimuli usually depends on the situation and the stimulus itself (Russell and Barrett, 1999), making the recognition more difficult without a clarifying context. In the closed-ended questionnaire the Fear and Sadness displays were recognized correctly to a similar extent in the Hungarian sample, 53 and 59% of the answers. While the statistical analysis showed that the scores given to the correct emotion differed significantly from the rest in the Fear display, for the Sadness display there was no significant difference between scores of the afraid and sad emotions. The mixture of the afraid and sad emotions was more prevalent in this display, but it was present in the case of the Fear display too, to a lesser extent. The mismatch between these two emotion displays could be the result of their similarly negative valence and some of the shared actions of the displays (pale color, size decrease). Similarly, the other prevalent mismatch was observed in the case of the Happiness display, in which participants also gave high scores to the angry emotion. This confusion might come from the similarities of intensity levels in the happy-angry emotions which are both high arousal emotions, and show some similar behavioral elements such as fast movement, perceived increase in size and vivid colors. Surprise is also a high-arousal emotion, which is reflected in its closeness to the angry and afraid scores in the cluster analysis. This reinforces the idea that the arousal level was a decisive factor for the scores given by the participants. The low certainty of the scores given for the surprise display can be the result of the ambiguousness of the emotion, as it has neither inherently positive nor negative valence.
When we look at cultural differences we can see that in most cases the participants from the two cultures recognized the displays similarly, but we also found differences. Hungarian and Japanese participants considered the neutral state to have a slightly different valence: while Hungarians scored it as more happy, Japanese participants perceived it to be somewhat sad. Additionally, the confusion matrix shows that Japanese participants were more successful in the recognition of Sadness than Hungarians, and they also gave high scores for sadness in the Fear display. As the human fine tuning step (see preliminary study in section Procedure) was conducted with Hungarians, it is possible that parameters such as the color values of the agent in these displays (mostly light blue or fading) or the rotation speed was considered low arousal by Japanese participants. The Anger and Surprise displays received similar scores in the two groups, reflecting the results already found in case of the full Hungarian subject group. Japanese participants were less successful in the recognition of Happiness than Hungarians, and they also tended to interpret it as more angry. The Happiness display of the agent was developed to express a high intensity emotion, but in many Asian cultures happiness is associated with a more relaxed, less intensive experience (Tsai et al., 2006; Miyamoto and Ma, 2011). Although the differences show the need for cultural fine tuning, the majority of displays received the highest scores for the correct emotions, indicating that the general rules used for creating the emotion expressions can be considered a good basis for emotion expression in artificial agents.
Although our results prove that using the biological rules for creating the expressions was generally successful, our study has some limitations which has to be considered. As we only wanted to test the concept, we only used one agent to which we created the emotion expressions, and all subjects saw the exact same emotion expression displays. These two aspects can lead to pseudoreplication effects (Heffner et al., 1996), which could have been an issue in case of a different research question, one that would set out to measure the efficiency of these displays in detail, and compare them to more convenient emotion expressions. In contrast, pseudoreplication is not a major issue in this current research, as we focused on testing the applicability of the theory. In the future we intend to investigate the implemented concept on various, broadly different artificial agents. We plan to embed a randomization effect which will generate slight variations within the set parameter limits for each emotion, providing variable displays that will also ensure the avoidance of pseudoreplication.
In the study we acquired data from both Hungarian and Japanese participants, which gave us the opportunity to conduct intercultural comparisons. Although we used matched subject groups for direct comparisons, these groups consisted of considerably less participants than the total Hungarian sample, and included only male participants; therefore the intercultural aspect of our study only opens new opportunities for more detailed cross-cultural research.
The emotion expression of artificial agents, especially in social robotics and ICT is becoming an increasingly researched area and can potentially lead to more diverse and robust human-machine interactions. Artificial agents are becoming more pervasive in our day-to-day life, and are expected to become even more widespread in the near future. With the growth of social robotics and the immersion of mobile applications in our daily tasks, ethical questions that seemed theoretical a few decades ago are of practical value today. It is expected that social robots will be used in e.g., elderly care, education, households and healthcare, in which these agents will interact with various people who are not familiar with the capabilities and limitations of robots. Communication technologies and especially mobile devices are already owned by the majority of the population in western countries (Donner, 2008) and have become the integral part of our culture and the way we interact with others (Srivastava, 2005). In case of long-term interactions with social robots or use of communication devices, people may attribute higher, anthropomorphic skills and cognitive features to these agents than what they really possess (Miklósi and Gácsi, 2012). A big emphasis should be put on truthful informing about agents that show behaviors associated with higher cognitive functions. It is known that people can get attached to their smart phones (Konok et al., 2016) and in certain areas social robotics is also encouraged to develop robots whose behavior facilitate the forming of attachment between the human and the robot in long-term interactions (Miklósi and Gácsi, 2012). The development of emotion expression for artificial agents requires a cautious approach which ensures that the new technologies that choose to integrate any kind of emotion expression in their devices could not be used to mislead or exploit users, especially the ones that generally have more difficulties understanding the concepts and limitations of these devices. We strongly believe that by developing less anthropomorphic emotion expressions, and by only using them within their constraints, this approach could be favorable. We plan to conduct research on the effects of any emotion expression in artificial agents, and encourage other researchers to do likewise.
Using general biological rules to create simple expressive behaviors for artificial agents proved to be a correct approach. We wish to establish the concept further by implementing and examining the biological approach in various agents with a diverse set of features and functions, both in connection with ICT applications and social robotics. In the current agent we did not implement disgust, as it is defined more as a sensory affect instead of an emotion (Panksepp, 2007), but it might be included in future studies. The development of new agents should also focus on the individualization of the agents. This could be achieved with the innovation of the agent's appearance and structure, while fine tuning of the agent (e.g., creating long-term personalities or short-term moods) would be possible by adjusting the display of the neutral state or the intensity of emotional displays. The development of continuous, adequately dynamic emotion expression is the aim of future studies.
Data Availability Statement
Datasets are available on request; The raw data supporting the conclusions of this manuscript will be made available by the authors, without undue reservation, to any qualified researcher.
MG, ÁM, TF, PK, and PB developed the main concept, design and the new approach of the study. MG, TF, and ÁM developed the ethological basis of the emotion expressions. GP implemented them as visual displays. MG, MN, and TF tested the participants with the questionnaires. BK, MG, VK, and TF conducted the statistical analysis. BK, VK, MG, ÁM, and TF wrote the manuscript. All authors contributed to the manuscript and approved the final version.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The research was supported by the ETOCOM project (TÁMOP-4.2.2-08/1/KMR-2008-0007) through TÁMOP, the Hungarian Scientific Research Fund (OTKA T049692) and the Hungarian Academy of Sciences (MTA 01 031); the Premium Post Doctorate Grant (460002) by the Office for Research Groups Attached to Universities and Other Institutions of the Hungarian Academy of Sciences in Hungary; and the National Research, Development, and Innovation Office grant (115862K) and (K120501).
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2018.01191/full#supplementary-material
Baranyi, P., Persa, G., and Csapo, A. (2011). Definition of cognitive infocommunications and an architectural implementation of cognitive infocommunications systems. World Acad. Sci. Eng. Technol. Int. Sci. Index 5, 376–380.
Becker-Asano, C., and Ishiguro, H. (2011). “Evaluating facial displays of emotion for the android robot geminoid F,” in IEEE SSCI 2011 - Symposium Series on Computational Intelligence. - WACI 2011 Workshop on Affective Computational Intelligence (Paris), 22–29. doi: 10.1109/WACI.2011.5953147
Buyukdura, J. S., McClintock, S. M., and Croarkin, P. E. (2011). Psychomotor retardation in depression: biological underpinnings, measurement, and treatment. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 35, 395–409. doi: 10.1016/j.pnpbp.2010.10.019
Chen, C.-C., and Fernald, R. D. (2011). Visual information alone changes behavior and physiology during social interactions in a cichlid fish (Astatotilapia burtoni). PLoS ONE 6:e20313. doi: 10.1371/journal.pone.0020313
de Boer, S. F., van der Vegt, B. J., and Koolhaas, J. M. (2003). Individual variation in aggression of feral rodent strains: a standard for the genetics of aggression and violence? Behav. Genet. 33, 485–501. doi: 10.1023/A:1025766415159
Descovich, K. A., Wathan, J. W., Leach, M. C., Buchanan-Smith, H. M., Flecknell, P., Farningham, D., et al. (2017). Facial expression: an under-utilised tool for the assessment of welfare in mammals. ALTEX Altern. Anim. Exp. 34, 409–429. doi: 10.14573/altex.1607161
Drummond, P. D., and Quah, S. H. (2001). The effect of expressing anger on cardiovascular reactivity and facial blood flow in Chinese and Caucasians. Psychophysiology 38, 190–196. doi: 10.1111/1469-8986.3820190
Durand, K., Gallay, M., Seigneuric, A., Robichon, F., and Baudouin, J. Y. (2007). The development of facial emotion recognition: the role of configural information. J. Exp. Child Psychol. 97, 14–27. doi: 10.1016/j.jecp.2006.12.001
Ekman, P., Friesen, W. V., O'Sullivan, M., Chan, A., Diacoyanni-Tarlatzis, I., Heider, K., et al. (1987). Universals and cultural differences in the judgments of facial expressions of emotion. J. Pers. Soc. Psychol. 53, 712–717. doi: 10.1037/0022-35126.96.36.1992
Evans, M. R., and Norris, K. (1996). The importance of carotenoids in signaling during aggressive interactions between male firemouth cichlids (Cichlasoma meeki). Behav. Ecol. 7, 1–6. doi: 10.1093/beheco/7.1.1
Fagerberg, P., Ståhl, A., and Höök, K. (2003). “Designing gestures for affective input: an analysis of shape, effort and valence,” in MUM 2003. Proceedings of the 2nd International Conference on Mobile and Ubiquitous Multimedia, 57–65. Available online at: http://citeseerx.ist.psu.edu/viewdoc/download?&rep=rep1&type=pdf
Feldmaier, J., Marmat, T., Kuhn, J., and Diepold, K. (2016). Evaluation of a RGB-LED-Based Emotion Display for Affective Agents. Available online at: http://arxiv.org/abs/1612.07303
Fitze, P. S., Cote, J., San-Jose, L. M., Meylan, S., Isaksson, C., Andersson, S., et al. (2009). Carotenoid-based colours reflect the stress response in the common lizard. PLoS ONE 4:e5111. doi: 10.1371/journal.pone.0005111
Frijda, N. H., Mesquita, B., Sonnemans, J., and Van Goozen, S. (1991). “The duration of affective phenomena or emotions, sentiments, and passions,” in International Review of Studies on Emotion, ed K. T. Strongman (Chichester: John Wiley & Sons), 187–225.
Gácsi, M., Kis, A., Faragó, T., Janiak, M., Muszynski, R., and Miklósi, Á. (2016). Humans attribute emotions to a robot that shows simple behavioural patterns borrowed from dog behaviour. Comput. Hum. Behav. 59, 411–419. doi: 10.1016/j.chb.2016.02.043
Gockley, R., Simmons, R., Wang, J., Busquets, D., and DiSalvo, C. (2004). “Grace and George: social robots at AAAI,” in Proceedings of AAAI'04. Mobile Robot Competition Workshop, 15–20. Available online at: http://www.aaai.org/Papers/Workshops/2004/WS-04-11/WS04-11-004.pdf
Gonsior, B., Sosnowski, S., Buss, M., Wollherr, D., and Kuhnlenz, K. (2012). “An emotional adaption approach to increase helpfulness towards a robot,” in 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (Vilamoura: IEEE), 2429–2436. doi: 10.1109/IROS.2012.6385941
Herrera-Pérez, J. J., Martínez-Mota, L., and Fernández-Guasti, A. (2008). Aging increases the susceptibility to develop anhedonia in male rats. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 32, 1798–1803. doi: 10.1016/j.pnpbp.2008.07.020
Kim, H., Somerville, L. H., Johnstone, T., Polis, S., Alexander, A. L., Shin, L. M., et al. (2004). Contextual modulation of amygdala responsivity to surprised faces. J. Cogn. Neurosci. 16, 1730–1745. doi: 10.1162/0898929042947865
Konok, V., Gigler, D., Bereczky, B. M., and Miklósi, Á. (2016). Humans' attachment to their mobile phones and its relationship with interpersonal attachment style. Comput. Hum. Behav. 61, 537–547. doi: 10.1016/j.chb.2016.03.062
Konok, V., Nagy, K., and Miklósi, Á. (2015). How do humans represent the emotions of dogs? The resemblance between the human representation of the canine and the human affective space. Appl. Anim. Behav. Sci. 162, 37–46. doi: 10.1016/j.applanim.2014.11.003
Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., and Matthews, I. (2010). “The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression,” in 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition – Workshops (San Francisco, CA: IEEE), 94–101. doi: 10.1109/CVPRW.2010.5543262
Michalak, J., Troje, N. F., Fischer, J., Vollmar, P., Heidenreich, T., and Schulte, D. (2009). Embodiment of sadness and depression—gait patterns associated with dysphoric mood. Psychosom. Med. 71, 580–587. doi: 10.1097/PSY.0b013e3181a2515c
Panksepp, J., Yates, G., Ikemoto, S., and Nelson, E. (1991). “Simple ethological models of depression: social-isolation induced ‘despair’ in chicks and mice,” in Animal Models in Psychopharmacology, eds B. Olivier, J. Mos, and J. L. Slangen (Basel: Birkhäuser Basel), 161–181. doi: 10.1007/978-3-0348-6419-0_15
Palagi, E., Burghardt, G. M., Smuts, B., Cordoni, G., Dall'Olio, S., Fouts, H. N., et al. (2016). Rough-and-tumble play as a window on animal communication. Biol. Rev. 91, 311–327. doi: 10.1111/brv.12172
Park, J., Barash, V., Fink, C., and Cha, M. (2013). “Emoticon style : interpreting differences in emoticons across cultures,” in Proceedings of the 7th International AAAI Conference on Weblogs and Social Media (ICWSM) (Cambridge, MA), 466–475.
Persa, G., Csapo, A., and Baranyi, P. (2012). “A pilot application for ethology-based CogInfoCom systems,” in 2012 IEEE 10th International Symposium on Applied Machine Intelligence and Informatics (SAMI) (Herl'any: IEEE), 477–482. doi: 10.1109/SAMI.2012.6209014
Russell, J. A., and Barrett, L. F. (1999). Core affect, prototypical emotional episodes, and other things called emotion: dissecting the elephant. J. Pers. Soc. Psychol. 76, 805–819. doi: 10.1037/0022-35188.8.131.525
Safdar, S., Friedlmeier, W., Matsumoto, D., Yoo, S. H., Kwantes, C. T., Kakai, H., et al. (2009). Variations of emotional display rules within and across cultures: a comparison between Canada, USA, and Japan. Can. J. Behav. Sci. Can. des Sci. du Comport. 41, 1–10. doi: 10.1037/a0014387
Song, S., and Yamada, S. (2017). “Expressing emotions through color, sound, and vibration with an appearance-constrained social robot,” in Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot Interaction – HRI'17 (New York, NY: ACM Press), 2–11. doi: 10.1145/2909824.3020239
Sundström, P., Ståhl, A., and Höök, K. (2005). “eMoto – affectively involving both body and mind,” in CHI'05 Extended Abstracts on Human Factors in Computing Systems – CHI'05 (New York, NY: ACM Press), 2005–2008. doi: 10.1145/1056808.1057078
Terada, K., Yamauchi, A., and Ito, A. (2012). “Artificial emotion expression for a robot by dynamic color change,” in 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication (IEEE), 314–321. doi: 10.1109/ROMAN.2012.6343772
Topál, J., Gácsi, M., Miklósi, Á., Virányi, Z., Kubinyi, E., and Csányi, V. (2005). Attachment to humans: a comparative study on hand-reared wolves and differently socialized dog puppies. Anim. Behav. 70, 1367–1375. doi: 10.1016/j.anbehav.2005.03.025
Trezza, V., Baarendse, P. J. J., and Vanderschuren, L. J. M. J. (2010). The pleasures of play: pharmacological insights into social reward mechanisms. Trends Pharmacol. Sci. 31, 463–469. doi: 10.1016/j.tips.2010.06.008
Verduyn, P., Delaveau, P., Rotgé, J.-Y., Fossati, P., and Van Mechelen, I. (2015). Determinants of emotion duration and underlying psychological and neural mechanisms. Emot. Rev. 7, 330–335. doi: 10.1177/1754073915590618
Vianna, D. M., and Carrive, P. (2005). Changes in cutaneous and body temperature during and after conditioned fear to context in the rat. Eur. J. Neurosci. 21, 2505–2512. doi: 10.1111/j.1460-9568.2005.04073.x
Keywords: emotion recognition, artificial agent, human-computer interaction, human-robot interaction, ethological approach, ethorobotics, artificial emotion expression
Citation: Korcsok B, Konok V, Persa G, Faragó T, Niitsuma M, Miklósi Á, Korondi P, Baranyi P and Gácsi M (2018) Biologically Inspired Emotional Expressions for Artificial Agents. Front. Psychol. 9:1191. doi: 10.3389/fpsyg.2018.01191
Received: 24 April 2018; Accepted: 20 June 2018;
Published: 13 July 2018.
Edited by:Anna Esposito, Università degli Studi della Campania “Luigi Vanvitelli” Caserta, Italy
Reviewed by:Rytis Maskeliunas, Kaunas University of Technology, Lithuania
Maria Teresa Riviello, International Institute for Advanced Scientific Studies “Eduardo R. Caianiello“, Italy
Copyright © 2018 Korcsok, Konok, Persa, Faragó, Niitsuma, Miklósi, Korondi, Baranyi and Gácsi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Beáta Korcsok, firstname.lastname@example.org