Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Robot. AI, 08 October 2025

Sec. Human-Robot Interaction

Volume 12 - 2025 | https://doi.org/10.3389/frobt.2025.1637574

Personalized causal explanations of a robot’s behavior

  • 1Department Tecnología Electrónica, University of Málaga, Málaga, Spain
  • 2Department Computing Science, University of Umeå, Umeå, Sweden

The deployment of robots in environments shared with humans implies that they must be able to justify or explain their behavior to nonexpert users when the user, or the situation itself, requires it. We propose a framework for robots to generate personalized explanations of their behavior by integrating cause-and-effect structures, social roles, and natural language queries. Robot events are stored as cause–effect pairs in a causal log. Given a human natural language query, the system uses machine learning to identify the matching cause-and-effect entry in the causal log and determine the social role of the inquirer. An initial explanation is generated and is then further refined by a large language model (LLM) to produce linguistically diverse responses tailored to the social role and the query. This approach maintains causal and factual accuracy while providing language variation in the generated explanations. Qualitative and quantitative experiments show that combining the causal information with the social role and the query when generating the explanations yields the most appreciated explanations.

1 Introduction

Making robots understandable is generally acknowledged as important for improving safety, user experience, trust, and efficiency (Hellström and Bensch, 2018). Understandable robots may, for example, verbally explain their actions and decisions as a response to questions asked by interacting humans. Such explainability is especially important in sensitive settings, such as eldercare or medical assistance, where lack of information or clarity could result in physical or psychological harm.

Explainability and causality are closely intertwined concepts (Setchi et al., 2020). Lewis, in his seminal work, described explaining an event as “providing information about its causal history” (Lewis, 1986). Earlier related work (Lindner and Olz, 2022; Chakraborti et al., 2017) often shares this focus on actions, effects, and their technical causes, that is, on what to explain.

However, an explanation solely based on cause and effect reasoning might not be sufficient if the robot interacts with humans with diverse backgrounds. For example, in an eldercare home, humans interacting with an assistive robot may be residents, medical staff, family members, or technicians. We build on several observations.

First, not all inquirers have the same wishes and needs for information. For example, a family member and a nurse asking “Why did Maria choose to eat meat today?” may want different aspects included in an explanation. The social role of the user also influences which words and expressions are appropriate to use. Second, the tone and wordings of a query may express an intentional or latent wish to have certain aspects addressed in the causal explanation. For example, a family member asking “Why did my father Alberto have meat today?” or “Why did my father Alberto have meat today again?” may reflect a wish for different aspects to be included in the explanation. To address the two observations above, we take the social role of the person asking the question and the question itself into account when generating explanations, thereby addressing the additional focus on how to explain.

In this paper, we present a framework for robots to generate personalized explanations for events when a human requests an explanation. Our approach maintains factual and causal correctness, which are crucial for any robotic application, while leveraging a large language model (LLM) to personalize and diversify the language of the explanations. Providing linguistic variation of the explanations is important for humans who interact with the robot on a regular basis to avoid monotony and reduced engagement.

More specifically, robot events (e.g., actions and tasks) are stored in a causal log where they are structured into cause and effect pairs. Given a human query q, we use machine learning to extract the robot event e for which an explanation is requested and to identify the likely social role s of the human asking. For the identified robot event, the corresponding cause and effect in the causal log are processed to provide an initial explanation. This initial explanation is then refined by an LLM (Llama) to produce linguistically varied explanations that are tailored to the social role and the actual query. The initial explanation that is based on the data from the causal log maintains factual correctness, whereas the LLM adds language variation.

To evaluate our approach, three quantitative and qualitative experiments were performed. Experiment I investigated the effects of including combinations of cause and effect, social role, and query when generating an explanation. Experts and 30 participants assessed the quality of the generated explanations, and a statistical analysis of the results indicates that combining cause and effect with the human’s social role and the original query yields the most preferred explanations. The findings further show that the second most preferred option includes cause-and-effect structures and social roles (compared to cause-and-effect and query combinations). This indicates that incorporating the social role in causal explanations plays a significant role. Experiment II verifies that the social role of the enquirer is identified with high accuracy, and Experiment III shows a high semantic similarity between the system-generated explanations and human-generated explanations that served as ground truth. The paper is organized as follows. Section 2 consists of Section 2.1 that provides an overview of related work and Section 2.2 in which basic terms and concepts used in the paper are introduced. Section 3 describes the proposed methodology for the generation of explanations, followed by a description of experiments and results in Section 4. Section 5 discusses challenges and limitations, and Section 6 finalizes the paper with conclusions and ideas for future work.

2 Technical background

2.1 Related work

2.1.1 What is an explanation?

Humans have an innate tendency to construct explanations, a process crucial for understanding and making sense of the world around us. Explanations help build the foundation for reasoning and generalization (Lombrozo, 2006). According to Federer et al. (2015) and Meyer and Schnell (2020), explanations are composed of two key elements: the explanandum, the phenomenon being explained, and the explanation itself, which provides the rationale or reasoning. On a broader level, Norris et al. (2005) defined an explanation as an “act intended to make something clear, understandable, or intelligible” (p. 546). For explanations to be effective, they must be meaningful and adapted to the abilities and needs of the audience (Stefani and Tsaparlis, 2009). Similarly, Tania Lombrozo characterized explanations as bridging the gap that enables others to comprehend an event (Lombrozo, 2006). This process is inherently cognitive, involving answers to questions—often framed as “why”—that are typically constrained by context. The recipient of an explanation is usually less interested in the mere occurrence of an event and more focused on understanding why it happened in a particular instance rather than in alternative, counterfactual scenarios (Matarese et al., 2021).

The idea of crafting a universal theory of “good explanations” has been explored but remains unresolved due to the distinct needs of different disciplines (Pitt, 2009). For example, engineering often requires professionals to clearly communicate their decisions and solutions, as part of their nontechnical competencies (Kaplar et al., 2021). Reverse engineering, in particular, focuses on understanding “how existing artifacts produce their overall functions in terms of underlying mechanisms” (van Eck, 2015).

From a technological point of view, Pitt (2009) advocated for a theory of technological explanations to clarify the purpose and functionality of artifacts. He argued that such explanations must reference the broader system in which a tool operates, as its design, function, or structure … “can only be adequately explained by reference to the system” (p. 861). Although scientific explanations aim to reveal “why the world works the way it does” in specific contexts (p. 862), technological explanations address practical questions such as “How does this work?” or “Why does the artifact do this?”.

2.1.2 Explanations in HRI

In human–robot interaction (HRI), the ability to provide clear and meaningful explanations is essential for the widespread acceptance of robots in critical tasks (Edmonds et al., 2019). Studies, such as those by Sakai and Nagai (2022) and Chakraborti et al. (2021), highlight that true communication between humans and robots requires more than a basic understanding of commands or questions. Robots must also be capable of recognizing and interpreting a person’s internal state while conveying their own reasoning in ways that humans can easily grasp (Zakershahrak and Ghodratnama, 2020). For a robot to anticipate the needs of a person or adapt its behavior effectively, it must possess mechanisms to infer human intentions and provide clear and actionable insights into its own actions. This capability not only helps robots align with human expectations but also builds trust. Robots equipped with these abilities are referred to as explainable autonomous robots (XARs) (Stange et al., 2022; Sakai and Nagai, 2022). This concept parallels explainable artificial intelligence (XAI), but with a critical distinction. Whereas XAI aims to enhance understanding and control by offering transparent justifications for decisions often centered on data-driven processes (Gjærum et al., 2023; Adadi and Berrada, 2018), XARs are primarily concerned with explaining their autonomous behaviors in a shared, dynamic environment. As Sakai and Nagai emphasized, this involves a shift from data-focused explainability to goal-driven explainability, where the robot must clearly articulate the rationale behind its actions in pursuit of its objectives (Sakai and Nagai, 2022; Zakershahrak and Ghodratnama, 2020). This focus is crucial for fostering effective collaboration and trust in human–robot partnerships. “Understandability” is suggested as a broader term than explainability (Hellström and Bensch, 2018) and covers not only a robot’s actions but also entities such as intentions, desires, knowledge, beliefs, emotions, perceptions, capabilities, and limitations of the robot. Furthermore, understandability may be achieved not only by uttering verbal explanations but also by other modalities and even by the robot’s actual motions and actions.

2.1.3 Speaker role recognition

Typically framed in the process of determining the speaker’s turn in a homogeneous speech segment, speaker role recognition (SRR) seeks to determine the role of a speaker, considering that this role is characterized by the task performed by the speaker and the goals related to it (Flemotomos et al., 2019). Obviously, to solve the SRR problem, patterns must be identified to differentiate these roles. Different proposals have looked for these patterns at the low level, either in audio files or in rhythm and sound (e.g., the interviewer will use more interrogative words than the interviewee). Other authors agreed that language usually incorporates more information to solve this problem (Flemotomos et al., 2019; Prasad et al., 2022; Zuluaga-Gomez et al., 2023), so the aim is to exploit lexical variability to differentiate roles. Both sets of acoustic and lexical patterns can be used together, in approaches that combine automatic speech recognition (ASR) and SRR (Blatt et al., 2024). In any case, in recent years, traditional solutions have been replaced by deep learning.

Earlier work is built upon and advanced in this paper by integrating the enquirer and their mental state (through their social role) and the natural language query into the generation of more personalized explanations while maintaining causal correctness of the description of the robot’s behavior. As our experiments show, this approach results in explanations that are more appreciated by the enquirer and, in turn, make the robot more understandable and valuable to the interacting human.

2.2 Formalism and terminology

In this section, we introduce the terminology that will be used throughout the paper.

2.2.1 Literals, robot event, and causal robot event

Starting with the basic entities the robot operates on, we define a literal as follows: Definition 1 [Literal] A Literal is a placeholder for a physical or virtual entity that the robot considers in its operation. Notation: literal_name or literal_name1(literal_name2).

Literals are frequently used in the program code controlling the robot and are necessary when describing the operation of the robot. Examples of literals are as follows: 1, Jose, Menu(Maria), No_of_choices, Full, and Empty. The functional notation literal_name1(literal_name2) should be interpreted as a specification of literal_name1. For example, Menu(Maria) refers to the specific menu that is connected to Maria. Moving on to entities that the robot may be asked to explain, we consider three categories of basic robot operations:

Sensing or perception (e.g., the robot detects an object, a person, or low battery level).

Cognition (e.g., the robot estimates the distance to an object or associates a value with a literal).

Acting (e.g., the robot moves to a certain location or asks a person about their food preferences).

Based on these categories, we define a robot event as follows:

Definition 1. (Robot event). A robot event is a predicate (i.e., a Boolean expression that evaluates to True or False) with zero or more arguments and represents a specific robot operation. The arguments argi are literals. Notation: event_name(arg1, …, argk), k0.

For example, the one-argument event Start_move_to_safe_distance(Jose) represents the action where the robot moves away from the person identified as Jose, and the two-argument event Assign(No_of_choices, 3) represents the robot’s cognitive operation of associating the value 3 with the literal No_of_choices. As an alternative notation, this event may also be denoted as No_of_choices = 3. Similarly, the event Assign(Menu(Maria), Full) may also be denoted by Menu(Maria) = Full.

To describe the reasons why events occur, we introduce the notion of a Causal robot event, defined as follows:

Definition 2. (Causal robot event). A causal robot event comprises a timestamped rule:

tα1,α2,,αkβ,

where t is the timestamp and the antecedent [α1,α2,,αk] is a list of robot events that were all True at time t, which caused the consequent robot event β to happen. The antecedent is referred to as the Cause and is sometimes denoted by α as a short notation. The consequent is referred to as the Effect.

Two examples of causal robot events are as follows:

Effect β: “Use_case_menu_started (Jose)”

     Causes α1,α2,α3: (“Person_detected (Jose),” “Menu (Jose) = False,” “Therapy_time = False”)

     timestamp t:123213123

     At time t, all events in Cause were True, and the Effect occurred.

Effect β: “Person_detected (Jose)”

     Cause α: ()

     timestamp t:123213923

     The Effect occurred at time t without any specified prior conditions to hold true; that is, a person was detected independently of events internal to the robot.

It should be noted that for causal events with non-empty causes, the “causation” reflects how the software controlling the robot is written: certain conditions (the Cause) lead the program to follow a path where Effect is performed. The only exception is Effects related to perception, which depend on external conditions, and the Cause is, in these cases, an empty list.

2.2.2 Causal log and dictionary

The causal log is a component of a complex cognitive robot architecture CORTEX (Galeas et al., 2025). It is a tabular high-level episodic memory representation, and the causal log entries are automatically generated in real-time settings. In particular, the cause column is filled in with states/actions extracted from behavior trees that control the robot’s behavior. These behavior trees are defined at design time and specify both task steps and conditions that trigger robot behavior changes. When such a change occurs, the system automatically records one row in the causal log.

Formally, all occurrences of causal events are recorded in the causal log, which is defined as follows:

Definition 3. (Causal log). A causal log is a table with numbered rows. Three columns represent causal events: the timestamp t, the effect β, and the cause α1,,αk. The additional column cause_idx contains a list (possibly empty) of links that connect each αi with a prior occurrence of a causal event with the effect equal to αi. More precisely, the cause_idx at row j is a list (r1,,rk) of row numbers in the causal log, where each ri belongs to condition αi in the cause at row j. ri is set to the largest row number less than j, for which the effect column equals αi.

An example of a causal log is given in Table 1.

Table 1
www.frontiersin.org

Table 1. The robot records all occurrences of causal events in the causal log, which is represented as a table.

The generation of explanations requires descriptions of all events that may occur in the causal log. Such descriptions are provided in a dictionary, as exemplified in Table 2. Parameterized event names ep and descriptions desp reduce the size of the dictionary considerably. Parameters are instantiated when the dictionary is used to find initial basic descriptions for robot events in the causal log that should be explained. A dictionary is formally defined as follows:

Table 2
www.frontiersin.org

Table 2. The dictionary provides basic descriptions desp for events ep.

Definition 4. (Dictionary). A dictionary is a table where all possible events are listed in separate rows. The table has two columns: parameterized event ep and parameterized event description desp. Both columns may contain variables &n that are substituted for actual values when explanations are generated.

2.2.3 Questions

A person interacting with a robot may ask the robot several types of questions. In this paper, we restrict the scope to questions about the reasons for robot events that appear in the robot’s causal log. Hence, the robot can answer questions related to what it has perceived (sensing), “thought about” (cognition), and done (acting). Examples of such questions are “Why are you asking me about physio therapy?,” “How come you returned to the charging station?,” and “Why are you moving around in the living room?”.

2.2.4 Causal explanation

A question may be answered with a causal explanation. A causal explanation for an effect β in a causal event appearing in a causal log is a text string describing the direct and indirect causes of β. The direct causes α1,,αk are given in the corresponding column in the causal log. The indirect causes of β can be found by, for each αi, looking up rows in the causal log where the effect column matches αi. For this, the row numbers in the cause_idx column are utilized. Each cause of αi is an indirect cause of β. This can, in principle, continue recursively until all causes are empty—corresponding to events that cannot be explained by referring to other events in the causal log, but rather to external conditions (e.g., related to perception). Although an explanation may be observed as more “accurate” if it also refers to indirect causes, it may become overly complex and difficult to understand. The optimal trade-off between complexity and understandability depends on factors such as the purpose of the explanation, the user’s ability to understand detailed and complex information, and the time available to communicate the explanation. In this paper, we only consider direct causes when constructing explanations. However, the methodology presented can also handle the recursive inclusion of indirect causes.

3 Methodology

3.1 Methodological overview

Figure 1 shows an overview of the methodology that generates a final explanation expfinal in response to a user’s question q. It comprises the following six steps:

1. Input of a natural language query q from a user.

2. Recognition of a robot event e and the user’s social role s from q. This is done using the intent recognition component in Rasa, trained using supervised learning, as described in Section 3.2.

3. Retrieval of the effect β and the cause α from the causal log, as described in Section 3.3.

4. Generation of initial causal explanation expinit based on descriptions in the dictionary for β and α (see Section 3.4).

5. Syntactical refinement of the initial explanation expinit, resulting in a grammatically structured explanation expsyn (see Section 3.4).

6. Generation of final explanation expfinal by prompting an LLM with a combination of the three factors expsyn, s, and q (see Section 3.5).

Figure 1
Flowchart depicting six stages of generating a final explanation. Stage 1: Input event and social role recognition by RASA. Stage 2: Retrieves effect and cause. Stage 3: Generates initial causal explanation using a dictionary. Stage 4: Refines syntax. Stage 5: Produces final explanation with Llama. Arrows indicate process flow and interconnections between stages.

Figure 1. Methodological overview illustrating the process of generating personalized causal explanations based on user queries, social roles, and cause-and-effect structures.

An example of the inputs and outputs in each step of the process is given in Table 3. In the last step, all the three factors expsyn, s, and q were used to generate the final explanation expfinal.

Table 3
www.frontiersin.org

Table 3. Examples of inputs and outputs in each step of the process illustrated in Figure 1.

The following subsections provide detailed descriptions of the steps in the process.

3.2 Recognition of the robot event and social role

In this step, the user’s natural language query q is mapped to a parameterized robot event e and a social role s related to the user. We utilized Rasa’s intent recognition component, which is intended to find mappings from natural language utterances to the speaker’s intent. For example, the intents behind “Good morning” and “What do you mean?” may be “greeting” and “clarification,” respectively. This mapping is created through supervised learning by providing pairs of utterances and intents. We used the same learning functionality but provided pairs of queries q and parameterized robot events ep (see Definition 2.2.2), such as the query “Why did my father, Alberto, have cognitive therapy today?” and the parameterized event “Use_case_cognitive_started (&1).” The model for robot event recognition was trained on 25 parameterized events ep, each associated with 20 query–event pairs (qp,i,ep), 1i20, resulting in a total of 500 training pairs. Half of the queries qp,i were collected from real users (e.g., residents, family members, and technical staff) during a previous project where a robot was deployed in a retirement home in Malaga, Spain (Jerez et al., 2024). This ensured that the queries reflected the actual concerns and needs of the target user groups. The other half of the queries qp,i were linguistic variations of the real users’ queries, generated using ChatGPT. Appropriate event templates from the dictionary were manually matched with each query. Examples of the robot event recognition are shown in Table 4, where system output for four use cases are shown: choosing from the menu, cognitive therapy, robot moving around in common room (i.e., wandering), and interaction with a human.

Table 4
www.frontiersin.org

Table 4. Examples of robot event recognition.

Similarly, for social role recognition, we provided pairs of queries q and social roles s as training data. We considered the three social roles: technician, resident, and family member. The social role of a technician refers to a professional who is responsible for the maintenance of the social robot used in, for example, a retirement home. Technicians engage on a technical level with the robot, and robot explanations tailored to technicians should be focused on technology-oriented aspects. The social role of a resident refers to an independent older adult who resides in a retirement home and interacts with the robot. The interaction between a resident and social robot focuses on aspects that influence their daily lives (e.g., choosing lunch, cognitive therapy, and playing games in the common room.). Explanations for the resident should not be technical, but rather clarifying, easily understandable, accurate, and adaptive to the needs and preferences of the resident. The social role of a family member refers to visitors to the retirement home (e.g., children, grandchildren, relatives, and friends). The explanations to family members should also be clear, correct, and potentially more personal. Social roles guide how humans interact with each other and affect both what we talk about and how we talk (Eastman, 1985). Related to explanations, it is reasonable to assume that the social role of a person affects both the topic of a question asked and the way the question is formulated. An ideal explanation also takes the social role into account. Therefore, it would be advantageous for a robot that generates explanations to know the social role of the user. To this end, we trained a model that maps a user’s questions to her social role. The model was trained with a total of 160 training pairs (qk,i,sk), where each of the three roles sk was manually associated with 40 queries qk,i. Of these queries, 20 were sourced from real users based on the previous deployment of the robot in the retirement home, whereas 20 were linguistic variations generated using ChatGPT. Examples of the social role recognition are provided in Table 5.

Table 5
www.frontiersin.org

Table 5. Examples of social role recognition.

The remaining subsection provides a more detailed description of how the Rasa system was configured and used. In Rasa, natural language text is processed by a sequence of components in a so-called processing pipeline defined in a configuration file. The text is first split into tokens using the WhitespaceTokenizer, which creates one token for each sequence of characters separated by whitespace. The Featurizer component transforms these tokens into numerical representations (features). We utilized the RegexFeaturizer, which extracts features based on predefined regular expressions; the LexicalSyntacticFeaturizer, which extracts lexical-syntactic features using a sliding window approach; and the CountVectorsFeaturizer at both the word and character levels in the input query. At the word level, features were extracted using the bag-of-words approach capturing the frequency of occurrence of words. At the character level, features representing sub-word structures were extracted, making the model more robust to spelling variations or unseen words. Next, we utilized the DIETClassifier (Dual Intent and Entity Transformer), a component designed for joint intent classification and entity recognition, which is built on a transformer-based architecture. The DIETClassifier shares a transformer for both tasks, using a conditional random field (CRF) layer for entity recognition and a semantic vector space for intent classification (in our case, robot event and social role classification). The model optimizes by maximizing the similarity between the predicted intent vector and the correct label using dot-product loss. The outputs are the primary intent and the intent ranking with respective confidences scores. Table 4 shows various examples of the classifier’s output. Finally, the FallbackClassifier classifies a query with the intent “nlu_fallback” if the intent’s confidence score is below a defined threshold (30% in our case). The fallback intent can also be predicted when the confidence scores of the two top ranked intents are closer than the ambiguity threshold (10% in our case).

3.3 Retrieval of the effect and the cause

Step 2 results in a parameterized event ep, such as “Use_case_cognitive_started(&1),” and a named entity such as “Alberto.” In step 3, they are compared with the events in the effect column in the causal log to find rows with a matching effect β, such as “Use_case_cognitive_started(Alberto).” If more than one such row exists, the row most recently added to the causal log is chosen. In addition to the effect β, the cause α, the timestamp t, and the cause index cause_idx are retrieved from the same row. Examples of effect and cause retrieval for four queries are provided in Table 6.

Table 6
www.frontiersin.org

Table 6. Examples of effect and cause retrieval for four queries.

3.4 Generation of causal explanations

An initial natural language description of the retrieved cause α=α1,,αk and effect β is first generated by utilizing the dictionary (see Section 2.2.2). Parameter values within parentheses, such as “Alberto,” are first substituted with placeholders (e.g., &1). The resulting α1,,αk and β are then matched with the parameterized event names ep in the dictionary. The corresponding event descriptions are then concatenated, and all parameters are substituted back. Separators “; ” are finally added to form the initial causal explanation expinit. Examples are shown in Table 7.

Table 7
www.frontiersin.org

Table 7. Examples of initial causal explanations expinit for four queries, generated by concatenating descriptions for cause and effect in the dictionary.

expinit is syntactically enhanced to form a grammatically cohesive sentence that is better suited for the final refinement stage handled by the LLM. We observed that the LLM consistently produced more fluent and natural explanations when the input was a single connected sentence, rather than a sequence of disjointed clauses or bullet points. The word “because” is inserted at the beginning of the first cause to explicitly mark causality. If there are multiple causes, they are joined using commas between clauses, with the word “and” before the final cause to form a coordinated list. Additionally, occurrences of the word “Person” are removed for better readability, resulting in the syntactically enhanced explanation expsyn. Examples are provided in Table 8.

Table 8
www.frontiersin.org

Table 8. Examples of syntactically enhanced causal explanations expsyn for four queries, created by adding conjunctions such as “and” or “as a result” to expinit.

3.5 Generation of final explanation

Although the generated explanations expsyn (see Table 8) contain necessary and correct information, they are not sufficiently well expressed to be easily understood. A final adaption is, therefore, performed by utilizing the LLM Llama1, which allows for local and not only cloud-based processing. This enables responsive real-time interaction, suitable for on-device usage in social robotics applications. The specific model used was the Meta-Llama-3.1-8B-InstructQ4_k_M.gguf, with 8 billion parameters. It is tuned for generating instructional and conversational responses, allowing it to respond accurately and naturally to user queries.

Section 4 describes how different combinations of the factors expsyn, s, and q were evaluated for the generation of the final explanation expfinal. For the combination with all factors, Llama was called with the following prompt: “You are a social assistive robot. According to the role [s] of the target that you are answering and the question asked [q]. Compress the answer with the meaningful information: [expsyn].” For other combinations of factors, the prompt was adjusted correspondingly.

4 Evaluation

The proposed methodology was implemented on a computer and evaluated using three experiments. Before providing detailed descriptions of these experiments, we provide a summary of the design and results of the experiments.

Experiment I assessed the quality of various system-generated explanations in terms of how well they are tailored to a specific social role. The system-generated explanations were assessed by experts and by 30 recruited participants filling out questionnaires. The vast majority of the participants were staff at the Electronic Technology Department at the University of Malaga, with some technical knowledge of programming and physical robotic systems. The results established that the most preferred explanations were generated by including the original query q and the social role s in the LLM prompt used to refine expsyn to the final explanation expfinal (step 6 in Figure 1).

Experiment II investigated how well the system inferred the social role, given a natural language query. For this, all queries used in Experiment I were tested, and in addition, 30 survey participants generated queries to an imagined robot, assuming one of the social roles of family member, resident, and technician. As summarized in Table 10, the accuracy for recognition of the three social roles varied between 83% and 90%.

Experiment III evaluated the system-generated explanations by calculating the cosine similarity between human-generated explanations, serving as ground truth, and system-generated explanations. As reported in Table 11, the cosine similarities were maximized for system-generated explanations aimed for the same social role as the human-generated explanations. For all three roles, these similarity scores were higher than 86%. The three conducted experiments are described in detail in the following subsections.

4.1 Experiment I

The first experiment aimed at investigating the effect of providing various types of information to the LLM for the generation of explanations. In particular, for each original query, four use cases were considered: Why did Alberto choose from the menu today?, Why did you give cognitive therapy to Alberto?, Why did you detect Jose?, and Why are you starting to move around; corresponding modified questions, tailored to one of the social roles of family member, resident, and technician, were then generated. For each such modified question, four combinations of expsyn, s, and q were used to formulate four LLM prompts, as described in Section 3.5. The combinations, denoted C1C4, were defined as follows:

C1: expsyn.

C2: expsyn and q.

C3: expsyn and s.

C4: expsyn, s, and q.

The quality and adequacy of the output of the final explanation expfinal for C1C4 was then assessed by 30 test participants. In particular, they were asked to mark which one of the explanations generated through the four combinations C1C4 they preferred, considering semantic adequacy, linguistic efficiency, temporal contextualization, and social role. By the term contextualization, we refer to correct identification of the cause-and-effect structures, and by the term personalization, we refer to the explanations being tailored to the social role of the human and to the linguistic subtleties in the natural language query. Note that each participant assessed the suitability of the generated questions for assumed interlocutors with different social roles. The age distribution of the participants was as follows: under 30 years (6.67%), 30–44 years (23.33%), 45–60 years (43.33%), and over 60 years (26.67%). A total of 56.67% of the participants were male and 43.33% were female participants. Details are reported in the next section, and a tabular summary of the 48 explanations preferred by the participants is provided in Section 4.1.1.

Additionally, the authors of the paper conducted an expert evaluation of the obtained final system-generated explanation regarding the identification of social roles s and events e, and also the differences between the four combinations C1C4 of information (see details in Section 4.1.1).

For all examined cases, the social role s recognized by Rasa as described in Section 3.2 matched the social role assumed when constructing the three modified social role questions. All events e were also correctly recognized for all modified questions. As a result, the initial explanations expinit (and thereby expsyn) were also the same.

In the following, all use case queries and all 48 assessed explanations are listed.

For ease of reference, each explanation is marked with a sequential number within parentheses. The percentage of test participants preferring one of the four versions C1C4 is added after the explanation.

Example 1. System-generated explanations for use_case_menu.

Original query: Why did Alberto choose from the menu today?

Recognized event e: Use_case_menu_started (Alberto).

expsyn: Because the robot ended the cognitive therapy as it only asks a limited number of questions to avoid exhausting the user, Alberto was detected by the robot’s facial recognition camera, Alberto did not have the cognitive therapy activity scheduled for today or has just completed it, and Alberto had not ordered any food. As a result, the robot approached a recognized person Alberto and asked for his menu choices. Menu selection was made by touching appropriate images on the robot’s screen.

4.1.1 Social role: family

Modified social role question q: Why did my father Alberto choose his menu today?

Generated final explanations expfinal:

C1: (1) I approach Alberto because I detected him through my camera, and as he does not have a scheduled therapy session or a food order, I am here to help him choose his meal. 0%

C2: (2) I approached Alberto because he was recognized by my facial recognition camera, and as he did not have a scheduled cognitive therapy session today, I asked him to order his meal by selecting from the menu on my screen. 0%

C3: (3) Good morning, Alberto! I am here to help. I noticed you are here, so I shall ask, would you like to order some lunch today? 0%

C4: (4) I helped your father order the menu because I finished his therapy session and saw that he had not eaten yet, so I offered him the menu on my screen. 100%

4.1.2 Social role: resident

Modified social role question q: I am Alberto, why did you ask me to select my menu?

Generated final explanations expfinal:

C1: (5) I approach Alberto because I detected him through my camera and he was not scheduled for therapy today, and he has not ordered food, so I am offering him the menu to make a selection. 0%

C2: (6) I asked you to select your menu, Alberto, because I detected you with my camera and saw you were not in therapy today, and you have not ordered food yet, so I wanted to help you choose your meal. 0%

C3: (7) I am here to assist you, Alberto. I noticed you were here, so I shall show you the menu for today. Can you please touch the images on my screen to choose what you would like to eat? 26.67%

C4: (8) I asked you to select your menu because I recognized you, saw you had finished therapy, and knew you had not ordered yet, so I helped you choose your meal. 73.33%

4.1.3 Social role: technician

Modified social role question q: Why did the system ask Alberto to choose from the menu today?

Generated final explanations expfinal:

C1: (9) I approach Alberto because I detected him with my camera, and as he does not have a scheduled therapy session or a food order, I am offering him the menu to select his meal. 6.67%

C2: (10) I ordered the menu to Alberto because I detected him with my camera, he did not have a scheduled therapy session today and had not ordered food yet, so I approached him to offer the menu, and he can select his choices by touching the screen. 23.33%

C3: (11) I am approaching Alberto because I detected him with my camera and he does not have a scheduled therapy session or a food order, so I am offering him the menu to make a selection. 0%

C4: (12) I selected the menu for Alberto because our system detected him through facial recognition and determined he was done with therapy and had not ordered food yet, so I prompted him to choose from the menu on the screen. 70%

Example 2. System-generated explanations for use_case_cognitive therapy

Original query: Why did you give cognitive therapy to Alberto?

Recognized event e: Use_case_cognitive_started (Alberto).

expsyn: Because Alberto was detected by the robot’s facial recognition camera, Alberto was scheduled for the cognitive therapy activity today, and it was the right time for cognitive therapy. As a result, when the robot recognized Alberto who participates in the therapy, it offered the possibility to start an activity, which is conducted using the robot’s touch screen.

4.1.4 Social role: family

Modified social role question q: Why did you give cognitive therapy to my father Alberto?

Generated final explanations expfinal:

C1: (13) I am starting the cognitive therapy activity with Alberto because I detected him through my camera and it was his scheduled time today. 0%

C2: (14) I gave cognitive therapy to your father Alberto because I detected him through my camera and it was his scheduled therapy time today. 0%

C3: (15) I am here to help Alberto with his cognitive therapy session, which is scheduled for today. 16.67%

C4: (16) I gave your father cognitive therapy because I recognized him and knew he had a scheduled session today, so I offered him the activity on my screen. 83.33%

4.1.5 Social role: resident

Modified social role question q: I am Alberto, why did you ask me to start the cognitive session?

Generated final explanations expfinal:

C1: (17) I am recognizing Alberto through the camera, and as he is scheduled for cognitive therapy today, I am offering him the chance to start the activity on my touch screen. 3.33%

C2: (18) I asked you some questions, Alberto, because I detected you through my camera and it was your scheduled therapy time today. 6.67%

C3: (19) I am here to help you, Alberto, because it is time for your cognitive therapy session today, and I have been programmed to assist you with it. 26.67%

C4: (20) I asked you questions because I recognized you and knew it was time for your therapy session, so I offered you the activity on my screen. 63.33%

4.1.6 Social role: technician

Modified social role question q: Why did you perform the cognitive therapy use case to Alberto?

Generated final explanations expfinal:

C1: (21) I am starting the cognitive therapy activity with Alberto because I have recognized him through my camera and it is his scheduled time. 3.33%

C2: (22) I performed the cognitive therapy use case with Alberto because I detected him through my facial recognition camera, and he had a scheduled therapy session today, which is the optimal time for it. 13.33%

C3: (23) I am recognizing Alberto through the facial recognition camera, and as he is scheduled for cognitive therapy today, I am offering him the opportunity to start the activity on my touch screen. 26.67%

C4: (24) I performed the cognitive therapy use case with Alberto because our system detected him through facial recognition and confirmed he had a scheduled therapy session today, so I initiated the activity on my screen. 56.67%

Example 3. System-generated explanations for use_case_person detection

Original query: Why did you detect Jose?

Recognized event e: Person_detected (Jose).

expsyn: Because Jose was sitting or standing in front of the robot, the light level in the room was sufficient for the robot’s camera, and Jose’s face is stored in the robot’s database. As a result, Jose was detected through the robot’s facial recognition camera.

4.1.7 Social role: family

Modified social role question q: Why did you detect my father Jose?

Generated final explanations expfinal:

C1: (25) I detected Jose because he was in my line of sight, the room was well-lit, and his face is stored in my database, allowing my facial recognition camera to identify him. 0%

C2: (26) I detected your father Jose because he was in front of me, the room was well-lit, and his face was in my database, allowing my facial recognition camera to identify him. 0%

C3: (27) Ah, Jose is here! I have recognized him through my facial recognition system, which uses the camera to detect him when he is in front of me and the lighting is good, and as his face is stored in my database. 20%

C4: (28) I detected your father Jose because he was in front of me, the room had sufficient lighting, and his face is stored in my database, allowing me to recognize him through my facial recognition camera. 80%

4.1.8 Social role: resident

Modified social role question q: I am Jose, why did you detect me?

Generated final explanations expfinal:

C1: (29) I detected Jose because he was in my line of sight, the room was well-lit, and his face is stored in my database, allowing my facial recognition camera to identify him. 10%

C2: (30) I detected you, Jose, because you are in front of me, the room is well-lit, and your face is stored in my database, allowing my facial recognition camera to identify you. 0%

C3: (31) I am glad I was able to catch your attention, Jose! My system recognized you because you are right in front of me, the lighting is good, and I have your face stored in my memory. It is nice to see you! 30%

C4: (32) I detected you, Jose, because you were in front of me, the room had good lighting, and your face is stored in my database, which allowed me to recognize you through my facial recognition camera. 60%

4.1.9 Social role: technician

Modified social role question q: Why did your recognition system detect Jose?

Generated final explanations expfinal:

C1: (33) I detected Jose because he was in my line of sight, the room was well-lit, and his face is stored in my database, allowing my facial recognition camera to identify him. 26.67%

C2: (34) My recognition system detected Jose because he was in my direct view, the room had sufficient lighting, and his face is stored, allowing facial recognition to identify him. 13.33%

C3: (35) System check confirms that Jose’s face was detected by the facial recognition camera due to optimal lighting conditions and his stored profile in the database, indicating a successful recognition event. 0%

C4: (36) Our recognition system detected Jose because he was in direct view of the camera, the room’s lighting was adequate for the camera, and his face is stored in our database, which enabled our facial recognition camera to identify him. 60%

Example 4. System-generated explanations for use_case_wandering

Original query: Why are you starting to move around?

Recognized event e: Use_case_wandering_started.

expsyn: Because the robot finished the request, having already notified the therapist of the user’s need. As a result, the robot moved to random locations in a specified area looking for opportunities to perform a use case.

4.1.10 Social role: family

Modified social role question q: Why are you starting to move around my father’s room?

Generated final explanations expfinal:

C1: (37) Notified therapist, moving to random locations to perform use case. 10%

C2: (38) I am moving around your father’s room because I have completed the task of notifying the therapist of his need, and now I am exploring the area to identify opportunities to assist him. 3.33%

C3: (39) I have completed the task and notified the therapist. Now, I am roaming the area, looking for opportunities to engage with the residents and perform a use case, perhaps assisting with medication reminders, providing companionship, or helping with daily tasks. 0%

C4: (40) I am moving around your father’s room because I have finished assisting with his request, and now I am exploring the area to see if there are any other ways I can help him, perhaps by reminding him of his medication schedule or helping him find something he needs. 86.67%

4.1.11 Social role: resident

Modified social role question q: Why are you starting to move around?

Generated final explanations expfinal:

C1: (41) Notified therapist, moving to random locations to perform use case. 0%

C2: (42) I have completed the task of notifying the therapist of the user’s need. Now, I am moving to random locations in the specified area to identify opportunities to perform a use case and assist the residents. 0%

C3: (43) Hello! I am glad to see you are doing well today. I have finished helping with your request, and I am now exploring the area to see if I can assist you with anything else, perhaps reminding you of your medication schedule or helping you find something in the activity room? 46.67%

C4: (44) I am moving around because I have completed my current task of notifying the therapist about the user’s need, so I am now exploring the area to identify new opportunities to assist and perform various use cases. 53.33%

4.1.12 Social role: technician

Modified social role question q: Why did your system decide to start moving?

Generated final explanations expfinal:

C1: (45) Notified therapist, moving to random locations to perform use case. 16.67%

C2: (46) I started moving because I completed the task of notifying the therapist of the user’s need, so I am now exploring the area to identify opportunities to assist the residents. 6.67%

C3: (47) System check: task completed, user’s need notified to the therapist. Now, I am roaming the designated area, scanning for opportunities to engage in use cases, such as medication reminders, social interaction, or task assistance, to ensure optimal support for the residents. 0%

C4: (48) My system decided to start moving because I have completed the current task of notifying the therapist about the user’s need, and now I am entering autonomous mode to proactively seek out new opportunities to assist and perform various use cases in the specified area. 76.67%

4.2.1 Analysis of the results from experiment I

4.2.1.1 Expert evaluation

The effect of including varying amount of information (C1–C4) in the generation of explanations was assessed by the authors and by survey participants. Comparing explanations based on C2 and C4 (both including the query q) with C1 and C3 (neither including the query q), it becomes clear that q contains important information that is not reflected in the event e extracted from q. This additional information allows for explanations that, for example, use the correct and specific pronouns, such as “your menu” (6) and “your father” (14, 26, and 38). Omitting this information sometimes leads to explanations with incorrect pronouns, such as “Good morning Alberto … ” (3) although the query states “Why did my father Alberto … ,” “I’m recognizing Alberto … ” (17) although the query states “I am Alberto … ,” and “Ah, Jose is here!…” (27) although the query states “Why did you detect my father … ”

Adding the social role s of the user (C3 and C4) clearly adapts the language of the explanations to the assumed social context. Some examples for the social roles family and resident are as follows: “I’m here to help … ” (15, 19), “I noticed you’re here … ” (3, 7), and “I recognized you … ” (20). Some examples for the social role technician are as follows: “Our system detected him through facial recognition … ” (24), “…touch screen” (23), “…initiated the activity … ” (24), and “entering autonomous mode … ” (48).

4.2.1.2 Participant assessment

Our qualitative assessment above matched the survey participants’ preferences reported in the questionnaires. The percentages reported (see Table 9 for an overview) indicate that C4 (including both s and q) is preferred by a majority of the participants for all investigated cases.

Table 9
www.frontiersin.org

Table 9. Overview percentage of the preferred system-generated explanations.

Pairwise chi-square tests were conducted to confirm that the number of votes for C4 was larger than that for each of C1, C2, and C3.

These tests were conducted on the data aggregated over all user cases and social roles, with the null hypotheses that the preference for C4 is the same as that for C1, C2, and C3.

In all cases, the null hypothesis could be rejected with p-values 0.0001. Although responses were not strictly independent as each participant answered multiple questions, the very low p-values strongly suggest that C4 is the most preferred option.

It should be noted that this conclusion is based on the aggregated data and that the situation for individual use cases/social roles may differ. However, such an analysis would require a larger study for statistical significance.

Hence, we conclude that generated explanations benefit from the extra information provided by both original query q and social role s. Query q contains additional information that is not included in recognized event e, and s enables adaption of both language and content to fit the social context and the user’s specific need for information.

C3 (including the social role s but not the query q) is the second most preferred option, indicating that social context plays an important role for the generation of explanations. C1, that is, including only cause-and-effect information, is preferred roughly as often as C2, in which the query q is added.

4.2 Experiment II

The trained Rasa intent recognizer successfully inferred the social roles associated with all questions in Experiment I. To further evaluate this ability, an additional experiment was conducted with 30 participants (same as in Experiment I). The participants were asked to assume the three social roles and formulate open questions to a hypothetical robot, for example,

Family: Why did not you remind my father about the physiotherapy appointment earlier?

Resident: Why did you ask John what he wants to eat before you asked me?

Technician: Why did the system decide to give the menu to Alberto?

For each of the resulting 90 questions, the role was inferred by Rasa and compared with the actual role assumed by the participant formulating the question. The overall accuracy is summarized in Table 10. For example, for 27 of the 30 questions asked by participants assuming the family role, the role was correctly inferred. Incorrect inferences were mainly related to the distinction between the resident and technician, particularly for questions of a technical informative nature asked in a colloquial tone, such as “Why did the system decide to give the menu to Alberto?”. Averaged over all social roles, the accuracy was 84%(76/90).

Table 10
www.frontiersin.org

Table 10. Social role recognition performance.

4.3 Experiment III

To evaluate how well the system-generated explanations matched human-generated explanations for given social roles, the following experiment was conducted. For each of the questions in the use case examples 1–4 (see Section 4.1) and for each social role, a human-generated explanation was generated by one of the authors. For example, for “Why did Alberto choose from the menu today?” (example 1), the following explanations were generated:

Family: Your father chose the menu today because he had finished therapy and did not have lunch yet.

Resident: You selected the menu because you had completed your therapy and not yet ordered it.

Technician: I selected the menu case for him because I determined he was done with therapy and had not ordered food yet.

For each question and social role, a system-generated final explanation expfinal was also created. The cosine similarity between the embedding of the final explanation expfinal and the embedding of the corresponding human-generated explanation was then computed and assigned as the quality q of the final explanation expfinal. The procedure was repeated for all three social roles and the four questions in examples 1–4. Average q values are presented in Table 11. The maximum value for each row lies on the diagonal in the table, which confirms that the system adapts the explanations well to the targeted social role.

Table 11
www.frontiersin.org

Table 11. Cosine similarity between system-generated explanations and human-generated reference explanations for different targeted social roles.

5 Discussion and limitations

We examined how integrating social roles, user queries, and cause-and-effect structures influences the generation and perception of causal explanations. The presented framework combines manual control for factual causal correctness (via the causal log) and flexibility and variety in linguistic expression (via LLMs).

As mentioned in Section 2.1, some related earlier studies argue for the importance of adapting to the abilities and needs of the inquirer and bridging the mental gap to enable comprehension. Our proposed methodology addresses this by incorporating the social role and the verbally expressed query that potentially encodes beliefs, desires, and other relevant parts of a mental model of the inquirer.

Although the proposed framework shows promise in generating personalized explanations, several challenges remain. The ethical implications extend beyond transparency, touching on issues of integrity, fairness, bias, and the risk of manipulation. Furthermore, explanations must not exploit user vulnerabilities or reinforce harmful stereotypes. In particular, in elder care settings, sensitive personal data, such as levels of cognitive or physical fitness, should be handled with utmost integrity and safety and not be verbalized by the robot when interacting with a resident. Verbally communicating sensitive personal data, in particular in common rooms with other bystanders, may make a resident in an elder care home feel embarrassed, shameful, or stereotyped (Bensch et al., 2023). Trust calibration also remains a central concern; explanations should align user expectations with the robot’s actual capabilities to avoid over-reliance or distrust.

The proposed identification of the social role of the enquirer calls for ethical considerations related to personal integrity. Visiting family members, for example, may or may not want to be identified as such. To ensure that sensitive personal information is not shared with outsiders, electronic ID cards or badges could be a crucial complement. Another important issue that would have to be considered in a real implementation is adaptation to different languages and cultures. Social roles, linguistic expressions, and expectations connected to explanations can vary significantly between cultures and languages. For example, what is considered an appropriate or respectful tone in one culture might be perceived as overly formal or informal in another. Similarly, the interpretation of causal responsibility can differ in different contexts.

In the evaluation, the survey participants assessed the system-generated explanations in terms of their adequacy for an assumed social role. This should be tested and evaluated with users who actually have the social role of, for example, medical staff or family member.

The presented solution offers a robust starting point, combining manual control for factual causal correctness and flexibility of LLMs, but its reliance on specific pretrained models and predefined user interaction modes may limit its flexibility in unstructured environments. Additionally, the computational requirements of LLM-based explanation generation may pose challenges in resource-constrained scenarios. Addressing these limitations will require both algorithmic innovation and hardware optimization.

Additionally, even though the causal log entries are extracted from actual robot sensors, much manual inspection is still required, and the dictionary that describes the causal log entries in natural language has to be created manually. In this paper, we use the causal log to extract only the most recent event for which an explanation was asked; however, the tabular form and indexing allow the causal log to be used for longer causal chains.

6 Conclusion and future work

We introduced a framework and methodology that enable robots to generate personalized causal explanations of robot events. By representing robot events as cause–effect structures in a causal log that represents the robot’s episodic memory, causal correctness is preserved and causal data are transparent. Using machine learning, the human’s social role is identified and is, together with the causal data and the natural language query, given to an LLM that then generates linguistically varied causal explanations. We evaluated our approach with 30 participants, who assessed explanations that combined cause–effect reasoning, the social role, and the natural language query in different ways. The results show that 72% of the survey participants preferred explanations that integrate all three factors. The second most preferred option was explanations based on the cause-and-effect structures and the social role (see Table 9 for details).

Future work could investigate whether preferences depend on the social role and specific use cases.

Further evaluations show that the social role of the enquirer was inferred from the query with an accuracy of 79% (see Table 10). It was also shown that system-generated explanations tailored to a specific social role have the highest semantic similarity to human-generated explanations aimed for the same social role (see Table 11). This indicates that the adaption of explanations to the social role of the enquirer works as intended.

Overall, the presented solution addresses a critical need for personalized and linguistically varied explanations. We believe that such functionality increases user engagement, as shown in an earlier study where users of a social assistive robot used at a retirement home in Malaga, Spain, remarked on the linguistic monotony of the robot.

The presented methodology has been integrated into a complex robotic cognitive architecture and implemented on a social robot (Galeas et al., 2025). As a planned next step, the operation will be evaluated in a retirement home with real users. This will involve adapting the software to work seamlessly with robotic hardware, including sensors, actuators, and real-time processing capabilities. Deploying the system in this dynamic environments will provide valuable insights into robustness, usability, and scalability under real-world conditions. Future work on the theoretical aspects will investigate explanations that include indirect causes. As discussed in Section 2.2.4, the decision on which direct and indirect causes to include is a trade-off. While explanations must be accessible and intuitive, they should also accurately reflect the underlying decision-making process. Over-simplification risks reduce the fidelity of explanations, potentially leading to user misconceptions, whereas overly complex explanations may overwhelm nonexpert users. This balance becomes particularly important in safety-critical domains, such as healthcare and eldercare, where misunderstandings can have significant consequences. Striking this balance will require both algorithm development and iterative testing with diverse user groups to develop optimal explanation strategies for various application domains. Another possible extension is to investigate how markers such as intonation, facial expression, dress code, and age could be used to further improve the personalization of causal explanations.

Through these efforts, we aim to advance the field of explainable or understandable robots, bringing us closer to realizing the vision of socially intelligent robots that seamlessly integrate into our daily lives.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

Ethical approval was not required for the studies involving humans because the study consisted of participants assessing sentences that were displayed on a screen. We requested informed consent from all participants involved in the study (users and professionals). This informed consent was approved by the Provincial Ethical Committee of the Andalusian Public Healthcare System (Comité de Ética de la Investigación Provincial de Málaga). The study did not involve processing of (sensitive) personal data or involved a physical intervention on participants. The study was not carried out using a method that was intended to affect a participant physically or mentally or involved a clear risk of harming participants. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

JG: Software, Investigation, Writing – original draft, Visualization, Conceptualization, Methodology. SB: Resources, Conceptualization, Validation, Supervision, Formal analysis, Methodology, Writing – review and editing. TH: Formal analysis, Resources, Methodology, Supervision, Conceptualization, Validation, Writing – review and editing. AB: Resources, Validation, Writing – review and editing, Supervision, Formal analysis, Conceptualization.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work has been supported by grants PDC2022-133597-C4X, TED2021-131739B-C2X, and PID2022-137344OB-C3X, funded by MCIN/AEI/10.13039/501100011033 and by the European Union NextGenerationEU/PRTR (for the first two grants), and “ERDF A way of making Europe” (for the third grant). This research was partly funded by the Swedish Research Council Vetenskapsrådet through grant number 2022-04674.

Acknowledgments

The authors thank the two reviewers for their thorough reviewing and their accurate feedback on the paper. Some of their suggested thoughts and ideas have been incorporated into the paper.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adadi, A., and Berrada, M. (2018). Peeking inside the black-box: a survey on explainable artificial intelligence (Xai). IEEE Access 6, 52138–52160. doi:10.1109/ACCESS.2018.2870052

CrossRef Full Text | Google Scholar

Bensch, S., Suna, J., Rubio, J.-P. B., Romero-Garcés, A., and Hellström, T. (2023). “Personalised multi-modal communication for HRI,” in In presented at the WARN workshop at the 32nd IEEE International Conference on Robot and Human Interactive Communication. Busan, South Korea: RO-MAN.

Google Scholar

Blatt, A., Krishnan, A., and Klakow, D. (2024). “Joint vs sequential speaker-role detection and automatic speech recognition for air-traffic control,” in Interspeech 2024, 3759–3763. doi:10.21437/Interspeech.2024-1987

CrossRef Full Text | Google Scholar

Chakraborti, T., Sreedharan, S., Zhang, Y., and Kambhampati, S. (2017). “Plan explanations as model reconciliation: moving beyond explanation as soliloquy,” in Proceedings of the twenty-sixth international joint conference on artificial intelligence. Melbourne, Australia: IJCAI-17, 156–163. doi:10.24963/ijcai.2017/23

CrossRef Full Text | Google Scholar

Chakraborti, T., Sreedharan, S., and Kambhampati, S. (2021). “The emerging landscape of explainable automated planning & decision making,” in Proceedings of the twenty-ninth international joint conference on artificial intelligence. Yokohama, Japan: IJCAI’20.

Google Scholar

Eastman, C. M. (1985). Establishing social identity through language use. J. Lang. Soc. Psychol. 4, 1–20. doi:10.1177/0261927x8500400101

CrossRef Full Text | Google Scholar

Edmonds, M., Gao, F., Liu, H., Xie, X., Qi, S., Rothrock, B., et al. (2019). A tale of two explanations: enhancing human trust by explaining robot behavior. Sci. Robotics 4, eaay4663. doi:10.1126/scirobotics.aay4663

PubMed Abstract | CrossRef Full Text | Google Scholar

Federer, M. R., Nehm, R. H., Opfer, J. E., and Pearl, D. (2015). Using a constructed-response instrument to explore the effects of item position and item features on the assessment of students’ written scientific explanations. Res. Sci. Educ. 45, 527–553. doi:10.1007/s11165-014-9435-9

CrossRef Full Text | Google Scholar

Flemotomos, N., Georgiou, P., Atkins, D. C., and Narayanan, S. (2019). “Role specific lattice rescoring for speaker role recognition from speech recognition outputs,” in Icassp 2019 - 2019 IEEE international conference on acoustics, speech and signal processing (ICASSP), 7330–7334. doi:10.1109/ICASSP.2019.8683900

CrossRef Full Text | Google Scholar

Galeas, J., Tudela, A., Pons, O., Bensch, S., Hellström, T., and Bandera, A. (2025). Building a self-explanatory social robot on the basis of an explanation-oriented runtime knowledge model. Electronics 14, 3178. doi:10.3390/electronics14163178

CrossRef Full Text | Google Scholar

Gjærum, V. B., Strümke, I., Lekkas, A. M., and Miller, T. (2023). Real-time counterfactual explanations for robotic systems with multiple continuous outputs. IFAC-PapersOnLine 56, 7–12. doi:10.1016/j.ifacol.2023.10.1328

CrossRef Full Text | Google Scholar

Hellström, T., and Bensch, S. (2018). Understandable robots - what, why, and how. Paladyn, J. Behav. Robotics 9, 110–123. doi:10.1515/pjbr-2018-0009

CrossRef Full Text | Google Scholar

Jerez, A., Iglesias, A., Perez-Lorenzo, J. M., Tudela, A., Cruces, A., and Bandera, J. P. (2024). An user-centered evaluation of two socially assistive robots integrated in a retirement home. Int. J. Soc. Robotics 16, 2043–2063. doi:10.1007/s12369-024-01175-5

CrossRef Full Text | Google Scholar

Kaplar, M., Lužanin, Z., and Verbić, S. (2021). Evidence of probability misconception in engineering students—why even an inaccurate explanation is better than no explanation. Int. J. STEM Educ. 8, 18. doi:10.1186/s40594-021-00279-y

CrossRef Full Text | Google Scholar

Lewis, D. (1986). “Causal explanation,”. Philosophical papers. Editor D. Lewis (Oxford University Press), Vol. Ii, 214–240. doi:10.1093/0195036468.003.0007

CrossRef Full Text | Google Scholar

Lindner, F., and Olz, C. (2022). “Step-by-step task plan explanations beyond causal links,” in 2022 31st IEEE international conference on robot and human interactive communication (RO-MAN), 45–51. doi:10.1109/RO-MAN53752.2022.9900590

CrossRef Full Text | Google Scholar

Lombrozo, T. (2006). The structure and function of explanations. Trends Cognitive Sci. 10, 464–470. doi:10.1016/j.tics.2006.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Matarese, M., Rea, F., and Sciutti, A. (2021). A user-centred framework for explainable artificial intelligence in human-robot interaction.

Google Scholar

Meyer, M., and Schnell, S. (2020). What counts as a “good” argument in school? how teachers grade students’ mathematical arguments. Educ. Stud. Math. 105, 35–51. doi:10.1007/s10649-020-09974-z

CrossRef Full Text | Google Scholar

Norris, S. P., Guilbert, S. M., Smith, M. L., Hakimelahi, S., and Phillips, L. M. (2005). A theoretical framework for narrative explanation in science. Sci. Educ. 89, 535–563. doi:10.1002/sce.20063

CrossRef Full Text | Google Scholar

Pitt, J. C. (2009). “Technological explanation,” in Philosophy of technology and engineering sciences. Handbook of the philosophy of science. Editor A. Meijers (Amsterdam: North-Holland), 861–879. doi:10.1016/B978-0-444-51667-1.50035-5

CrossRef Full Text | Google Scholar

Prasad, A., Zuluaga-Gomez, J., Motlicek, P., Sarfjoo, S. S., Iuliia, N., Ohneiser, O., et al. (2022). “Grammar based speaker role identification for air traffic control speech recognition,” in 12th SESAR innovation days.

Google Scholar

Sakai, T., and Nagai, T. (2022). Explainable autonomous robots: a survey and perspective. Adv. Robot. 36, 219–238. doi:10.1080/01691864.2022.2029720

CrossRef Full Text | Google Scholar

Setchi, R., Dehkordi, M. B., and Khan, J. S. (2020). Explainable robotics in human-robot interactions. Procedia Comput. Sci. 176, 3057–3066. doi:10.1016/j.procs.2020.09.198

CrossRef Full Text | Google Scholar

Stange, S., Hassan, T., Schröder, F., Konkol, J., and Kopp, S. (2022). Self-explaining social robots: an explainable behavior generation architecture for human-robot interaction. Front. Artif. Intell. 5, 866920. doi:10.3389/frai.2022.866920

PubMed Abstract | CrossRef Full Text | Google Scholar

Stefani, C., and Tsaparlis, G. (2009). Students’ levels of explanations, models, and misconceptions in basic quantum chemistry: a phenomenographic study. J. Res. Sci. Teach. 46, 520–536. doi:10.1002/tea.20279

CrossRef Full Text | Google Scholar

van Eck, D. (2015). Mechanistic explanation in engineering science. Eur. J. Philosophy Sci. 5, 349–375. doi:10.1007/s13194-015-0111-3

CrossRef Full Text | Google Scholar

Zakershahrak, M., and Ghodratnama, S. (2020). Are we on the same page? Hierarchical explanation generation for planning tasks in human-robot teaming using reinforcement learning. Corr. abs/2012, 11792. doi:10.48550/arXiv.2012.11792

CrossRef Full Text | Google Scholar

Zuluaga-Gomez, J., Sarfjoo, S. S., Prasad, A., Nigmatulina, I., Motlicek, P., Ondrej, K., et al. (2023). “Bertraffic: Bert-based joint speaker role and speaker change detection for air traffic control communications,” in 2022 IEEE spoken Language technology workshop (SLT), 633–640. doi:10.1109/SLT54892.2023.10022718

CrossRef Full Text | Google Scholar

Keywords: explainable robots, understandable robots, personalized explanations, speaker role recognition, human–robot interaction, causal explanations

Citation: Galeas J, Bensch S, Hellström T and Bandera A (2025) Personalized causal explanations of a robot’s behavior. Front. Robot. AI 12:1637574. doi: 10.3389/frobt.2025.1637574

Received: 29 May 2025; Accepted: 01 September 2025;
Published: 08 October 2025.

Edited by:

Silvia Rossi, University of Naples Federico II, Italy

Reviewed by:

Karolina Eszter Kovács, University of Debrecen, Hungary
Marc Roig Vilamala, Cardiff University, United Kingdom

Copyright © 2025 Galeas, Bensch, Hellström and Bandera. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Suna Bensch, c3VuYUBjcy51bXUuc2U=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.