- 1CHI – Chair of Health Informatics, TUM University Hospital, Munich, Germany
- 2MCML – Munich Center for Machine Learning, Munich, Germany
- 3GLAM – Group on Language, Audio, & Music, Imperial College London, London, United Kingdom
- 4MDSI – Munich Data Science Institute, Munich, Germany
Digital mental health interventions (DMHIs) have become increasingly prominent as scalable solutions to address global mental health needs. However, many existing tools lack the emotional sensitivity required to foster meaningful engagement and therapeutic effectiveness. Affective computing, a field focused on designing systems capable of detecting and responding to human emotions, offers promising advancements to the emotional responsiveness of these digital interventions. This narrative review examines how affective computing methods such as emotion recognition, sentiment analysis, emotion synthesis, and audiovisual and physiological signal processing, are being integrated into DMHIs to enhance user engagement and improve clinical outcomes. The findings suggest that emotionally adaptive systems can strengthen user engagement, simulate empathy, and support more personalized care. Early studies indicate potential benefits in terms of symptom reduction and user satisfaction, though clinical validation remains limited. Challenges such as algorithmic bias, privacy concerns, and the need for ethical design frameworks continue to shape the development of this emerging field. By synthesizing current trends, technological advancements, and ethical considerations, this review highlights the potential of affective computing in digital mental health and identifies key directions for future research and implementation.
1 Introduction
Affective computing, which is considered as the development of computational systems that detect, interpret, and respond to human emotions (1), has rapidly advanced digital mental health (2–4). Emotion-aware technologies are increasingly used to personalize care, enhance engagement, and improve outcomes across psychological conditions. Digital mental health interventions (DMHIs), which encompass technology-based tools such as mobile apps, wearable sensors, and conversational agents, aim to assess, monitor, or treat mental health disorders (5, 6). When integrated with affective computing, these systems gain the ability to dynamically adjust to users’ emotional states, offering more responsive and individualized support (7). From stress-sensing wearables (8) to sentiment-aware chatbots (9), this fusion reshapes human–technology interaction, particularly in clinical contexts where emotional understanding is critical to therapeutic success.
Despite its growing importance, comprehensive reviews focused specifically on affective computing within digital mental health remain scarce. Existing literature tends to either survey affective computing broadly without focusing on psychological interventions (10–12), or examine DMHIs without considering the application or the underlying mechanisms of affective computing (13–15). As a result, the unique opportunities, challenges, and ethical implications of integrating affective computing into mental health care remain underexplored.
This narrative review addresses that gap by critically examining emotionally responsive DMHIs—systems that extend beyond emotion detection to adapt their behavior in real time based on users’ affective states. We present several use-cases targeted directly at patients, but also at supporting clinicians. The focus is on monitoring systems and AI-driven chatbots, while evaluating their technical foundations, clinical evaluation, and level of emotional reactivity.
By centering on emotionally adaptive systems, we aim to bridge the gap between computational emotion modeling and the practical realities of digital mental health care. We also highlight ethical considerations, including transparency, emotional manipulation, privacy, and cultural validity. Ultimately, this review offers an interdisciplinary roadmap for researchers and clinicians seeking to responsibly apply affective computing to improve digital mental health outcomes.
2 Methods
To investigate annual publication trends and describe the growth of research at the intersection of affective computing and mental health, a structured literature search was conducted using the Web of Science Core Collection and Scopus databases; with 2,106 and 3,353 results respectively. Query terms were applied to titles, abstracts, and keywords. The search period spans from 1997, when Rosalind Picard first introduced the term affective computing (1), to the end of 2024. To focus on computational models designed to train machines in emotion recognition, rather than on studies aimed at understanding human emotion recognition abilities, we restricted the subject areas to Computer Science and Engineering. Duplicate records (2,569) were identified and removed using the systematic review tool Rayyan (16). The search strategy is summarized in Table 1.

Table 1. Search strategy for bibliometric analysis of publications at the intersection of affective computing and mental health.
3 Background
To highlight the growing applicability and scientific interest in affective computing within digital mental health, a comprehensive overview of its evolution is provided, followed by an examination of its key components: emotion modeling, sensing, and adaptation. This foundation offers the necessary context for understanding current challenges and emerging trends in affective computing-based technologies for DMHIs.
3.1 The evolution of affective computing in digital mental health
To provide an illustrative overview of the rapidly increasing trend in annual scientific production concerning affective computing and mental health, the progression of publications is shown in Figure 1. Prior to 2010, the number of publications was sparse, reflecting the early theoretical phase of affective computing, which had yet to find concrete applications in mental health. This first stage (1997–2010) was summarized by Picard’s seminal work, which formally defined affective computing as the study of systems capable of recognizing, interpreting, expressing, and regulating human emotions (1). During this phase, research focused primarily on human–computer interaction, with Ekman’s theory of six basic emotions (17) serving as the primary labeling framework. Early studies explored various overt modalities: Lyons et al. (18) applied Gabor wavelets to facial image analysis; Schuller et al. (19) employed prosodic and acoustic features to classify speech-based emotion; and Pang et al. (20) introduced machine learning for sentiment polarity detection in textual film reviews to give but a few examples. The Kismet robot developed at the MIT Media Lab (21) integrated facial motor expressions and vocal prosody, demonstrating a prototype of “affective dialogue” between humans and machines. While foundational for validating whether machines could perceive emotional cues, these efforts remained largely confined to controlled environments and healthy populations, with limited integration into real-world mental health contexts or deeper affective state modeling.

Figure 1. Annual publication counts from Web of Science and Scopus illustrating the growth of research at the intersection of affective computing and mental health from 1997 to 2024.
The second stage (2011–2017) marks a shift from theoretical exploration to practical application. In 2010, the launch of the IEEE Transactions on Affective Computing signified the formal recognition of affective computing as an independent research domain comprising multiple subfields (22). This milestone elevated the academic standing of the field and laid the groundwork for its expansion into healthcare and mental well-being. This is reflected in a steady increase in publications over these years. On the technical front, the proliferation of mobile devices and wearable sensors provided a viable data foundation for affective computing in digital mental health. For example, devices like the Empatica E4 enabled continuous capture of physiological signals such as electrodermal activity and photoplethysmography (23). Similarly, the Affectiva SDK facilitated real-time facial expression analysis on mobile platforms, expanding the capacity for in-situation emotion sensing (24). These technologies led to the development of several representative datasets, including DAIC-WOZ for assessing anxiety, depression, and PTSD symptoms (25); AVEC 2016 for multimodal depression risk modeling (26); and SWELL-KW for stress recognition and contextual behavior analysis (27). Based on these datasets, a number of empirical studies related to mental health have been conducted (28–30).
This period also benefited from increased accessibility to traditional machine learning techniques. Models based on support vector machines, random forests, and k-nearest neighbors were widely used for early affect classification across speech, text, and physiological modalities (31). However, these approaches typically relied on static feature representations and lacked the capacity to model temporal dynamics, contextual dependencies, or individual variability in affective responses. As noted by D’Mello (32), affective systems that do not account for emotion trajectories and situational antecedents fall short of the needs within mental health contexts. This limitation underscored a broader debate within the field: whether affective computing should remain focused on perceptual signal classification or evolve toward deeper emotional intelligence, involving causal modeling and semantic understanding. Since 2018, affective computing has entered a new phase in digital mental health, driven by the adoption of deep learning. As these methods matured, various neural architectures, including convolutional neural networks, recurrent neural networks, and transformers, have increasingly been applied to recognizing emotion-related mental health conditions such as depression, anxiety, and stress, as well as to clinical alignment tasks. Rejaibi et al. (33) employed MFCC features and a recurrent neural network on the DAIC-WOZ depression detection task. Their model achieved state-of-the-art performance at the time, demonstrating the practical value of acoustic features for predicting emotional severity. Ray et al. (34) proposed a multi-layer attention network for AVEC 2019, using convolutional neural networks to extract audio features and LSTMs for textual encoding. A multi-level attention mechanism fused audio, visual, and textual inputs, achieving notable improvements over the original baselines in depression regression. Wu et al. (35) developed a transformer-based self-supervised learning framework for emotion recognition using wearable signals. Their results demonstrated superior performance in classifying stress-related states compared to fully supervised methods, with improved generalizability.
More recently, transformer and then large language models (LLMs) have begun to play a role in affective computing for digital mental health (36). In social media contexts, Bucur et al. (37) introduced a time-enhanced multimodal transformer model that integrates CLIP image embeddings with EmoBERTa textual representations and applies time2vec encoding to model posting intervals, aiming to predict user-level depression risk. The model achieved state-of-the-art F1 scores of 0.931 on a popular multimodal Twitter corpus and 0.902 on the MultiRedditDep dataset. Additionally, LLMs have been used for interactive emotional intervention (3). These technological advancements, along with increased recognition of mental health due to the COVID-19 pandemic, explain the exponential growth in scientific output after 2018.
In sum, the evolution of affective computing in digital mental health reflects not only algorithmic progress but also the beginning of a broader shift from recognizing surface-level emotional cues to aspiring to understand their generative mechanisms and enabling adaptive interventions. While much of the field still relies on direct mappings between (multimodal) cues and emotion states, the increasing integration of context, temporality, and personalized dynamics suggests a transition toward richer, semantically informed affective computing; moving from the perceptual layer (“What you feel”), through the modeling layer (“Why you feel”), toward the intervention layer (“What should be done”). Therefore, affective computing now plays an increasingly central role in the development of digital mental health closed-loop systems, which are cybernetic frameworks that continuously sense mood, update personalized models, and deliver adaptive interventions in real time (38, 39). Nonetheless, realizing this loop still presents substantial theoretical and methodological challenges.
3.2 Technical key components of affective computing
We complement existing works, such as the systematic review by Wang et al. (12) and the survey by Afzal et al. (40), which comprehensively cover the technical properties and quantitative performance of key affective computing methods across various modalities like voice and text, by introducing affective computing from a human–computer interaction perspective.
3.2.1 Emotion modeling
Emotion modeling provides the theoretical foundation for affective computing systems to understand human psychological states. Current approaches are primarily grounded in two major emotion theories. Basic emotion theory categorizes emotions into discrete classes, such as Ekman’s six universal emotions (17). Due to its clear label structure and low annotation cost, this framework has been widely adopted in classification tasks across speech, image, and text modalities. It is also integrated into many mental health monitoring systems. However, the theory originates from Western emotional expressions and fails to capture the diversity of emotional signals in clinical or cross-group contexts: Pampouchidou et al. (41) found notable performance differences across gender subgroups in automatic depression detection using discrete labels; Alghowinem et al. (42) further showed that models trained on single-cultural datasets underperform in cross-cultural settings. This can be explained by the currently dominating opinion that emotions are not entirely universal across cultures, but can be perceived differently (43). However, basic emotion theory does not account for such variability. Additionally, there is strong evidence that Ekman’s six emotions do not sufficiently capture the complexity of facial expression (44–46), highlighting the limitations of this framework.
Alternatively, the dimensional emotion theory following Wilhelm Wundt’s foundational work represents emotional states along three axes (valence and arousal, which are popularly used) and sometimes a third axis of dominance (47). This framework is widely used in subfields of affective computing such as speech emotion recognition (48) and text-based emotion recognition (49), where it serves as a regression target in deep learning models. In digital mental health, several studies have extended this approach to clinical emotion modeling. For instance, Ahmed et al. (50) used wearable physiological data (e.g., electrodermal activity, heart rate variability) to predict both depression severity and valence—arousal scores. Their tri-modal model reached high accuracy and F1 score performance for severity classification and valence detection, demonstrating the feasibility of dimensional emotion tracking, though explainability and annotation costs remain problematic.
Beyond these two mainstream models, compound emotion theory was proposed by Du et al. (51). It argues that humans express not only basic emotions but also blends of them, such as “joy–surprise” or “anger–surprise.” Their team identified 21 compound emotions using the Facial Action Coding System (FACS), showing distinct muscular patterns and high inter-rater agreement. While empirical applications in mental health remain limited, compound emotions offer a more fine-grained representational framework for clinical emotion monitoring. Nevertheless, a certain facial movement can express more than one emotional category and conveys besides emotional further social information (52). Similarly, the psychological constructionist theory proposes that emotions are made of several components that are not specific to emotion such as sensory stimulation (53). These different perspectives show that modeling emotions is far from trivial and can contain additional layers that are currently rarely considered in affective computing. Therefore, emotion modeling in digital mental health must carefully consider which psychological framework is most appropriate in the given setting and try to balance label interpretability, expressive granularity, and cross-cultural adaptability.
In current digital mental health applications, emotion modeling typically serves as a theoretical and representational backbone rather than a direct supervision signal. Most clinical systems rely on standardized psychological scales as primary labels or evaluation endpoints, such as PHQ-9 for depression, GAD-7 for anxiety, and DSM-5 for PTSD (54). These instruments quantify persistent symptom severity, whereas emotion models capture momentary affective dynamics. Some studies have begun to explore how emotional labels can complement or predict changes in longer-term symptom scores (55, 56). This dual-layer structure, combining real-time emotion inference with validated clinical assessment tools, offers a promising direction for personalized, longitudinal mental health monitoring.
3.2.2 Emotion sensing and recognition
Emotion sensing and recognition constitutes a central component of affective computing based on emotion modeling. It aims to objectively identify and quantify individuals’ emotional states through multimodal signals. Facial expression analysis today typically employs deep neural network architectures, increasingly enhanced by self-supervised pretraining methods to extract salient static and dynamic features (57). Speech emotion recognition has evolved from traditional acoustic representations (e.g., MFCCs, mel-spectrograms) combined with deep models to capture local temporal patterns, to the use of pretrained frameworks such as wav2vec 2.0 for richer contextual modeling (58, 59).
Physiological emotion modeling leverages biosignals such as electrodermal activity and photoplethysmography, with deep learning techniques used to construct personalized affective representations (60). Text-based emotion recognition has also advanced significantly. Early approaches relied on handcrafted features such as bag-of-words, TF-IDF, and sentiment lexicons. More recent models based on pretrained language transformers (e.g., BERT, LLaMA) have achieved superior performance by capturing semantic ambiguity and emotional metaphor through contextual learning and prompt-based adaptation (61).
Additionally, multimodal systems integrate visual, vocal, physiological, and textual cues, often employing cross-attention or alignment mechanisms to achieve semantic fusion (62). Empirical studies suggest that multimodal approaches generally outperform unimodal models, showing promise in modeling complex psychological states and facilitating more precise affective alignment (63).
Despite these advances, the generalizability of emotion sensing techniques to clinical settings remains limited. For example, a cross-corpus study found that emotion recognition models trained exclusively on data from healthy individuals performed at or below chance level when applied to clinical corpora such as DAIC-WOZ. Minimal fine-tuning on target data was necessary to restore performance, underscoring the presence of strong distributional shifts between source and target domains (42).
Further studies reveal that environmental variables, such as background noise and room reverberation, can reduce unweighted recall by up to 20%, indicating limited robustness in real-world deployment (64). These results suggest that general-purpose models cannot be reliably applied to populations with conditions such as anxiety, depression, or PTSD, where emotional expression, linguistic style, and physiological responses often diverge markedly from those of healthy individuals. Moreover, pharmacological interventions may further alter these manifestations (65), potentially rendering standard annotation schemes inadequate or biased. To address these challenges, it is necessary to adopt strategies such as few-shot adaptation, domain transfer, and subject-specific baseline modeling, at both the representation and label-design levels, to improve robustness, fairness, and interpretability in clinical applications.
3.2.3 Affective adaptation
Affective adaptation, also referred to as emotional responsiveness, describes the ability of affective computing systems to dynamically adjust their behavior or responses in real time based on a user’s emotional state and context. Unlike passive emotion recognition, affective adaptation enables personalized and contextually relevant interactions. This is typically achieved through techniques such as domain adaptation, few-shot and/or fine-tuning, or individual baseline calibration (66, 67).
Emotionally adaptive virtual agents such as Woebot dynamically adjust their dialogue strategies to deliver empathetic and psychologically safe interactions (68). These systems demonstrate that affective adaptation enhances not only model performance but also user engagement, trust, and therapeutic effectiveness. As such, affective adaptation represents a critical step in moving from emotion sensing to active, personalized intervention in digital mental health.
However, cross-domain studies have shown that facial and textual emotion recognition models experience substantial performance degradation when cultural or population-specific variations are not accounted for. Domain adaptation methods have proven effective in mitigating these biases and restoring model performance (69). In physiological modalities, multi-source alignment frameworks (particularly for EEG and electrodermal activity) have enabled successful cross-subject transfer while reducing reliance on large-scale annotations (70). These findings highlight that general-purpose models alone are insufficient to capture the emotional variability observed in clinical populations. As such, adaptation techniques are essential for improving robustness, fairness, and inclusivity in emotion-aware digital health systems.
4 Affective computing applications in the mental health domain
Building on the historical and technical background, we now turn to a structured overview of how affective computing is applied across different use cases in digital mental health.
4.1 Typology
To structure the possible applications of affective computing in digital mental health systems, we define a typology considering the following dimensions: for whom the system is primarily designed, what the purpose of the system is, and how it responds to the patient’s emotional states. An illustration of this typology can be seen in Figure 2, where the first dimension distinguishes between patient-facing systems, which are used directly by individuals for self-monitoring or therapeutic support, and therapist-facing systems, which assist clinicians with affective insights or decision support.
The second dimension differentiates between the clinical objectives: early identification via screening, detection of existing mental disorders, and interventions to improve the mental state. Often, several of these goals are coupled in one system, as for instance, screening the current state of the patient enables adaptive interventions. Affective computing can be leveraged in screening by automatically detecting subtle emotional and behavioral markers indicative of early mental health changes through multimodal sensing. For diagnosis, it can provide objective, continuous assessments of emotional states to complement clinical evaluation, helping to identify symptom severity or subtype differentiation. In intervention, affective computing enables systems to respond dynamically to users’ emotional states by adapting content, providing empathic feedback, or delivering personalized therapeutic strategies with the goal of improving user engagement and treatment outcomes.
The third dimension concerns the system’s level of affective adaptivity. Passive systems simply detect or log emotional states without adapting their behavior; reactive systems adjust responses based on detected emotions, such as shifting tone; while active systems aim to guide the user in their emotional experience, by, for instance, offering personalized emotional coaching to promote regulation and behavior change. In the following, we will focus on the application of affective computing in interventions, as this often entails the technical implementations given in screening or diagnosis.
4.2 Patient-facing passive adaptivity
When considering passive patient-facing systems that aim to increase mental health by capturing emotional states, the primary application is digital self-monitoring. Research indicates that people who regularly monitor their emotions are better at managing stress and controlling emotional reactions (71, 72). Additionally, monitoring emotions over time can be part of screening or diagnosis as (early) signs of illnesses can be detected, which in turn allows for earlier and potentially more effective interventions. Conventional self-monitoring often requires individuals to manually track their emotions, behaviors, and surrounding circumstances using tools like paper diaries, mood logs, or structured symptom checklists, then, these records can later be examined with clinicians during therapy (73, 74). Despite their clinical value, such traditional approaches can be time-consuming, prone to memory inaccuracies, and often fail to capture emotional experiences as they happen.
Therefore, passive and semi-passive digital tools, including smartphones, wearable devices, and experience sampling apps are employed to track emotional states continuously and in real time (75). For example, studies using real-time emotional tracking methods have found that individuals with mood disorders experience greater emotional stability when they engage in self-monitoring (76) and adolescents that use mobile self-monitoring tools can reduce symptoms of depression (77).
Affective computing systems, explained in Figure 3, leverage such multi-modal data sources. They integrate diverse streams of input such as facial expressions (e.g., detecting smiles, frowns, or expressions of surprise or distress), vocal tone (e.g., pauses, pitch, speed, and tone), physiological signals (e.g., heart rate variability, galvanic skin response, respiration rate, and EEG), and behavioral data (e.g., app usage, movement patterns via accelerometers, typing speed), and digital journaling inputs (textual capture of emotions, experiences, or personal thoughts using digital tools), to detect emotional states with greater accuracy and contextual awareness (1, 78). Affective computing technologies not only reduce the burden of manual tracking, but also facilitate self-awareness by enabling continuous data collection and trend analysis, and furthermore hold potential for early prediction of mental health deterioration or disease onset. The analysis results allow for instantaneous feedback, helping users to become aware of mood fluctuations and potential emotional triggers as they happen (79). Furthermore, users and clinicians can be notified of concerning trends, surpassing the delayed feedback typical of conventional approaches.
Therefore, machine learning algorithms are applied to identify mood patterns, behavioral signals, and emotional irregularities. For example, facial recognition algorithms can analyze micro-expressions to identify emotions like sadness or anger (80), while voice analysis tools detect affective cues from pitch, tempo, and intonation (81). Wearable devices can capture physiological signals such as heart rate variability or galvanic skin response, which are strongly correlated with emotional arousal and stress (82). Additionally, natural language processing can interpret emotional valence in textual or spoken language (83). Overall, a solid technological foundation exists, to detect affect from diverse modalities.
4.3 Patient-facing reactive adaptivity
While passive affective computing systems support emotional self-monitoring by detecting and presenting emotional states to users, reactive systems take this further by adapting their behavior in response to those detected emotions. This is achieved by analyzing affective data in real time and dynamically adjusting the system’s outputs, interface, or prompts based on the user affect. The adaptivity in speaking tone can be used to simulate empathy, as for instance, emotionally aware conversational agents like Wysa,1 and Replika2 to create safe environments that enable users to articulate their feelings and engage in structured emotional reflection (68).
One of the key approaches to improving emotional human-computer interaction is to simulate mirroring, which refers to the unconscious tendency to reflect or mimic another individual’s emotions. This psychological phenomenon can be leveraged in affective computing-based systems to assist users in self-reflection and emotional awareness. For example, Rajcic and McCormack (84) discuss how artificially intelligent systems equipped with mirroring capabilities can foster empathy and promote deeper emotional engagement.
In addition to mirroring, active listening constitutes another central strategy. Unlike passive reception, active listening involves attentively processing users’ verbal and emotional cues to facilitate meaningful exchanges of ideas and emotions. According to Roshanaei et al. (85), artificial intelligence (AI) that actively listens can better navigate empathy, making interactions more personalized and emotionally resonant. The importance of active listening in communication has also been emphasized by Bodie et al. (86), who highlights its central role in fostering understanding and rapport. Furthermore, Oertel et al. (87) propose an engagement-aware, attentive AI listener designed specifically for multi-user interactions, thereby extending active listening’s benefits to more complex social settings.
Technically, reactive systems often operate via rule-based logic or machine learning models that personalize feedback based on real-time affective data (88, 89). These adaptive mechanisms rely on predefined thresholds or models trained on longitudinal data to decide when and how to intervene. By aligning system behavior with users’ affective states, reactive systems help bridge the gap between emotional awareness and behavioral adjustment, providing more targeted and relevant support compared to passive feedback alone.
4.4 Patient-facing active adaptivity
In contrast to passive self-monitoring or reactive feedback systems, active affective computing systems go beyond recognizing and responding to emotional states: They intervene deliberately to guide users through structured emotional development processes. These systems are designed to promote behavioral change, emotional regulation, and therapeutic progress. On a short term basis, this can include recommending just-in-time micro-interventions, which are small behavioral suggestions that aim to increase the well-being in the moment. For instance, the system may recommend listening to AI-generated music based on the patient’s current emotion state (90), or a conversational agent can guide through mindfulness and breathing exercises (91) when stress is detected.
Long-term interventions are often drawing on established psychological frameworks like Cognitive Behavioral Therapy (CBT), where the users can receive personalized support based on AI-driven analysis of their emotional patterns, language use, and engagement levels (7), such as adaptive feedback, tailored coping strategies, and dynamically adjusted task difficulty (92). A prime example of active emotional adaptivity is the aforementioned Woebot, a CBT-based conversational agent3 that offers users structured mental health support by combining mood tracking with real-time psychoeducation and cognitive restructuring techniques. Unlike reactive systems that offer situational prompts, Woebot initiates therapeutic exercises, challenges cognitive distortions, and tracks progress over time, representing an active attempt to influence emotional and behavioral outcomes.
Another important emotional adaptive approach is emotion regulation coaching. Shi (93) present frameworks in which AI assists users in managing their emotions effectively through personalized coaching techniques. It has been also extensively reviewed by Sadka and Antle (94), who demonstrate its widespread application across different domains. Furthermore, empirical studies provide evidence of the effectiveness of AI-driven emotion regulation coaching applied in schools, where such systems help cultivate an adaptive emotional atmosphere conducive to learning and well-being (95, 96).
Additionally, Badia et al. (97) proposed the Emotional Labyrinth, an architecture designed for affective-driven procedural content generation in virtual reality environments, aimed at facilitating more effective emotional regulation. Alipour et al. (98) explored Model-Free Reinforcement Learning mechanisms through which these systems can induce behavioral change in users. For example, the adaptive user interface can intelligently guide users toward safe areas during emergency training by responding to their emotional states, helping them make calm and effective decisions under stress. Numerous efforts have focused on these aspects, with comprehensive surveys of user-centered design approaches and evaluation methodologies provided by Martins et al. (99). Owing to their adaptability, such systems are more readily accepted by users and yield considerable gains in both task performance and perceived usability.
Moreover, affective computing can be leveraged to directly target the treatment of emotional disorders (100), and patient surveys indicate a proportion of openness to such collaboration in mental health care (101, 102). Research includes methods based on multimodal data, gated sequential modeling architectures to extract continuous features of the data, and end-to-end learning systems with video as the direct input and emotion as output (19, 103). Such AI is now being incorporated into therapeutic interventions themselves and offering support (104). These systems can deliver CBT modules, track patient mood over time, and even provide real-time feedback during sessions, which in turn strengthens the foundation for personalized and adaptive therapy, as emotional insights enable therapeutic systems to dynamically tailor interventions to an individual’s current affective state, thereby enhancing treatment relevance and engagement (78). This shift also highlights the technological capabilities of AI in the broader landscape of mental health treatment, reshaping how care is accessed and delivered.
4.5 Therapist facing applications
While digital mental health interventions are generally conceptualized as direct-to-user technologies, therapists remain the conventional and central providers of psychological care. Therefore, a system is considered therapist-facing when the core interpretation and decision-making responsibilities lies with the therapist. Also, in these cases affective computing can be used for improvement.
Effective and timely intervention in mental health care relies heavily on continuous monitoring of patients, as explained in a previous section. This monitoring can be done in-session and out-of-session. Remote monitoring performed out-of-session allows the therapist to gain new insights beyond the boarder of their direct contact with the patients. Recent advancements have introduced wearable technologies integrated with AI, such as electronic skin devices, which can continuously monitor physiological indicators like cortisol levels and skin conductance to assess stress responses (8). The passive monitoring of emotional states during daily activities, such as studying or working, provides the clinician with a broader context for intervention (105). With in-session monitoring affective computing can enhance the therapeutic process by analyzing unstructured clinical session data using models such as ChatGPT (106) to provide the clinicians with real-time insights into patient emotions as the session progresses (107, 108).
Therapeutic support can also be provided remotely through teletherapy: The mental health services and counseling that are provided via the video or phone call. It enables the delivery of care beyond traditional in-person sessions. Teletherapy not only facilitates continuous monitoring of patients when they are not physically present but also offers real-time feedback and intervention during virtual interactions. Affective computing models have been increasingly integrated into teletherapy platforms to enhance these capabilities, enabling more scalable mental health care delivery by for instance providing real-time facial emotional recognition (109). Lenartowicz (110) reviewed the role of AI in teletherapy, emphasizing its potential to improve accessibility and convenience, including the use of virtual reality (VR) interventions to target specific psychological conditions. For instance, simulating feared environments to help treat phobias, or providing immersive scenarios to support PTSD therapy (111, 112). Therefore, research and implementation of AI-assisted teletherapy continue to progress (113).
Furthermore, previous studies have explored potential in assisting clinicians by recommending personalized therapy through large-scale data analysis (68, 114). Deep learning approaches have been investigated for selecting appropriate therapies such as CBT and predicting the relationship between CBT and social anxiety to inform future treatment choices (115). A notable example is that facial expression-based depression detection methods have been proposed to enhance depression diagnosis and improve the quality of counseling (116). Overall, these use-cases show that affective computing does not only carry a lot of potential in directly interacting with patients to improve their well-being, but also to support clinicians in their work.
4.6 Clinical evaluation
As affective computing-based DMHIs become more integrated into mental health applications, clinical evaluation of their appropriateness and effectiveness in outcomes is central. However, only a small number of studies have clinically assessed recent solutions. An overview of performed clinical evaluations is presented in Table 2. For example, Heinz et al. (3) conducted a randomized controlled trial using a chatbot named Therabot with individuals experiencing depression, anxiety, or eating disorders. Results indicated significant reductions in depressive and anxious symptoms, with improvements in emotional recovery and therapeutic alliance comparable to traditional psychological treatments. Similarly, (68) demonstrated that Woebot reduced symptoms for depression and anxiety significantly within two weeks. Likewise, the chatbot Wysa was found effective in real-world usage, lowering psychological distress through guided journaling and reflective dialogue (9). Limbic Access, deployed in the United Kingdom’s National Health Service, utilizes AI-driven assessments to support clinicians based on patient text inputs and clinical records to screen for depression, anxiety, and PTSD while offering a conversational agent to improve recovery rates (4). In the case of psychosis, Avatar Therapy (2) employs virtual avatars to represent hallucinatory voices; randomized controlled trials demonstrate that this approach can significantly reduce distress and hallucination frequency compared to standard treatment.
Applications for monitoring, which have been evaluated for performance rather than clinical outcome, include PhysioFormer, which integrates physiological signals (e.g., electrocardiogram and electrodermal activity) from the WESAD dataset to detect stress and affective states, achieving close to perfect classification accuracy in controlled lab settings (117) clearly indicating an overly optimistic experimental design, and FacePsy, a mobile tool, which analyzes facial expressions and eye-tracking data to identify depressive symptoms, shows promising clinical accuracy (AUROC 81%) in patient trials (118). MoodRhythm, which was assessed in a pilot study (119), combines passive sensing (e.g., GPS, accelerometer) with mood self-reports to provide rhythm-based mood feedback, leading to better self-awareness in bipolar patients. Overall, the actual application of affective computing-based systems in mental healthcare is only emerging. While the technological performance is increasing, large scale studies on the effect on clinical outcome and user engagement are missing.
4.7 Ethical concerns and considerations for clinical applications
The gap between affective computing research and its clinical application is besides ethical skepticism rooted in limited technological performance. Although, reviews for specific applications like monitoring mental health with smart wearable report high performance values ranging from 78%–97% (120), more recent work criticizes the lack of standardized approaches to validate systems, which complicates independent verification and comparison of systems, and also empathizes that current emotion recognition models often perform well under controlled laboratory conditions but face significant performance reduction when applied to in-the-wild data (121). Furthermore, Monteith et al. (122) caution that the commercial deployment of emotion AI may exacerbate social stigma and discrimination because AI models may be biased. Similarly, Hernandez et al. (123) point out that the increasing scale of AI deployment and the shift in agency from experts to lay users, coupled with regulatory shortcomings, may lead to unforeseen risks. A related concern is raised by Shimo (124), who argue that current affective AI systems are often developed and deployed under the assumption of minimal variation in emotional expression across human populations. This limited understanding of human and cultural diversity may compromise the systems’ ability to accurately recognize and interpret the emotions of marginalized or underrepresented groups. From a data ethics perspective, Durovic and Corno (125) emphasize the privacy risks posed by affective computing-based systems, particularly due to its reliance on large volumes of personal data for training and inference. In addition, Chu et al. (126) explore the psychological implications of emotional attachment to AI, warning that perceived relationships between users and machines could lead to confusion and emotional dependency. More critically, Devi et al. (127) suggest that such attachments may even pose threats to personal identity and self-conception.
Besides the technical limitations large ethical concerns exist, particularly regarding patient privacy, informed consent, and emotional manipulation (128, 129). Moreover, a critical debate persists regarding whether AI systems can genuinely exhibit authentic emotions. Vyas (130) has investigated the levels of trust and satisfaction among AI users, revealing complex and nuanced perspectives on this issue. Moreover, Glikson and Woolley (131) and Yang and Rau (132) provide comprehensive reviews of the literature on human trust in affective computing systems, demonstrating that the display of adaptive emotional responses by AI can considerably influence users’ trust, with appropriately aligned emotional expressions generally enhancing perceived trustworthiness and user satisfaction. However, discrepancies or perceived inauthenticity in the AI’s emotional adaptation may foster skepticism and ultimately diminish user trust in the system (131). Similarly, Rubin et al. (133) recently found that identical empathic responses are rated more empathic and supportive when believed to be human-generated than AI-generated, highlighting potential challenges in integrating affective computing systems into emotional caring situations.
Furthermore, the display of emotions does not only influence the user’s trust, but also their engagement. Yu et al. (134) examined the effects of emotional displays on user engagement, demonstrating that emotionally expressive content can enhance user attention and promote a wider range of products through visually captivating designs. Similarly, Maduku et al. (135) investigated the relationship between customer emotions and engagement in the context of AI assistant usage, concluding that positive emotional experiences and engagement with digital voice assistants considerably influence customer loyalty. In line with these findings, Chang and Herath (136) showed that AI systems capable of recognizing, interpreting, and responding to human emotions can effectively strengthen emotional engagement and foster greater trust in human-AI interactions. Beyond assessment of the users, the opinion of clinicians is of high importance, as they can recommend and apply systems. Research conducted in Japan indicates that attitudes toward the use of affective computing systems in healthcare positively correlate with individuals’ familiarity with the technology (137). On a global scale, Doraiswamy et al. (138) found opinions from 791 psychologists from 22 different countries showing that while healthcare professionals do not oppose the integration of such tools, many maintain that these technologies are unlikely to replace human clinicians in delivering truly empathetic care.
In order to improve the application of affective computing, Mohammad (139) provided a structured ethical checklist, highlighting the risks of privacy violations, emotional exploitation, and societal division. He advised integrating fairness, accountability, and explainability at early design stages. Saeidnia et al. (140) further argues that AI-driven emotional interventions should embed ongoing ethical review processes and stakeholder participation. They also call for regular assessments of algorithmic bias, data protection, and stratified impact. In response to these issues, regulatory bodies worldwide are gradually introducing relevant legislation (141). Furthermore, Löchner et al. (142) propose the TEQUILA framework to guide responsible development of digital mental health interventions, and Latif et al. (143) present promises and perils of AI-based emotion recognition to encourage prosocial development. Consequently, future efforts must prioritize improving emotion detection accuracy in-the-wild, while maintaining ethical standards and user privacy, as well as ensuring a level transparency that increases patient and clinician’s trust.
5 Future directions
With the ongoing evolution of affective computing in digital mental health, we suggest that future development should prioritize four key directions: multimodal interaction, explainable modeling, personalized models, and integration with healthcare systems.
First, multimodal input and output will be central to the next stage of development, as affective content is distributed across multiple channels. Compared to unimodal systems, multimodal architectures offer improved recognition accuracy and better contextual awareness (144). Nevertheless, this introduces the challenge of fusing incongruent modalities; for example, a patient may verbally express that they are feeling good but at the same time shake their head. That said, this obstacle can still carry potential, as from a psychological perspective such incoherencies can indicate discrepancies between expressed and experienced emotions, or conflicting self-states (145), and might therefore highlight content that is of therapeutic interest. On the other hand, multimodal output can introduce embodied agents, such as social robots or avatars that respond to the detected emotions through multiple channels like facial expressions, gestures, and gaze while answering verbally. Such agents may serve as empathic companions, co-therapists, or emotion regulation aids, supporting users during stress or therapy by mirroring affect or maintaining calming behavior. Research suggests they can improve engagement, reduce loneliness, and support emotional awareness, particularly in vulnerable populations such as children or older adults (93). However, such systems introduce new risks such as over-attachment and emotional dependency, which require careful application.
Second, explainable AI is essential in digital health applications. In clinical systems, lack of explainability can reduce clinicians’ trust in system outputs and hinder patients’ understanding and acceptance of recommendations, ultimately affecting treatment outcomes. Explainable models improve transparency, fairness, and communication efficiency, support shared decision-making, and promote patient engagement (146). Existing studies have explored various techniques such as visualizing attention in medical imaging in order to understand the central parts of the AI’s decision (147), as well as salient features and counterfactual explanations to prevent clinicians from over-relying on incorrect AI outputs (148). However, when applied to some mental illnesses (such as anxiety and depression) where the primary modalities are audio, physiological signals or brief facial micro-expressions, general-purpose explainable AI methods often struggle. These modalities lack visual structures, making outputs hard to interpret and inaccessible to non-technical users (149). There is thus a high need for task-, modality-, and user-specific explainability frameworks that produce clinically relevant explanations.
Third, personalized modeling is critical to long-term effectiveness and fairness in affective computing systems, as emotional responses vary considerably across individuals, influenced by factors such as demographic attributes and cultural contexts (150). Likewise, mental disorders like depression can manifest differently across cultures (151). Additionally, mental illnesses like schizophrenia can affect the display of emotions, where in the case of flat affect, there is barely or no external emotional display, while the subjective experience of emotions is not diminished (152). Such individual and disease-specific differences need to be accounted for, as uncertainty in emotion recognition could lead to inappropriate emotional adaptation from the system. Additionally, mental disorders have an individual progression, which is reflected by the moving target problem (153). The user’s emotional and mental state evolves due to therapeutic progress, life events, or symptom fluctuation, as is the case for flat affect in patients who have suffered from psychosis (154). Although these adaptations to culture, disease, and personal progression are central for a high-quality therapy, such specifications of affective computing models are still only under development and require further research.
Finally, the clinical deployment of affective computing systems is lacking behind the start-of-the-art in research and requires deeper exploration at both system integration and application levels. Future models should interface with electronic health records, support telemedicine platforms, and be embedded within intervention planning workflows (155). Ensuring closed-loop mechanisms for data standardization, privacy compliance, and clinical acceptability will be essential to integration. Recently, privacy-preserving federated learning frameworks have been introduced into physiological and speech-based emotion recognition. These architectures allow model training across distributed devices without sharing raw data, enabling a balance between data protection and cross-device generalization (156), and therefore pose a promising approach for DMHIs. Furthermore, an increase in clinical deployment requires clinical validation and large-scale longitudinal studies in order to ensure therapeutic value. Concluding, the application of affective computing for mental health requires further technical improvements such as multimodal integration and personalization, as well as an increase in explainability and safety to increase trust, while clinical trials are outstanding to ensure an actual benefit of the application of affective computing.
6 Discussion and conclusion
This review has outlined how affective computing technologies are increasingly integrated in digital mental health interventions, spanning both patient- and therapist-facing applications. These systems aim to recognize, interpret, and respond to users’ emotional states in real-time to support self-monitoring, enhancing engagement, and enable more personalized and adaptive interventions (107–109). A central distinction in this review is between passive, reactive, and active adaptivity. Nevertheless, the boundaries between these types are often blurred in practice. Despite increasing sophistication, emotion detection technologies still face validity and reliability concerns, especially across diverse populations. Current algorithms often struggle with cultural variation, individual differences, and context sensitivity (121, 124). Without accurate and interpretable emotion recognition, downstream interventions may misfire and therefore undermine user trust and therapeutic efficacy (131, 132).
While much attention in digital mental health has focused on direct-to-user tools, we highlight the growing role of therapist-facing applications. Affective computing technologies can support clinicians by offering real-time emotional insights, informing diagnosis, enabling personalized care, and extending monitoring beyond therapy sessions (4, 107). Such hybrid models preserve the therapist’s expertise while enriching it with algorithmic support. However, these systems raise new questions around accountability, interpretability, and the therapist’s role in AI-augmented decision-making (110, 139). If affective feedback is incorrect or biased, how should clinicians interpret or override it? Integration of explainable AI methods and clinician-in-the-loop frameworks will be essential for ensuring these systems remain clinically meaningful and ethically deployable (140).
Moreover, large ethical challenges are being faced, as empathetic AI responses may risk emotional dependency or manipulation, especially when systems are designed to mimic human emotion (126, 127). Moreover, the commercial deployment of affective systems raises concerns around bias, privacy, and social inequality, since models trained on non-diverse data may misinterpret emotions in marginalized groups (122, 124), while real-time affective monitoring could impose surveillance risks if not properly regulated (125). Additionally, commercial mental health applications face the risk of being designed in a way that keeps users attached in order to increase customer numbers instead of designing systems that aim to promote recovery in the fastest and safest way. Consequently, transparency in AI systems as well as their development should be of highest interest, to ensure safe deployment with high clinical value.
In the future affective computing will likely play an increasing role in both patient-facing interventions and clinical workflows. Key directions for development include multimodal architectures, explainable and personalized models, and integration with healthcare infrastructure. In particular, embodied agents, such as emotionally expressive avatars and robots, may offer new modes of support by combining affective sensing with naturalistic, human-like interactions (93). Nevertheless, affective computing technologies should be viewed not as replacements for human care, but as tools to enhance and extend it. To enhance the trust, clinical evaluation of the effect of affective computing on the clinical outcome require extensive attention in the future, as it has only been done scarcely so far. Concluding, the success of affective computing for mental health will depend on interdisciplinary collaboration across AI, psychology, and clinical practice, as well as robust evaluation frameworks.
Author contributions
MS: Writing – review & editing, Writing – original draft, Visualization, Conceptualization, Methodology, Data curation. YL: Writing – review & editing, Writing – original draft. SM: Writing – original draft, Visualization, Writing – review & editing. QS: Writing – original draft, Writing – review & editing. BS: Writing – review & editing, Supervision.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work has received funding from the DFG’s Reinhart Koselleck project No. 442218748 (AUDI0NOMOUS).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Footnotes
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
2. Garety PA, Edwards CJ, Jafari H, Emsley R, Huckvale M, Rus-Calafell M, et al. Digital AVATAR therapy for distressing voices in psychosis: the phase 2/3 AVATAR2 trial. Nat Med. (2024) 30:3658–68. doi: 10.1038/s41591-024-03252-8
3. Heinz MV, Mackin DM, Trudeau BM, Bhattacharya S, Wang Y, Banta HA, et al. Randomized trial of a generative AI chatbot for mental health treatment. NEJM AI. (2025) 2:AIoa2400802. doi: 10.1056/AIoa2400802
4. Rollwage M, Juchems K, Habicht J, Carrington B, Hauser T, Harper R. Conversational AI facilitates mental health assessments and is associated with improved recovery rates. medRxiv [Preprint]. 2022–11 (2022).
5. Boucher EM, Harake NR, Ward HE, Stoeckl SE, Vargas J, Minkel J, et al. Artificially intelligent chatbots in digital mental health interventions: a review. Expert Rev Med Devices. (2021) 18:37–49. doi: 10.1080/17434440.2021.2013200
6. Lattie EG, Adkins EC, Winquist N, Stiles-Shields C, Wafford QE, Graham AK. Digital mental health interventions for depression, anxiety, and enhancement of psychological well-being among college students: systematic review. J Med Internet Res. (2019) 21:e12869. doi: 10.2196/12869
7. Striegl J, Richter JW, Grossmann L, Bråstad B, Gotthardt M, Rück C, et al. Deep learning-based dimensional emotion recognition for conversational agent-based cognitive behavioral therapy. PeerJ Comput Sci. (2024) 10:e2104. doi: 10.7717/peerj-cs.2104
8. Xu C, Song Y, Sempionatto JR, Solomon SA, Yu Y, Nyein HY, et al. A physicochemical-sensing electronic skin for stress response monitoring. Nat Electron. (2024) 7:168–79. doi: 10.1038/s41928-023-01116-6
9. Inkster B, Sarda S, Subramanian V. An empathy-driven, conversational artificial intelligence agent (Wysa) for digital mental well-being: real-world data evaluation mixed-methods study. JMIR mHealth uHealth. (2018) 6:e12106. doi: 10.2196/12106
10. Greene S, Thapliyal H, Caban-Holt A. A survey of affective computing for stress detection: evaluating technologies in stress detection for better health. IEEE Consum Electron Mag. (2016) 5:44–56. doi: 10.1109/MCE.2016.2590178
11. Pepa L, Spalazzi L, Capecci M, Ceravolo MG. Automatic emotion recognition in clinical scenario: a systematic review of methods. IEEE Trans Affect Comput. (2021) 14:1675–95. doi: 10.1109/TAFFC.2021.3128787
12. Wang Y, Song W, Tao W, Liotta A, Yang D, Li X, et al. A systematic review on affective computing: emotion models, databases, and recent advances. Inf Fusion. (2022) 83:19–52. doi: 10.1016/j.inffus.2022.03.009
13. Eisenstadt M, Liverpool S, Infanti E, Ciuvat RM, Carlsson C. Mobile apps that promote emotion regulation, positive mental health, and well-being in the general population: systematic review and meta-analysis. JMIR Ment Health. (2021) 8:e31170. doi: 10.2196/31170
14. Lehtimaki S, Martic J, Wahl B, Foster KT, Schwalbe N. Evidence on digital mental health interventions for adolescents and young people: systematic overview. JMIR Ment Health. (2021) 8:e25847. doi: 10.2196/25847
15. Philippe TJ, Sikder N, Jackson A, Koblanski ME, Liow E, Pilarinos A, et al. Digital health interventions for delivery of mental health care: systematic and comprehensive meta-review. JMIR Ment Health. (2022) 9:e35159. doi: 10.2196/35159
16. Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan—a web and mobile app for systematic reviews. Syst Rev. (2016) 5:210. doi: 10.1186/s13643-016-0384-4
17. Ekman P, Dalgleish T, Power M. Basic emotions. In: Dalgleish T, Powerm MJ, editors. Handbook of Cognition and Emotion. San Francisco, USA: John Wiley & Sons Ltd (1999). p. 45–60. doi: 10.1002/0470013494.ch3
18. Lyons M, Akamatsu S, Kamachi M, Gyoba J. Coding facial expressions with Gabor wavelets. In: Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition. IEEE (1998). p. 200–5.
19. Schuller B, Rigoll G, Lang M. Hidden Markov model-based speech emotion recognition. In: 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings.(ICASSP’03). IEEE (2003). Vol. 2, II–1.
20. Pang B, Lee L, Vaithyanathan S. Thumbs up? Sentiment classification using machine learning techniques. In: Proceedings of the ACL-02 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics (2002). p. 79–86.
21. Breazeal C. Emotion and sociable humanoid robots. Int J Hum Comput Stud. (2003) 59:119–55. doi: 10.1016/S1071-5819(03)00018-1
22. D’Mello S, Kappas A, Gratch J. The affective computing approach to affect measurement. Emot Rev. (2018) 10:174–83. doi: 10.1177/1754073917696583
23. Kye S, Moon J, Lee J, Choi I, Cheon D, Lee K. Multimodal data collection framework for mental stress monitoring. In: Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2017 ACM International Symposium on Wearable Computers. (2017). p. 822–9.
24. McDuff D, Mahmoud A, Mavadati M, Amr M, Turcot J, Kaliouby R. AFFDEX SDK: a cross-platform real-time multi-face expression recognition toolkit. In: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems. (2016). p. 3723–6.
25. Gratch J, Artstein R, Lucas GM, Stratou G, Scherer S, Nazarian A, et al. The distress analysis interview corpus of human and computer interviews. In: LREC. Reykjavik (2014). Vol. 14. p. 3123–8.
26. Valstar M, Gratch J, Schuller B, Ringeval F, Lalanne D, Torres MT, et al. AVEC 2016: depression, mood, and emotion recognition workshop and challenge. In: Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge. (2016). p. 3–10.
27. Koldijk S, Sappelli M, Verberne S, Neerincx MA, Kraaij W. The swell knowledge work dataset for stress and user modeling research. In: Proceedings of the 16th International Conference on Multimodal Interaction. (2014). p. 291–8.
28. Sriramprakash S, Prasanna VD, Murthy OR. Stress detection in working people. Procedia Comput Sci. (2017) 115:359–66. doi: 10.1016/j.procs.2017.09.090
29. Cummins N, Vlasenko B, Sagha H, Schuller B. Enhancing speech-based depression detection through gender dependent vowel-level formant features. In: Artificial Intelligence in Medicine: 16th Conference on Artificial Intelligence in Medicine, AIME 2017, Vienna, Austria, June 21-24, 2017, Proceedings 16. Springer (2017). p. 209–14.
30. Pampouchidou A, Simantiraki O, Fazlollahi A, Pediaditis M, Manousos D, Roniotis A, et al. Depression assessment by fusing high and low level features from audio, video, and text. In: Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge. (2016). p. 27–34.
31. Jadhav N, Sugandhi R. Survey on human behavior recognition using affective computing. In: 2018 IEEE Global Conference on Wireless Computing and Networking (GCWCN). IEEE (2018). p. 98–103.
32. D’mello SK, Kory J. A review and meta-analysis of multimodal affect detection systems. ACM Comput Surv. (2015) 47:1–36. doi: 10.1145/2682899
33. Rejaibi E, Komaty A, Meriaudeau F, Agrebi S, Othmani A. MFCC-based recurrent neural network for automatic clinical depression recognition and assessment from speech. Biomed Signal Process Control. (2022) 71:103107. doi: 10.1016/j.bspc.2021.103107
34. Ray A, Kumar S, Reddy R, Mukherjee P, Garg R. Multi-level attention network using text, audio and video for depression prediction. In: Proceedings of the 9th International on Audio/Visual Emotion Challenge and Workshop. (2019). p. 81–8.
35. Wu Y, Daoudi M, Amad A. Transformer-based self-supervised multimodal representation learning for wearable emotion recognition. IEEE Trans Affect Comput. (2023) 15:157–72. doi: 10.1109/TAFFC.2023.3263907
36. Amin MM, Cambria E, Schuller BW. Will affective computing emerge from foundation models and general artificial intelligence? A first evaluation of ChatGPT. IEEE Intell Syst. (2023) 38:15–23. doi: 10.1109/MIS.2023.3254179
37. Bucur AM, Cosma A, Rosso P, Dinu LP. It’s just a matter of time: detecting depression with time-enriched multimodal transformers. In: European Conference on Information Retrieval. Springer (2023). p. 200–15.
38. Rahmani AM, Lai J, Jafarlou S, Azimi I, Yunusova A, Rivera AP, et al. Personal mental health navigator: Harnessing the power of data, personal models, and health cybernetics to promote psychological well-being. Front Digit Health. (2022) 4:933587. doi: 10.3389/fdgth.2022.933587
39. Henderson K, Reihm J, Koshal K, Wijangco J, Sara N, Miller N, et al. A closed-loop digital health tool to improve depression care in multiple sclerosis: iterative design and cross-sectional pilot randomized controlled trial and its impact on depression care. JMIR Formative Res. (2024) 8:e52809. doi: 10.2196/52809
40. Afzal S, Khan HA, Piran MJ, Lee JW. A comprehensive survey on affective computing: Challenges, trends, applications, and future directions. IEEE Access. (2024) 12:96150–68. doi: 10.1109/ACCESS.2024.3422480
41. Pampouchidou A, Simantiraki O, Vazakopoulou CM, Chatzaki C, Pediaditis M, Maridaki A, et al. Facial geometry and speech analysis for depression detection. In: 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE (2017). p. 1433–6.
42. Alghowinem S, Goecke R, Epps J, Wagner M, Cohn JF. Cross-cultural depression recognition from vocal biomarkers. Interspeech. (2016) 1943–7. doi: 10.21437/Interspeech.2016-1339
43. Gendron M, Crivelli C, Barrett LF. Universality reconsidered: diversity in making meaning of facial expressions. Curr Dir Psychol Sci. (2018) 27:211–9. doi: 10.1177/0963721417746794
44. Durán JI, Fernández-Dols JM. Do emotions result in their predicted facial expressions? A meta-analysis of studies on the co-occurrence of expression and emotion. Emotion. (2021) 21:1550. doi: 10.1037/emo0001015
45. Fernández-Dols JM, Crivelli C. Emotion and expression: naturalistic studies. Emot Rev. (2013) 5:24–9. doi: 10.1177/1754073912457229
46. Fernandez-Dols JM, Sanchez F, Carrera P, Ruiz-Belda MA. Are spontaneous expressions and emotions linked? An experimental test of coherence. J Nonverbal Behav. (1997) 21:163–77. doi: 10.1023/A:1024917530100
47. Wundt WM, Wundt W. Grundzüge der Physiologischen Psychologie. Leipzig: Engelmann (1874). Vol. 1.
48. Bautista JL, Shin HS. Speech emotion recognition model based on joint modeling of discrete and dimensional emotion representation. Appl Sci. (2025) 15:623. doi: 10.3390/app15020623
49. Mendes GA, Martins B. Quantifying valence and arousal in text with multilingual pre-trained transformers. In: European Conference on Information Retrieval. Springer (2023). p. 84–100.
50. Ahmed A, Ramesh J, Ganguly S, Aburukba R, Sagahyroon A, Aloul F. Investigating the feasibility of assessing depression severity and valence-arousal with wearable sensors using discrete wavelet transforms and machine learning. Information. (2022) 13:406. doi: 10.3390/info13090406
51. Du S, Tao Y, Martinez AM. Compound facial expressions of emotion. Proc Natl Acad Sci. (2014) 111:E1454–62. doi: 10.1073/pnas.1322355111
52. Barrett LF, Adolphs R, Marsella S, Martinez AM, Pollak SD. Emotional expressions reconsidered: challenges to inferring emotion from human facial movements. Psychol Sci Public Interest. (2019) 20:1–68. doi: 10.1177/1529100619832930
53. Gendron M, Feldman Barrett L. Reconstructing the past: a century of ideas about emotion in psychology. Emot Rev. (2009) 1:316–39. doi: 10.1177/1754073909338877
54. Ahmadi A, Galusha JM, Ponder WN, Carbajal J, Schuman DL, Whitworth J, et al. Validation of the PCL-5, PHQ-9, and GAD-7 in a sample of first responders. J Occup Environ Med. (2023) 65:467–76. doi: 10.1097/JOM.0000000000002823
55. Suppakhun Y, Boonkuson N, Thongbai N. Depression screening with PHQ-9 combined with facial emotion analysis. In: 2024 International Technical Conference on Circuits/Systems, Computers, and Communications (ITC-CSCC). IEEE (2024). p. 1–5.
56. Ren X, Su S, Jiang W, Wang Y, Yao J, Su Y, et al. Reliability and validity of facial expression recognition in the screening of generalized anxiety disorder. J Affect Disord. (2025) 369:538–46. doi: 10.1016/j.jad.2024.10.022
57. Sun L, Lian Z, Liu B, Tao J. MAE-DFER: efficient masked autoencoder for self-supervised dynamic facial expression recognition. In: Proceedings of the 31st ACM International Conference on Multimedia. (2023). p. 6110–21.
58. Wani TM, Gunawan TS, Qadri SAA, Kartiwi M, Ambikairajah E. A comprehensive review of speech emotion recognition systems. IEEE Access. (2021) 9:47795–814. doi: 10.1109/ACCESS.2021.3068045
59. Wagner J, Triantafyllopoulos A, Wierstorf H, Schmitt M, Burkhardt F, Eyben F, et al. Dawn of the transformer era in speech emotion recognition: closing the valence gap. IEEE Trans Pattern Anal Mach Intell. (2023) 45:10745–59. doi: 10.1109/TPAMI.2023.3263585
60. Luzzani G, Buraioli I, Guglieri G, Demarchi D. EDA, PPG and skin temperature as predictive signals for mental failure by a statistical analysis on stress and mental workload. IEEE Open J Eng Med Biol. (2024) 6:248–55. doi: 10.1109/OJEMB.2024.3515473
61. Pereira P, Moniz H, Carvalho JP. Deep emotion recognition in textual conversations: a survey. Artif Intell Rev. (2025) 58:1–37. doi: 10.1007/s10462-024-11010-y
62. Pan B, Hirota K, Jia Z, Dai Y. A review of multimodal emotion recognition from datasets, preprocessing, features, and fusion methods. Neurocomputing. (2023) 561:126866. doi: 10.1016/j.neucom.2023.126866
63. Udahemuka G, Djouani K, Kurien AM. multimodal emotion recognition using visual, vocal and physiological signals: a review. Appl Sci. (2024) 14:8071. doi: 10.3390/app14178071
64. Mitra V, Tsiartas A, Shriberg E. Noise and reverberation effects on depression detection from speech. In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE (2016). p. 5795–9.
65. Pringle A, Harmer CJ. The effects of drugs on human models of emotional processing: an account of antidepressant drug treatment. Dialogues Clin Neurosci. (2015) 17:477–87. doi: 10.31887/DCNS.2015.17.4/apringle
66. Tervonen J, Nath RK, Pettersson K, Närväinen J, Mäntyjärvi J. Cold-start model adaptation: evaluation of short baseline calibration. In: Adjunct Proceedings of the 2023 ACM International Joint Conference on Pervasive and Ubiquitous Computing & the 2023 ACM International Symposium on Wearable Computing. (2023). p. 417–22.
67. Jaques N, Rudovic O, Taylor S, Sano A, Picard R. Predicting tomorrow’s mood, health, and stress level using personalized multitask learning and domain adaptation. In: IJCAI 2017 Workshop on Artificial Intelligence in Affective Computing. PMLR (2017). p. 17–33.
68. Fitzpatrick KK, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment Health. (2017) 4:e7785. doi: 10.2196/mental.7785
69. Singhal P, Gokhale S, Shah A, Jain DK, Walambe R, Ekart A, et al. Domain adaptation for bias mitigation in affective computing: use cases for facial emotion recognition and sentiment analysis systems. Discov Appl Sci. (2025) 7:229. doi: 10.1007/s42452-025-06659-1
70. Guo W, Xu G, Wang Y. Multi-source domain adaptation with spatio-temporal feature extractor for EEG emotion recognition. Biomed Signal Process Control. (2023) 84:104998. doi: 10.1016/j.bspc.2023.104998
71. Gross JJ. Emotion regulation: current status and future prospects. Psychol Inq. (2015) 26:1–26. doi: 10.1080/1047840X.2014.940781
73. Stone AA, Shiffman S, Schwartz JE, Broderick JE, Hufford MR. Patient compliance with paper and electronic diaries. Control Clin Trials. (2003) 24:182–99. doi: 10.1016/S0197-2456(02)00320-3
74. Thiele C, Laireiter AR, Baumann U. Diaries in clinical psychology and psychotherapy: a selective review. Clin Psychol Psychother. (2002) 9:1–37. doi: 10.1002/cpp.302
75. Ben-Zeev D, Scherer EA, Wang R, Xie H, Campbell AT. Next-generation psychiatric assessment: using smartphone sensors to monitor behavior and mental health. Psychiatr Rehabil J. (2015) 38:218. doi: 10.1037/prj0000130
76. Myin-Germeys I, van Os J, Schwartz JE, Stone AA, Delespaul PA. Emotional reactivity to daily life stress in psychosis. Arch Gen Psychiatry. (2001) 58:1137–44. doi: 10.1001/archpsyc.58.12.1137
77. Kauer SD, Reid SC, Crooke AHD, Khor A, Hearps SJC, Jorm AF, et al. Self-monitoring using mobile phones in the early stages of adolescent depression: randomized controlled trial. J Med Internet Res. (2012) 14:e1858. doi: 10.2196/jmir.1858
78. Calvo RA, D’Mello S. Affect detection: an interdisciplinary review of models, methods, and their applications. IEEE Trans Affect Comput. (2010) 1:18–37. doi: 10.1109/T-AFFC.2010.1
79. Larsen ME, Nicholas J, Christensen H. A systematic assessment of smartphone tools for suicide prevention. PLoS One. (2016) 11:e0152285. doi: 10.1371/journal.pone.0152285
80. Ekman P. Data from: an argument for basic emotions. Cogn Emot. (1992) 6(3-4):169–200. doi: 10.1080/02699939208411068
81. Schuller B, Batliner A, Steidl S, Seppi D. Recognising realistic emotions and affect in speech: state of the art and lessons learnt from the first challenge. Speech Commun. (2011) 53:1062–87. doi: 10.1016/j.specom.2011.01.011
82. Sano A, Picard RW. Stress recognition using wearable sensors and mobile phones. In: 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction. IEEE (2013). p. 671–6.
83. Tausczik YR, Pennebaker JW. The psychological meaning of words: LIWC and computerized text analysis methods. J Lang Soc Psychol. (2010) 29:24–54. doi: 10.1177/0261927X09351676
84. Rajcic N, McCormack J. Mirror ritual: an affective interface for emotional self-reflection. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. (2020). p. 1–13.
85. Roshanaei M, Rezapour R, El-Nasr MS. Talk, listen, connect: navigating empathy in human-AI interactions. arXiv [Preprint]. arXiv:2409.15550 (2024).
86. Bodie GD, Vickery AJ, Cannava K, Jones SM. The role of “active listening” in informal helping conversations: impact on perceptions of listener helpfulness, sensitivity, and supportiveness and discloser emotional improvement. West J Commun. (2015) 79:151–73. doi: 10.1080/10570314.2014.943429
87. Oertel C, Jonell P, Kontogiorgos D, Mora KF, Odobez JM, Gustafson J. Towards an engagement-aware attentive artificial listener for multi-party interactions. Front Rob AI. (2021) 8:555913. doi: 10.3389/frobt.2021.555913
88. Hu X, Qian M, Cheng B, Cheung YK. Personalized policy learning using longitudinal mobile health data. J Am Stat Assoc. (2021) 116:410–20. doi: 10.1080/01621459.2020.1785476
89. Gheibi O, Weyns D, Quin F. Applying machine learning in self-adaptive systems: a systematic literature review. ACM Trans Auton Adapt Syst. (2021) 15:1–37. doi: 10.1145/3469440
90. Shen L, Zhang H, Zhu C, Li R, Qian K, Tian F, et al. Enhancing emotion regulation in mental disorder treatment: an AIGC-based closed-loop music intervention system. IEEE Trans Affect Comput. (2025) 16:2245–60. doi: 10.1109/TAFFC.2025.3557873
91. Mitsea E, Drigas A, Skianis C. Artificial intelligence, immersive technologies, and neurotechnologies in breathing interventions for mental and emotional health: a systematic review. Electronics. (2024) 13:2253. doi: 10.3390/electronics13122253
92. Huang H, Zeng W, Wang T, Liu Y, Cao C, Li R, et al. AI-enhanced HTP test analysis and emotion recognition for personalized psychotherapy interventions. In: International Conference on Human-Computer Interaction. Springer (2025). p. 301–13.
93. Shi L. The integration of advanced AI-enabled emotion detection and adaptive learning systems for improved emotional regulation. J Educ Comput Res. (2025) 63:173–201. doi: 10.1177/07356331241296890
94. Sadka O, Antle A. Interactive technologies for emotion regulation training: a scoping review. Int J Hum Comput Stud. (2022) 168:102906. doi: 10.1016/j.ijhcs.2022.102906
95. Akintayo OT, Eden CA, Ayeni OO, Onyebuchi NC. Integrating AI with emotional and social learning in primary education: developing a holistic adaptive learning ecosystem. Comput Sci IT Res J. (2024) 5:1076–89. doi: 10.51594/csitrj.v5i5.1116
96. Galea E. Advancing AI-driven pedagogy through the regulation of emotions: a vision for school practitioners. Ital J Health Educ Sports Inclusive Didactics. (2024) 8. doi: 10.32043/gsd.v8i3.1085
97. Badia SB, Quintero LV, Cameirao MS, Chirico A, Triberti S, Cipresso P, et al. Toward emotionally adaptive virtual reality for mental health applications. IEEE J Biomed Health Inf. (2018) 23:1877–87. doi: 10.1109/JBHI.2018.2878846
98. Alipour M, Moghaddam MT, Vaidhyanathan K, Kjærgaard MB. Toward changing users behavior with emotion-based adaptive systems. In: Proceedings of the 31st ACM Conference on User Modeling, Adaptation and Personalization. (2023). p. 85–95.
99. Martins GS, Santos L, Dias J. User-adaptive interaction in social robots: a survey focusing on non-physical interaction. Int J Soc Rob. (2019) 11:185–205. doi: 10.1007/s12369-018-0485-4
100. Li Y, Shao S, Milling M, Schuller BW. Large language models for depression recognition in spoken language integrating psychological knowledge. arXiv [Preprint]. arXiv:2505.22863 (2025).
101. Mehta A, Niles AN, Vargas JH, Marafon T, Couto DD, Gross JJ. Acceptability and effectiveness of artificial intelligence therapy for anxiety and depression (Youper): longitudinal observational study. J Med Internet Res. (2021) 23:e26771. doi: 10.2196/26771
102. Robertson C, Woods A, Bergstrand K, Findley J, Balser C, Slepian MJ. Diverse patients’ attitudes towards artificial intelligence (AI) in diagnosis. PLOS Digit Health. (2023) 2:e0000237. doi: 10.1371/journal.pdig.0000237
103. Tzirakis P, Trigeorgis G, Nicolaou MA, Schuller BW, Zafeiriou S. End-to-end multimodal emotion recognition using deep neural networks. IEEE J Sel Top Signal Process. (2017) 11:1301–9. doi: 10.1109/JSTSP.2017.2764438
104. Gual-Montolio P, Jaén I, Martínez-Borba V, Castilla D, Suso-Ribera C. Using artificial intelligence to enhance ongoing psychological interventions for emotional problems in real-or close to real-time: a systematic review. Int J Environ Res Public Health. (2022) 19:7737. doi: 10.3390/ijerph19137737
105. Vistorte AOR, Deroncele-Acosta A, Ayala JLM, Barrasa A, López-Granero C, Martí-González M. Integrating artificial intelligence to assess emotions in learning environments: a systematic literature review. Front Psychol. (2024) 15:1387089. doi: 10.3389/fpsyg.2024.1387089
106. OpenAI. Data from: ChatGPT (June 26 version). Large language model (2025). Available online at: https://chat.openai.com/.
107. Meier ST. Initial evaluation of an AI-augmented progress monitoring and outcome assessment. Couns Psychother Res. (2025) 25:e12869. doi: 10.1002/capr.12869
108. Vijay K, Raghakeerthana R, Milinda RR, Thusheel S. AI-powered mental health assessment using emotion detection for real-time analysis. In: 2025 International Conference on Computational, Communication and Information Technology (ICCCIT). IEEE (2025). p. 530–5.
109. Singh J, Shelke NA, Ghosh S, Nagasri A, Yadav A, Elangovan M. Deep learning integration in teletherapy for modernizing mental health support. In: 2024 2nd World Conference on Communication & Computing (WCONF). IEEE (2024). p. 1–6.
110. Lenartowicz R. Technology in mental health treatment: a review of teletherapy, mobile health apps, virtual reality, and artificial intelligence. Intersect Stanford J Sci Technol Soc. (2023) 17. https://ojs.stanford.edu/ojs/index.php/intersect/article/view/2859
111. Maples-Keller JL, Bunnell BE, Kim SJ, Rothbaum BO. The use of virtual reality technology in the treatment of anxiety and other psychiatric disorders. Harv Rev Psychiatry. (2017) 25:103–13. doi: 10.1097/HRP.0000000000000138
112. Emmelkamp PM, Meyerbröker K. Virtual reality therapy in mental health. Annu Rev Clin Psychol. (2021) 17:495–519. doi: 10.1146/annurev-clinpsy-081219-115923
113. Hebbar S, Vandana B. Artificial intelligence in future telepsychiatry and psychotherapy for e-mental health revolution. In: Computational Intelligence in Medical Decision Making and Diagnosis. Boca Raton, FL: CRC Press (2023). p. 39–60. doi: 10.1201/9781003309451-3
114. Jain A, Jain A. AI-based emotion detection system in healthcare for patient. Gen Artif Intell Biomed Smart Health Inf. (2025) 455–70. doi: 10.1002/9781394280735.ch23
115. Al-Abyadh MHA, Hoang VT. Emotion AI: cognitive behavioral therapy for teens having some mental health disorders. In: Garg M, Koundal D, editors. Emotional AI and Human-AI Interactions in Social Networking. London: Academic Press (2024). p. 169–89. doi: 10.1016/B978-0-443-19096-4.00001-8
116. Nixon D, Mallappa VV, Petli V, HosgurMath S, K SK. A novel AI therapy for depression counseling using face emotion techniques. Glob Transit Proc. (2022) 3:190–4. doi: 10.1016/j.gltp.2022.03.008
117. Wang Z, Wu W, Zeng C. Physioformer: integrating multimodal physiological signals and symbolic regression for explainable affective state prediction. arXiv [Preprint]. arXiv:2410.11376 (2024).
118. Islam R, Bae SW. FacePsy: an open-source affective mobile sensing system-analyzing facial behavior and head gesture for depression detection in naturalistic settings. Proc ACM Hum-Comput Interact. (2024) 8:1–32. doi: 10.1145/3676505
119. Matthews M, Abdullah S, Murnane E, Voida S, Choudhury T, Gay G, et al. Development and evaluation of a smartphone-based measure of social rhythms for bipolar disorder. Assessment. (2016) 23:472–83. doi: 10.1177/1073191116656794
120. Long N, Lei Y, Peng L, Xu P, Mao P. A scoping review on monitoring mental health using smart wearable devices. Math Biosci Eng. (2022) 19:7899–919. doi: 10.3934/mbe.2022369
121. D’Mello SK, Booth BM. Affect detection from wearables in the “real” wild: fact, fantasy, or somewhere in between? IEEE Intell Syst. (2023) 38:76–84. doi: 10.1109/MIS.2022.3221854
122. Monteith S, Glenn T, Geddes J, Whybrow PC, Bauer M. Commercial use of emotion artificial intelligence (AI): implications for psychiatry. Curr Psychiatry Rep. (2022) 24:203–11. doi: 10.1007/s11920-022-01330-7
123. Hernandez J, Lovejoy J, McDuff D, Suh J, O’Brien T, Sethumadhavan A, et al. Guidelines for assessing and minimizing risks of emotion recognition applications. In: 2021 9th International Conference on Affective Computing and Intelligent Interaction (ACII). IEEE (2021). p. 1–8.
124. Shimo S. Risks of bias in AI-based emotional analysis technology from diversity perspectives. In: 2020 IEEE International Symposium on Technology and Society (ISTAS). IEEE (2020). p. 66–8.
125. Durovic M, Corno T. The privacy of emotions: from the GDPR to the AI Act, an overview of emotional AI regulation and the protection of privacy and personal data. In: Ebers M, Sein K, editors. Privacy, Data Protection and Data-driven Technologies. London: Routledge (2024). p. 368–404.
126. Chu MD, Gerard P, Pawar K, Bickham C, Lerman K. Illusions of intimacy: emotional attachment and emerging psychological risks in human-AI relationships. arXiv [Preprint]. arXiv:2505.11649 (2025).
127. Devi S, Kaur S, Goel D, Yadav B. Navigating the perils: exploring identity threats confronting emotional artificial intelligence. In: Tikadar S, Liu H, Bhattacharya P, Bhattacharya S, editors. Humanizing Technology With Emotional Intelligence. Hershey, PA: IGI Global Scientific Publishing (2025). p. 433–42. doi: 10.4018/979-8-3693-7011-7.ch022
128. Kenku AA, Tafida AM, Umar SS. Ethical challenges in tele-therapy: a modern perspective. Int J Psychother Afr. (2025) 10:35–50.
129. Devillers L, Cowie R. Ethical considerations on affective computing: an overview. Proc IEEE. (2023) 111:1445–58. doi: 10.1109/JPROC.2023.3315217
130. Vyas AM. Exploring the impact of “emotion-recognition-AI” on consumer trust and satisfaction. In: 2024 IEEE International Students’ Conference on Electrical, Electronics and Computer Science (SCEECS). IEEE (2024). p. 1–6.
131. Glikson E, Woolley AW. Human trust in artificial intelligence: review of empirical research. Acad Manage Ann. (2020) 14:627–60. doi: 10.5465/annals.2018.0057
132. Yang M, Rau PLP. Trust building with artificial intelligence: comparing with human in investment behaviour, emotional arousal and neuro activities. Theor Issues Ergon Sci. (2024) 25:593–614. doi: 10.1080/1463922X.2023.2290019
133. Rubin M, Li JZ, Zimmerman F, Ong DC, Goldenberg A, Perry A. Comparing the value of perceived human versus AI-generated empathy. Nat Hum Behav. (2025) 1–15. doi: 10.1038/s41562-025-02247-w
134. Yu J, Dickinger A, So KKF, Egger R. Artificial intelligence-generated virtual influencer: examining the effects of emotional display on user engagement. J Retailing Consum Serv. (2024) 76:103560. doi: 10.1016/j.jretconser.2023.103560
135. Maduku DK, Rana NP, Mpinganjira M, Thusi P, Mkhize NHB, Ledikwe A. Do AI-powered digital assistants influence customer emotions, engagement and loyalty? An empirical investigation. Asia Pac J Mark Logistics. (2024) 36:2849–68. doi: 10.1108/APJML-09-2023-0935
136. Chang F, Herath D. From interaction to relationship: the role of memory, learning, and emotional intelligence in AI-embodied human engagement. In: 2025 20th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE (2025). p. 1269–73.
137. Ho MT, Le NTB, Mantello P, Ho MT, Ghotbi N. Understanding the acceptance of emotional artificial intelligence in Japanese healthcare system: a cross-sectional survey of clinic visitors’ attitude. Technol Soc. (2023) 72:102166. doi: 10.1016/j.techsoc.2022.102166
138. Doraiswamy PM, Blease C, Bodner K. Artificial intelligence and the future of psychiatry: insights from a global physician survey. Artif Intell Med. (2020) 102:101753. doi: 10.1016/j.artmed.2019.101753
139. Mohammad SM. Ethics sheet for automatic emotion recognition and sentiment analysis. Comput Linguist. (2022) 48:239–78. doi: 10.1162/coli/a/00433
140. Saeidnia HR, Hashemi Fotami SG, Lund B, Ghiasi N. Ethical considerations in artificial intelligence interventions for mental health and well-being: ensuring responsible implementation and impact. Soc Sci. (2024) 13:381. doi: 10.3390/socsci13070381
141. van Kolfschooten H, van Oirschot J. The EU artificial intelligence act (2024): implications for healthcare. Health Policy. (2024) 149:105152. doi: 10.1016/j.healthpol.2024.105152
142. Löchner J, Carlbring P, Schuller B, Torous J, Sander L. Digital interventions in mental health: an overview and future perspectives. Internet Interv. (2025) 40:100824. doi: 10.1016/j.invent.2025.100824
143. Latif S, Ali HS, Usama M, Rana R, Schuller B, Qadir J. AI-based emotion recognition: promise, peril, and prescriptions for prosocial path. arXiv [Preprint]. arXiv:2211.07290 (2022).
144. Ezzameli K, Mahersia H. Emotion recognition from unimodal to multimodal analysis: a review. Inf Fusion. (2023) 99:101847. doi: 10.1016/j.inffus.2023.101847
145. Porter S, Ten Brinke L, Wallace B. Secrets and lies: involuntary leakage in deceptive facial expressions as a function of emotional intensity. J Nonverbal Behav. (2012) 36:23–37. doi: 10.1007/s10919-011-0120-7
146. Sun Q, Akman A, Schuller BW. Explainable artificial intelligence for medical applications: a review. ACM Trans Comput Healthc. (2025) 6:1–31. doi: 10.48550/arXiv.2412.01829
147. Rao A, Aalami O. Towards improving the visual explainability of artificial intelligence in the clinical setting. BMC Digit Health. (2023) 1:23. doi: 10.1186/s44247-023-00022-3
148. Lee MH, Chew CJ. Understanding the effect of counterfactual explanations on trust and reliance on ai for human-AI collaborative clinical decision making. Proc ACM Hum Comput Interact. (2023) 7:1–22. doi: 10.1145/3610218
149. Akman A, Sun Q, Schuller BW. Improving audio explanations using audio language models. IEEE Signal Process Lett. (2025) 32:741–5. doi: 10.1109/LSP.2025.3532218
150. Mauss IB, Butler EA. Cultural context moderates the relationship between emotion control values and cardiovascular challenge versus threat responses. Biol Psychol. (2010) 84:521–30. doi: 10.1016/j.biopsycho.2009.09.010
151. Goodmann DR, Daouk S, Sullivan M, Cabrera J, Liu NH, Barakat S, et al. Factor analysis of depression symptoms across five broad cultural groups. J Affect Disord. (2021) 282:227–35. doi: 10.1016/j.jad.2020.12.159
152. Gur RE, Kohler CG, Ragland JD, Siegel SJ, Lesko K, Bilker WB, et al. Flat affect in schizophrenia: relation to emotion processing and neurocognitive measures. Schizophr Bull. (2006) 32:279–87. doi: 10.1093/schbul/sbj041
153. Hatherley J. A moving target in AI-assisted decision-making: dataset shift, model updating, and the problem of update opacity. Ethics Inf Technol. (2025) 27:1–14. doi: 10.1007/s10676-025-09829-2
154. Evensen J, Røssberg JI, Barder H, Haahr U, Hegelstad WV, Joa I, et al. Flat affect and social functioning: a 10 year follow-up study of first episode psychosis patients. Schizophr Res. (2012) 139:99–104. doi: 10.1016/j.schres.2012.04.019
155. Maleki Varnosfaderani S, Forouzanfar M. The role of AI in hospitals and clinics: transforming healthcare in the 21st century. Bioengineering. (2024) 11:337. doi: 10.3390/bioengineering11040337
Keywords: affective computing, digital mental health interventions, artificial intelligence, affective adaptation, emotion recognition
Citation: Schlicher M, Li Y, Murthy SMK, Sun Q and Schuller BW (2025) Emotionally adaptive support: a narrative review of affective computing for mental health. Front. Digit. Health 7:1657031. doi: 10.3389/fdgth.2025.1657031
Received: 30 June 2025; Accepted: 18 September 2025;
Published: 15 October 2025.
Edited by:
Jennifer Apolinário-Hagen, Heinrich Heine University of Düsseldorf, GermanyReviewed by:
Srishti Goel, Yale University, United StatesShanky Goyal, Chandigarh Engineering College Jhanjeri, India
Copyright: © 2025 Schlicher, Li, Murthy, Sun and Schuller. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Michelle Schlicher, bWljaGVsbGUuc2NobGljaGVyQHR1bS5kZQ==
†These authors have contributed equally to this work