- 1Department of Psychiatry and Neurosciences, Charité – Universitätsmedizin Berlin, Campus Mitte, Berlin, Germany
- 2clare&me GmbH, Berlin, Germany
- 3Department of Psychiatry, Psychotherapy and Psychosomatics, St Joseph Hospital Berlin Weissensee, Berlin, Germany
This study examined the characteristics, motives, expectations, and attitudes of users interested in artificial intelligence (AI) self-help provided by the bot Clare®, a conversational AI for mental health support, and explored the development of a working alliance. A cross-sectional survey of 527 English-speaking self-referred users revealed high levels of anxiety (69%), depression (59%), severe stress (32%), and loneliness (86%). The participants expressed positive attitudes toward digital mental health solutions, with key motives including avoiding embarrassment (36%) and concerns about appearance in face-to-face consultations (35%). Expectations focused on emotional support (35%) and expressing feelings (32%). A strong working alliance was established within 3–5 days (Working Alliance Inventory-Short Report, M = 3.76, SD = .72). These findings highlight the potential of conversational AI in providing accessible and stigma-free support, informing the design of human-centric AI in mental health. Future research should explore long-term user outcomes and clinical large language model integration with traditional mental health services.
1 Introduction
A major barrier to the implementation and scale-up of mental health services is the persistent global shortage of mental health workers, currently estimated at 1.18 million professionals (1, 2). This workforce gap significantly limits service availability, with approximately 85% of individuals with mental health disorders in low- and middle-income countries receiving no care (3). Additional barriers, such as stigma, limited access, and a preference for self-care, deter many from seeking help, while others face delays due to a shortage of mental health professionals, particularly in rural and low-income areas (4, 5). Therefore, self-care interventions that focus on mental health and wellbeing have shifted into focus (6). Meanwhile, artificial intelligence (AI) and machine learning technologies are rapidly advancing in mental healthcare, offering new opportunities for diagnosis and treatment (7–11). Conversational AI (CAI), powered by natural language processing (NLP) and machine learning (ML), shows promise in reaching large, underserved populations (12). CAI can provide psychological care to vulnerable groups, such as the elderly, adolescents, and those avoiding traditional treatments due to financial or social barriers, such as fear of stigmatization (13, 14). Existing industry-driven CAIs, such as Woebot® or Wysa®, targeting depression, anxiety, and stress (12, 15, 16), show promising results in terms of symptom reduction (16) or bonding experience (17).
While early CAIs, such as ELIZA, an early rule-based chatbot developed by computer scientist Joseph Weizenbaum in the 1960s, primarily used (18), newer systems such as Wysa®, Woebot®, and Clare® leverage large language models (LLMs) to enable more flexible and context-aware conversations (19). These models can respond to unanticipated user inputs and adapt dynamically during interactions (20). Most digital solutions for mental health have been struggling with low engagement and high dropout rates (21). Proactive CAI, combining machine learning and LLMs, has potential to allow for an autonomous and digital provision of evidence-based therapeutic interventions that may improve adherence (22). AI, including LLMs, generates flexible and context-dependent responses and thereby mimics therapeutic conversations (23).
Despite their potential, CAIs also present ethical and safety concerns, particularly regarding patient autonomy, trust, and the risk of over-reliance on technology (13, 24). Additionally, research in this field is still scarce and only a very small number of CAIs incorporating LLMs have been tested, such as the industry-based mental health chatbot, Youper, which is based on rule-based and generative AI (25). In particular, advancements in proactive CAI have shifted the focus to topics such as conversational safety, response appropriateness, and AI alignment [see (19, 26) for an extensive overview]. Users' opinions about AI self-help are critical for successful implementation; hence, a research-driven framing of CAI appears to be crucial. The success of these AI tools depends on managing user expectations and addressing these concerns effectively (14). Moreover, we have limited knowledge of the sociodemographics, needs, and preferences of users of conversational AI for self-help, including their age and gender distribution and their expectations and motivations for seeking AI-driven mental health support.
This study explores the demographics, psychological wellbeing, motives, and attitudes of Clare® users, a conversational agent integrating rule-based and generative AI. Further, we assess users’ motives for using and expectations of AI self-help, perceptions of online interventions in general, and bonding effects after initial interactions with Clare®. Notably, in this study, Clare® was implemented as a voice-based conversational agent, enabling spoken interactions rather than solely text-based communication. Our findings aim to inform research-driven expectation management and guide the responsible design of human-centric CAIs.
2 Materials and methods
2.1 Clare®—the intervention itself
2.1.1 Onboarding
Once participation was confirmed, access to Clare® was granted. The participants initiated the onboarding process via a sign-up website. This included an initial onboarding call with Clare® to schedule future calls. Subsequently, the users could trigger on-demand calls with Clare® through a dedicated study website.
2.1.2 CAI-delivered intervention
Clare®, developed by clare&me GmbH, is an AI mental health support agent with emotional intelligence, available as a chat and voice agent on mobile phones. It offers 24/7 anonymous access without installation, simulating human-like, empathetic conversations. Clare® provides content based on cognitive behavioral therapy (CBT), self-compassion, and mindfulness (27–29). In this study, Clare® was used as a voice-only conversational agent.
Clare® integrates CBT techniques, including Socratic dialogue, cognitive restructuring, and framing, using a rule-based model developed by clinical psychologists to match user input with relevant exercises (e.g., meditation for stress). Users can choose the interaction length and mode (verbal or chat). For example, in response to “I feel tired and stressed,” Clare® identifies issues (e.g., sleep difficulties or stress) and provides suitable exercises (e.g., sleep hygiene or stress management). Examples of the content in Clare® are shown in Table 2. The design of Clare® was guided by key ethical principles, including:
• Accountability and transparency: Clearly disclosing the AI's functionalities, establishing boundaries for interaction, and educating users to manage expectations and foster trust.
• Data security and privacy: Adhering to established privacy standards to safeguard user information.
• Human-centered ethics: Aligning with ethical frameworks, such as the American Psychiatric Association (APA) Ethics Code, and prioritizing a human-in-the-loop approach to ensure appropriate oversight in mental health contexts.
2.1.3 Exposure to Clare®
Participants could freely determine their level of interaction with Clare®. However, to be included in the analysis for working alliance, users were required to have had at least one onboarding call (see Table 2 and Section 2.4). To participate in the 8-week study, users were required to complete four questionnaires: at baseline, 3–5 days later (t1b), in week 4 (t2), and in week 8 (t3), and to be exposed to Clare® at least once in each interval between the measurement points. Exposure was defined as a bot-delivered phone call and characterized by the extent of participant interaction, including call frequency and duration. Observed behavioral patterns will not be analyzed in depth in this study but will be examined in a future publication.
2.1.4 Engagement metrics
Engagement was quantified using two metrics: the number of bot-delivered phone calls per week and the average duration of these automated calls. All the call sessions included in the engagement metrics were automated interactions between the users and the voice-based bot. Participants were considered active users if they completed at least one bot-delivered phone call (including the onboarding session) with Clare® within 1 week.
2.1.5 Technological base of Clare®
Clare® is a CAI built on multiple fine-tuned LLMs, developed by clare&me GmbH using various open-source LLMs. Fine-tuning involved collaboration with conversation designers and psychologists to ensure the communication aligns with therapeutic ethical guidelines, focusing on relationship formation, transparency, and limitations. Clare® operates independently, accepting text and voice inputs, with voice transcriptions processed using NLP to extract key information about emotions and context. Clare® engages users to assess their mental state and suggests relevant exercises, offering alternatives if declined. Interaction duration typically ranges from 5 to 45 min, depending on the iterative process and exercise length.
2.1.6 Safety system
Clare® employs a robust safety framework during onboarding and ongoing use, emphasizing automatic crisis detection. A custom moderation application programming interface (API) filters inputs and outputs, flagging inappropriate content before LLM processing. Key safety features include the following.
2.1.6.1 Crisis intervention
If suicidality or severe distress is detected, users are directed to psychological support hotlines and blocked from further use to ensure safety.
2.1.6.2 Conduct enforcement
Users are reminded to interact respectfully and maintain a civil tone.
The safety system ensures appropriate referrals in emergencies and enforces conduct guidelines. A conservative approach prioritizes overflagging, with human moderators reviewing flagged content to minimize false positives.
2.2 Study design and procedure
Here we report cross-sectional data from a longitudinal study set within a larger research frame assessing the feasibility and effectiveness of the voice and chatbot Clare® that is yet to be published. In the Clare® trial, self-report assessments were performed at baseline (t1), after initial contact (t1b), after 4 weeks (t2), and after 8 weeks (t3). Further, we report baseline data from assessment points t1 and t1b for working alliance [Working Alliance Inventory-Short Report (WAI-SR)] (Table 1).
2.2.1 Sample recruitment and data collection
Participants were recruited through English-language advertisements on Facebook, Instagram, LinkedIn, and Google Ads, using language targeting based on users' browser or account settings. No demographic targeting was applied. Digital access and literacy were assessed via questions on internet and mobile phone access, and prior use of Clare® or similar mental health tools. After providing informed consent, participants began the baseline survey (t1). Eligible participants met the inclusion criteria, completed the first survey, and had at least one onboarding call and interaction with Clare®.
2.2.1.1 Inclusion criteria
The participants were 18–65 years old, proficient in English, provided electronic consent and contact information, and had access to a mobile internet device. For the working alliance assessment, participants were required to schedule and complete an onboarding phone call with Clare® (approximately 3–6 min, see Table 2). Continued study participation over 8 weeks required completion of four questionnaires: a baseline questionnaire, a short questionnaire 3–5 days after onboarding with Clare® to assess the working alliance, and two follow-up questionnaires at 4-week intervals (t2, t3). In addition, the participants were required to engage with Clare® at least once between each measurement point. Behavioral engagement was defined as interactions between the users and Clare®, such as scheduling, receiving, and completing phone calls with Clare® (Figure 1).

Figure 1. Flowchart. WAI-SR, Working Alliance Inventory-Short Revised. clare&me logo used with permission from https://www.clareandme.com/.
2.2.1.2 Exclusion criteria
Participants with acute suicidality, identified during onboarding by Clare®'s safety system, were excluded (see Section 2.6.5 for details). Participants who did not successfully onboard with Clare® were excluded from the working alliance analysis (t1b); individuals who did not fulfill the minimum behavioral engagement with Clare® were also excluded from this analysis.
2.3 Ethics approval
The study protocol was approved by the Ethics Committee of Campus Charité Mitte, Berlin, Germany (EA1/109/22).
2.4 Assessment instruments
At baseline (t1), 527 participants were assessed after confirming their eligibility (Figure 1). The online survey, administered via SoSciSurvey between 20 November 2023 and 21 April 2024, included 57 questions plus demographic data to capture participant characteristics and behaviors (Table 1). Although several additional variables were collected as part of the broader exploratory study (Table 1), only those most relevant to the research questions are reported here. Both the 4-item Patient Health Questionnaire (PHQ-4) and 10-item Patient Health Questionnaire (PHQ-D) were chosen to capture a broad range of psychological distress. The PHQ-4 provided a quick screen for anxiety and depression, while the PHQ-D, including the PHQ stress module, offered deeper insight into specific psychosocial stressors, such as relationship difficulties, financial concerns, and work-related stress. This combination ensured efficient screening and a broad assessment of distress, anxiety, and depression in the general population (30). The following assessment instruments were selected to operationalize the constructs of the perception of AI and digital therapy, mental health, bonding, and affective engagement.
2.4.1 Perception of AI and digital therapy
2.4.1.1 Motives and expectations
Participants were asked about their “motives for using Clare” and their “expectations of using Clare,” and multiple answers could be given. Motives and expectations for AI-based mental health support were assessed using 11 predefined options each. The expectation list (e.g., fast and easy access, anonymity, and reduced emphasis on appearance) and motive list (e.g., problem-solving opportunities, gaining clarity, and obtaining information) were adapted from Eichenberg (31), who used them in an online self-help study. These lists were translated from German to English (see Supplementary Tables B.1 and B.2 for all items and abbreviations).
2.4.1.2 Attitudes Toward Psychological Online Interventions (APOI)
The Attitudes toward Psychological Online Interventions Questionnaire (APOI) assesses the attitudes toward psychological online interventions in general (32). The questionnaire includes four subscales: skepticism and risk perception, confidence in effectiveness, technologization threat, and anonymity benefits. The anonymity subscale evaluates users' views on online interventions vs. traditional therapy. Participants respond to 16 statements using a dichotomous scale.
2.4.2 Mental health outcomes
2.4.2.1 PHQ-4
The PHQ-4 assesses depression and anxiety, with items from PHQ-2 and the Generalized Anxiety Disorder scale (GAD-2) rated on a 4-point scale (33). A score of 3 or higher on any subscale indicates a positive screening result, while the total PHQ-4 score (ranging from 0 to 12) classifies psychological distress as none (0–2), mild (3–5), moderate (6–8), or severe (9–12).
2.4.2.2 UCLA Loneliness Scale (3-item version)
The 3-item UCLA Loneliness Scale is a shortened version of the original scale assessing loneliness and social isolation (34). Participants respond to 3 items using a 4-point scale (1 = never, 2 = rarely, 3 = sometimes, 4 = always). The responses are reverse-coded where appropriate, and the final score is calculated by averaging the ratings, with higher scores indicating greater loneliness.
2.4.2.3 PHQ-D stress module (items 121 – 12j in the PHQ-D)
The PHQ stress module, a component of the PHQ-D, consists of 10 items assessing psychosocial stressors such as relationship difficulties, financial concerns, and work-related stress. Participants rate each item on a scale from 0 (not affected) to 2 (severely affected), yielding a total score between 0 and 20. The score reflects the severity of stress: 0–4 (minimal), 5–9 (mild), 10–14 (moderate), and 15–20 (severe) (35).
2.4.2.4 3-item Social Phobia Inventory
The 3-item Social Phobia Inventory (Mini-SPIN) is a brief version of the Social Phobia Inventory assessing fear, avoidance, and physiological distress. Participants rate three items on a 5-point scale (0 = not at all, 4 = extremely), and a cutoff score of 6 or greater has been suggested for social anxiety (36).
2.4.2.5 Satisfaction with Life Scale
The Satisfaction with Life Scale (SWLS) is a brief 5-item measure of global life satisfaction. Items are rated on a 7-point scale (1 = strongly disagree to 7 = strongly agree), yielding a total score between 5 and 35. A score of 20 suggests a moderate level of life satisfaction, with higher scores indicating greater satisfaction (37).
2.4.3 Bonding and affective engagement
2.4.3.1 Working Alliance Inventory-Short Revised
Participants self-reported affective engagement and bonding with the AI using the 12-item WAI-SR. This is a measure of subjective therapeutic alliance, including subscales for bond, goal, and task (38), based on Bordin's (39) working alliance theory. “Therapist” was replaced with “Clare®” to assess the user alliance with the relational agent (Supplementary Table C.1), following prior research approaches on CAI and therapeutic bonding (17, 40). Per the methods of Jasper et al. (41), bond scores of ≥3.45 were considered high, as seen in Darcy et al. (17).
2.5 Data analysis
Data were analyzed using IBM SPSS Statistics Version 28. Descriptive statistics summarized sociodemographic characteristics, clinical symptoms (UCLA, PHQ-D, PHQ-4, SWLS, and Mini-SPIN), and prior experience with digital mental health tools. Attitudes toward online therapy (APOI) were reported as percent agreement on a dichotomous scale (yes/no). Agreement with motives and expectations was assessed with yes/no answers and analyzed by ordering items based on the percentage of affirmative responses.
The t1b sample (n = 348) was descriptively analyzed for working alliance (WAI-SR) and its correlation with loneliness (UCLA). Bivariate Pearson correlation coefficients were calculated to assess the strength and direction of the relationships between working alliance (WAI-SR) and loneliness (UCLA), and with satisfaction with life (SWLS), anxiety and depression (PHQ-4), stress (PHQ-D), and social anxiety (Mini-SPIN). A two-sided independent t-test was conducted to compare WAI-SR scores between women (n = 176) and men (n = 168).
To examine the differences between completers and non-completers, a dropout analysis was conducted. Descriptive statistics (age and sex) were reported for both groups. Given the unequal sample sizes and violation of the homogeneity of variances (as indicated by Levene's tests), Welch's t-tests were used to compare mental health indicators (e.g., depression, anxiety, social anxiety, loneliness, and working alliance) between groups.
3 Results
3.1 Participant characteristics
Of the 604 people screened, 527 met the study criteria and were included in the baseline sample (t1). The mean age was 36.2 years (SD = 9.39; range = 18–64 years) as described in Table 3.
3.1.1 Baseline clinical symptomatology
An overview of participants' scores for the clinical variables (UCLA, PHQ-D, PHQ-4, SWLS, Mini-SPIN) is presented in Table 4. The mean PHQ-4 score was 7.9 (SD = 2.14), with 37% showing a moderate elevation (score = 6–8) and 47.4% exhibiting a severe elevation (score = 9–12) in anxiety and depression symptoms. For 68.7% of participants, the PHQ-4 scores indicated anxiety disorders, while 59.2% showed scores suggesting depressive disorders (≥3 on PHQ-4 subscales). The mean PHQ-D score was 12.6 (SD = 3.31), with 54.1% having moderate psychosocial stress (score 6–8) and 31.9% experiencing severe stress (score 9–12). The mean UCLA score was 6.95 (SD = 1.46), with 85.9% classified as “lonely” (score ≥6). The mean SWLS score was 21.2 (SD = 7.10), with 20.3% being (extremely) dissatisfied (score ≥10) and 30.2% (extremely) satisfied (score ≥26) with their life. The mean Mini-SPIN score was 8.70 (SD = 2.38), with 84.4% at risk for generalized social anxiety disorder (score ≥6).
3.2 Engagement patterns
We explored participant engagement by examining call frequency and duration across both mid-treatment (n = 53) and post-treatment (n = 21) samples. Engagement was tracked over a 4-week period in the mid-treatment phase and an 8-week period in the post-treatment phase. For a detailed overview, refer to Supplementary Tables E.1–E.4.
3.2.1 Call engagement over time
In the mid-treatment sample (n = 53), both the frequency and duration of calls declined over the 4 weeks. The average number of calls in week 1 was 1.77 (SD = 1.52). Call frequency dropped to 0.68 (SD = 0.96) in week 2 and continued to decrease in week 3 to 0.49 (SD = .64) and in week 4 to 0.40 (SD = .57), indicating a reduction in participant engagement over time. A similar decline was observed in call duration. In week 1, the average call length was 3.35 min (SD = 4.37), decreasing to 2.01 (SD = 3.23) in week 2, 1.25 (SD = 2.55) in week 3, and 1.45 (SD = 2.95) in week 4.
3.2.2 Call engagement in the post-treatment sample
Participants in the post-treatment sample (n = 21) showed sustained interaction with Clare® across the full 8-week period. In week 1, the average number of calls was 2.10 (SD = 1.48). Although call frequency declined over time, the participants received approximately one call per week through week 5. A more noticeable drop occurred in weeks 6 and 7, with call frequency at 0.52 (SD = 0.75), followed by a slight recovery in week 8 to 0.57 (SD = .51). Call duration followed a similar trend. In week 1, the average call lasted 3.96 min (SD = 5.10). The duration remained relatively high through week 4 (range: 2.41–3.25 min), declined in weeks 5–7, and rebounded slightly in week 8 to 2.90 min (SD = 4.34).
3.3 Dropout analysis
Comparisons between completers (n = 21) and non-completers (n = 348) showed that non-completers reported significantly higher mental health distress, with most completers being female. Both groups initially formed a strong working alliance with the bot, but non-completers did not maintain engagement. Similar patterns were found when comparing individuals who completed t2 (n = 53) to non-completers (n = 316). A detailed overview of the dropout analysis is provided in Supplementary Material.
3.4 Previous experience and attitudes toward psychological online interventions
Of the 527 participants, only 8 (1.52%) had used other digital mental health tools, and 7 (1.33%) had previously interacted with Clare®. Positive attitudes toward online interventions were generally more prevalent than negative ones, except for crises, where 64.33% (n = 339) favored traditional psychotherapy (Figure 2).

Figure 2. Attitudes toward Psychological Online Interventions Questionnaire (APOI; yes-votes in order of decreasing agreement; t1; n = 527). Items are abbreviated for illustrative reasons. Abbreviations are defined in Supplementary Table A.1. Answers are on a dichotomous scale (yes/no). The selection of multiple motives was possible. Green indicates positive and orange negative attitudes toward psychological online interventions. POI, psychological online interventions; TP, traditional psychotherapy.
3.5 Motives and expectations of AI therapy
The primary motives were avoiding embarrassment when discussing problems with humans (35.7%, n = 188) and receiving advice regardless of appearance (35.29%, n = 186). Lesser motives included reduced commuting (20.5%, n = 108) and anonymity (19.6%, n = 102) (Figure 3). The main expectations were emotional support (35.5%, n = 187) and expressing feelings (32.5%, n = 171). Fewer participants expected referrals to local contacts (14.3%, n = 75) or information on AI self-help (14%, n = 74) (Figure 4).

Figure 3. Motives for seeking AI advice (yes-votes in order of decreasing agreement; t1; n = 527). Items are abbreviated for illustrative reasons. Abbreviations are defined in Supplementary Table B.1. Answers are on a dichotomous scale (yes/no). The selection of multiple motives was possible.

Figure 4. Expectations of AI advice (yes-votes in order of decreasing agreement; t1; n = 527). Items are abbreviated for illustrative reasons. Abbreviations are defined in Supplementary Table B.1. Answers are on a dichotomous scale (yes/no). The selection of multiple motives was possible.
After 3–5 days (t1b), users reported their working alliance with Clare® (WAI-SR; M = 3.76; SD = 0.72; see Table 5). All the subscales of the working alliance—total, goal, task, and bonding—showed positive and significant correlations with loneliness (Table 6). Male participants (n = 168) reported higher mean scores (M = 3.88) than female participants (n = 176, M = 3.65) for overall working alliance. The difference between the groups was small to moderate [t (348) = −3.17, p = 0.002, d = −0.34], as was the difference in the goal [t (348) = −2.40, p = .017, d = −0.26], task [t (348) = −4.09, p < 0.001, d = −0.44], and bonding subscales [t (342) = −2.14, p = 0.033, d = −0.23].

Table 6. Association between loneliness at baseline and bonding experience after 3–5 days of using Clare®.
Furthermore, the total working alliance scores (WAI-SR) showed statistically significant low positive correlations with psychosocial stress (PHQ-D; r = .337, p < .001), anxiety and depression (PHQ-4; r = .368, p < .001), and social anxiety (Mini-SPIN; r = .336, p < .001). A negligible, but still significant, positive correlation was also observed with life satisfaction (SWLS; r = .097, p = .041). Correlation strengths are interpreted according to the classification in a previous study (42).
4 Discussion
The presented study explored the characteristics of individuals interested in conversational AI for self-help, and their psychological wellbeing, motives, and expectations of AI therapy, and the attitudes toward online therapy in general among users of a conversational AI for mental health support (Clare®), and their working alliance and bonding with Clare®.
4.1 Demographic profile of users of CAI
4.1.1 Age and gender
In this study, 53% of participants identified as female, closely aligning with findings by Darcy et al. (17) with 57.5% female participation of self-referred users of the CBT-based conversational agent Woebot® within an age range of 18–78 years. Studies on AI interventions with self-referred samples are scarce, often focusing on college or clinical populations. For example, Fitzpatrick et al. (16) reported an average age of 22.2 years (67% female) among college students, while another study (43) reported a mean age of 22.9 years (70% female) in a clinical sample using the AI Tess®. This reflects broader trends where women are more likely to seek mental health services. Unlike previous CAI studies that focused on young female adults (12, 16, 44), our sample shows a balanced gender distribution and spans a broad age range. This may suggest that chatbots such as Clare® may effectively address mental health issues across genders, potentially reducing stigma and enhancing participation. While both men and women are attracted to AI for self-help, gender-specific approaches could further improve the effectiveness of digital mental health interventions. Additionally, our findings challenge the notion that AI is primarily a solution for Generation Z (45), indicating that AI tools can enhance mental health support for many generations.
4.1.2 Psychological wellbeing of users of Clare® at baseline
The sample exhibited high anxiety and depression levels, comparable to rates in the general populations of the UK, Germany, and the US (46–48). Despite this, no acute crises were detected during onboarding. Moderate psychosocial stress was reported by half of the participants, and severe stress by 32%. Loneliness was significant, affecting 86% of the participants, raising public health concerns (49, 50). Loneliness is recognized as detrimental to health, and while CAIs, including chatbots, have the potential for social support, their impact on loneliness remains under-researched (9), but could be a central therapeutic target of CAIs.
4.1.3 Drop-out and engagement patterns
This study found that individuals with higher levels of distress were less likely to stay engaged with the AI-based support, aligning with the bot's intended scope for moderate distress. While initial working alliances were strong, sustained engagement proved challenging, suggesting the importance of maintaining this connection for continued participation.
Differences in engagement patterns suggest that sustained contact plays a critical role in participant retention. While the mid-treatment sample showed a sharp drop in call frequency and duration after week 1, the post-treatment sample maintained more consistent engagement across all 8 weeks. These findings highlight the importance of early and ongoing interaction in supporting adherence and reducing dropout.
4.1.4 Attitudes toward digital mental health support
We observed high acceptance of psychological online therapy, with 71% finding it effective, useful, and helpful. However, only 3% reported using other digital mental health support. Previous research indicates that favorable attitudes do not always lead to high engagement (51, 52), and initial interest often fades due to the “novelty effect” (53). To be considered effective, interventions should aim for positive long-term effects (54). User engagement with Clare® may be partly influenced by a novelty effect. While we examined whether the working alliance was sustained over time, we did not directly assess this effect.
Future studies should aim to distinguish novelty-driven from sustained engagement, for example, by comparing early and later outcomes, including a familiar control group, or using a mixed methods approach to gain a deeper behavioral understanding—separating novelty-period responses from sustained motivation or engagement (53).
Furthermore, positive attitudes are linked to better implementation and efficacy (55, 56). Therefore, managing expectations, ensuring transparency, and educating users on CAI's limitations and possibilities are vital for sustaining its positive impact. These measures may prevent negative experiences and attitudes that may undermine the positive effects of such interventions. Interestingly, 69% of participants had no confidentiality concerns with online therapy, and 64% found online platforms more secure than traditional methods. While previous guidelines emphasize confidentiality in e-therapy (57), this was less of a concern for Clare® users. Additionally, 62% found it easier to disclose feelings online and were more likely to discuss online therapy with friends than traditional therapy. AI-driven self-help tools may offer a valuable alternative for individuals who find conventional counseling or mHealth interventions inaccessible.
Approximately 40% of the participants worried that online interventions may worsen their loneliness, a concern relevant as many users were already experiencing significant loneliness. This concern reflects broader debates on AI and social isolation; while some see human-AI interactions as a remedy, others warn of increased societal withdrawal (58). Critics suggest that although CAI may temporarily ease isolation, over-reliance could deepen loneliness and alienate individuals from human contact (59), raising concerns about “AI delusion,” and the need for safe AI responses (26). Additionally, 33% questioned the professionalism of digital mental health tools and 64% preferred traditional therapy over online options in crisis, highlighting the need for further research and careful management of expectations of CAI for mental health.
4.2 Motives and expectations of AI self-help
The primary motivations for seeking AI-based mental health support included avoiding face-to-face contact, managing nervousness and embarrassment, and maintaining control. Participants expected that AI would offer emotional support, facilitate self-expression, aid in self-assessment, and improve relationships. Other benefits included increased confidence and knowledge of local support options.
First encounters with therapists can trigger insecurities and feelings of shame (60), a universal emotion linked to perceived inadequacy (61). Shame has been a barrier in therapeutic contexts, particularly for trauma-related cases (62) or male patients (63). Research indicates that AI chatbots can reduce feelings of shame, nervousness, and distress, promoting the realization that seeking help is not shameful (64). Moreover, a recent study showed that there is no difference in self-reported intimacy of self-disclosure between human and chatbot conditions. While chatbots were associated with less fear of judgment, humans were perceived as more trustworthy (65).
This may imply that conversational AI systems should prioritize emotionally safe, non-judgmental environments to facilitate self-disclosure and reduce help-seeking barriers. Despite comparable levels of disclosure to chatbots and humans, lower trust in chatbots underscores the need for transparent design, consistent behavior, and empathetic responses. Incorporating human-in-the-loop mechanisms and human-centered design approaches for bots (23, 66) can enhance oversight and trust. As the exact purpose of bots and the nature of the relationship a user has with them are not always obvious and understood (13), user education and digital literacy are essential for managing expectations and supporting informed use. CAIs may serve as effective adjuncts to traditional care, particularly in early engagement.
Contrary to earlier findings, anonymity and personalized information were less crucial for Clare® users. While anonymity often enhances self-disclosure in digital settings (67), recent studies suggest that perceived emotionality in chatbots can reduce self-disclosure (68). This highlights the potential benefits of AI's limited emotional engagement and artificiality.
4.3 Working alliance with Clare®
Our preliminary findings suggest that users develop a strong bond with Clare® within 3–5 days, indicating an initial positive working alliance. This effect appears particularly pronounced in individuals with high loneliness scores. Moreover, we found that a stronger working alliance was moderately associated with a higher symptom burden. Given the well-documented link between the therapeutic alliance and positive psychotherapy outcomes (69), further research is needed to examine the stability and clinical relevance of this bond over time. This level of bonding is consistent with other conversational agents such as Wysa® and Woebot® (17, 70). Beatty et al. (70) reported a mean WAI-SR score of 3.64 (SD 0.81) among users (n = 1,205), and Darcy et al. (17) observed similar scores among Woebot® users (n = 36,070), with a mean bond subscale score of 3.8 (SD 1.0), where bond scores of ≥3.45 are considered high (41). These working alliance scores are comparable to in-person outpatient psychotherapy (71) and group CBT (41).
Notably, our study found higher bonding scores among male participants (M = 3.9) compared to women (M = 3.7), contrasting with Darcy et al. (17), where higher scores were reported among women. The finding of higher bonding scores among male participants may reflect that men often report greater difficulty talking to mental health professionals, such as psychologists (72). Men may feel more comfortable engaging with AI-driven therapy. Shame linked to a psychotherapeutic dialogue (73) and less fear of perceived judgement in AI interactions (65) may reduce barriers to emotional expression, fostering stronger bonds. Tailoring AI interventions to address gender-specific needs could improve engagement and accessibility, especially for men. Higher loneliness in users correlated with stronger bonding with Clare®, linking AI bonding to clinical variables. This highlights AI's potential to meet emotional needs and emphasizes the importance of considering dependency and unhealthy attachments in users with mental health issues (74).
The early formation of a bond with the AI agent appears promising, but raises important questions about its long-term stability. Notably, early therapeutic alliance ratings have shown limited predictive value, highlighting the need to conceptualize the alliance as a dynamic process that unfolds over the course of treatment. In internet-based cognitive behavioral therapy (ICBT) for depression and anxiety, evidence on the relationship between the alliance and clinical outcomes remains mixed; while some studies report no significant association (75, 76), others found correlations at various stages, including early (77) and mid-treatment (78, 79). These inconsistencies suggest that the function and formation of the therapeutic alliance may differ in digital compared to face-to-face settings. Future research should investigate whether the alliance formed with digital agents is sustained over time and whether it contributes meaningfully to clinical improvement.
In this study, the WAI-SR was adapted by replacing “therapist” with the AI agent “Clare®,” allowing for the assessment of alliance-like perceptions in a digital context. However, this raises concerns about the construct validity of relational constructs, such as “bond,” when interacting with non-human agents. While alliance scores may reflect perceptions of responsiveness or trustworthiness, their comparability to traditional human-delivered therapy is limited. Future research should explore the validity of these constructs in digital interactions.
4.4 General implications of our results for CAI
While Clare® offers promising potential in mental health applications, it is essential to acknowledge the ongoing ethical and safety concerns surrounding AI in therapeutic contexts. Issues such as patient autonomy, over-reliance, and trust remain significant challenges (13, 24, 80). The use of proactive conversational agents capable of topic shifting and generating novel treatment plans (19) presents additional complexities, including ensuring conversational safety, preventing inappropriate responses, and maintaining alignment with therapeutic objectives. Clare® was developed through interdisciplinary collaboration involving data engineers, developers, researchers, psychologists, and conversation designers to address these concerns and key ethical principles (see Section 2.1.2).
The use of AI in caregiving, especially in clinical and therapeutic settings, raises critical ethical and safety concerns about the effectiveness of psychological support and human-AI relationships. These concerns are particularly pronounced for vulnerable groups, such as low-income or minority populations, where emotional dependency on AI could pose substantial risks (81, 82). This highlights the need for stringent ethical standards (14, 83, 84). CAIs, especially those based on LLMs, face significant safety and ethical challenges, including preserving patient autonomy, reducing manipulation risks, and ensuring proper user-technology relationships and privacy (13, 24).
Users' expectations of AI mental health support are not fully understood, raising concerns about AI capabilities and the potential for misunderstanding or deception. Ensuring safety, reliability, risk management, and expectation clarity is crucial (14). Proactive CAI systems may be seen as intrusive, affecting user comfort (66). Ethical issues, such as biased training data (85) and harmful advice (86), can harm user outcomes and erode public trust (23). Clear guidelines and risk assessments are needed to address these concerns and ensure responsible AI use, as demonstrated by models such as Clare®, which incorporates safety measures and human support connections.
4.5 Limitations and future studies
While this baseline data provides valuable insights, several limitations should be considered: the sample was predominantly from Western countries (the UK, Germany, and the US), which limits the generalizability of the findings to other cultural contexts. Attitudes toward mental health, trust in digital technologies, and communication norms vary across cultures (87), influencing user engagement and perceptions of CAI. Misunderstandings about a chatbot's purpose or differing care expectations can also affect engagement and perceived usefulness (13). Socio-cultural and political contexts further shape expectations and concerns about AI. Embedding cultural values is therefore essential (88). Future research should aim to recruit more diverse samples by partnering with international research institutions, translating interventions into multiple languages, and adapting content to be culturally sensitive. These steps help ensure CAIs are effective and acceptable across a wide range of cultural contexts.
The recruitment method may have led to an inherent bias, particularly in terms of digital access, language proficiency, and platform-specific reach. Social media platforms differ in user demographics and usage patterns; to mitigate this bias and foster sample diversity, we deliberately employed a range of platforms to engage participants from varied age groups, gender identities, and educational backgrounds.
Given that only 1.52% of participants had used other digital mental health tools, and only 1.33% had interacted with Clare® previously, the novelty of the platform may have influenced initial perceptions and alliance ratings. Users unfamiliar with such tools may have rated their alliance more positively due to a lack of comparison with other platforms. Future studies should examine how prior experience with digital mental health tools may shape user expectations and alliance development, the sustainability of the working alliance over time, and whether familiarity leads to different patterns of engagement and alliance formation.
Adapting the WAI-SR for AI interactions may affect the interpretability and construct validity of the alliance subscales, particularly the bond dimension. The results should, therefore, be interpreted with caution, and future work should explore whether new alliance measures tailored to digital agents are warranted.
This baseline paper relies on self-reported data, which can be subject to biases such as social desirability and recall bias. Furthermore, self-reported data is limited, as we did not ask for a diagnosis or prior or current treatment. Another limitation of this study is the use of surveys administered every 4 weeks instead of Ecological Momentary Assessment (EMA), which could better minimize recall bias, enhance ecological validity, and capture real-time behaviors in natural environments in future research (89, 90).
Moreover, questions remain about the sustainability of the therapeutic alliance of the CAI with its users over time, which needs further investigation. Given the exploratory nature of this study, participant attrition was expected. Future research should investigate factors contributing to attrition, including demographic and psychological variables, and explore strategies to improve engagement, especially for those with higher distress. A limitation of the study is that expectations towards Clare® were only assessed at the beginning of participation (t1). Therefore, it is not possible to determine whether expectations changed over time or if they were influenced by the actual usage experience. While a comparison of expectations between the completers and dropouts could be made, it remains unclear to what extent these expectations influenced actual usage behavior or dropout.
The scope of this baseline paper is limited to reporting baseline data and initial working alliance, assessed 3–5 days after onboarding, a timeframe consistent with prior studies (16, 70, 79). Interaction data (e.g., call duration and frequency per week) is also included. Follow-up measurements will be presented in a subsequent publication. This will include the mobile Agnew Relationship Measure (mARM), a measure of the therapeutic relationship, which was administered only after the participants had interacted with Clare® (at weeks 4 and 8, i.e., t2 and t3). The mARM assesses experiences with digital mental health interventions. Based on a review of existing research, no studies were found that employed this measure after a short period of interaction. While both the mARM and WAI-SR assess the therapeutic alliance, the WAI-SR has been used in digital mental health studies following brief interactions (3–5 days), supporting its cross-sectional applicability. Given the novelty of this field, a review of the item content led to the conclusion that the mARM requires more prolonged interaction to yield valid measurements and is most meaningful after at least 4 weeks of use, which is a key methodological consideration for future research.
A future paper will explore the users’ interaction data alongside changes in psychological distress and therapeutic alliance over 4- and 8-week periods. Given the small sample size, analyses will be primarily descriptive, supplemented by t-tests to explore potential group differences. Variables of interest will include alliance scores and changes in psychological distress. In addition, subgroup analyses will examine usage frequency and its association with these outcomes, offering preliminary insights into potential patterns of engagement and effect.
5 Conclusions
Users with mental health challenges are attracted to and bond with CAI for self-help, seeking emotional support without concerns about shame or physical appearance. Future research should examine change mechanisms, such as the working alliance, and their impact on the overall wellbeing of users. Addressing ethical considerations, relationship dynamics, and risk management is crucial for effective CAIs. Understanding user expectations and user behavior of CAI will enhance education on AI capabilities and limitations.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving human participants were reviewed and approved by the Ethics Committee of Charité – Universitätsmedizin Berlin, Department of Psychiatry and Neurosciences. All procedures were conducted in accordance with local legislation and institutional requirements. Written informed consent was obtained from all participants prior to their participation in the study.
Author contributions
LS: Conceptualization, Data curation, Methodology, Visualization, Formal analysis, Project administration, Writing – original draft, Writing – review & editing. TK: Formal analysis, Visualization, Writing – review & editing. SK: Writing – review & editing, Supervision.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the clare&me GmbH.
Conflict of interest
LS was employed by clare&me GmbH.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdgth.2025.1576135/full#supplementary-material
References
1. Kakuma R, Minas H, van Ginneken N, Poz MRD, Desiraju K, Morris JE, et al. Human resources for mental health care: current situation and strategies for action. Lancet. (2011) 378(9803):1654–63. doi: 10.1016/S0140-6736(11)61093-3
2. Otu A, Yaya S. Uncovering the collateral impacts of COVID-19 on maternal mental health. Reprod Health. (2022) 19(1):115. doi: 10.1186/s12978-022-01427-5
3. Endale T, Qureshi O, Ryan GK, Esponda GM, Verhey R, Eaton J, et al. Barriers and drivers to capacity-building in global mental health projects. Int J Ment Health Syst. (2020) 14(1):89. doi: 10.1186/s13033-020-00420-4
4. Mojtabai R, Olfson M, Sampson NA, Jin R, Druss B, Wang PS, et al. Barriers to mental health treatment: results from the national comorbidity survey replication. Psychol Med. (2011) 41(8):1751–61. doi: 10.1017/S0033291710002291
5. Thomas KC, Ellis AR, Konrad TR, Holzer CE, Morrissey JP. County-level estimates of mental health professional shortage in the United States. Psychiatr Serv. (2009) 60(10):1323–8. doi: 10.1176/ps.2009.60.10.1323
6. World Health Organization. WHO Guideline on Self-Care Interventions for Health and Wellbeing. Geneva: World Health Organization (2022).
7. D’Alfonso S, Lederman R, Bucci S, Berry K. The digital therapeutic alliance and human-computer interaction. JMIR Ment Health. (2020) 7(12):e21895. doi: 10.2196/21895
8. Koutsouleris N, Hauser TU, Skvortsova V, De Choudhury M. From promise to practice: towards the realisation of AI-informed mental health care. Lancet Digit Health. (2022) 4(11):e829–40. doi: 10.1016/S2589-7500(22)00153-4
9. Li H, Zhang R, Lee Y-C, Kraut RE, Mohr DC. Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being. NPJ Digit Med. (2023) 6:236. doi: 10.1038/s41746-023-00979-5
10. Minerva F, Giubilini A. Is AI the future of mental healthcare? Topoi. (2023) 42(3):1–9. doi: 10.1007/s11245-023-09932-3
11. Torous J, Blease C. Generative artificial intelligence in mental health care: potential benefits and current challenges. World Psychiatry. (2024) 23(1):1. doi: 10.1002/wps.21148
12. Chan WW, Fitzsimmons-Craft EE, Smith AC, Firebaugh M-L, Fowler LA, DePietro B, et al. The challenges in designing a prevention chatbot for eating disorders: observational study. JMIR Format Res. (2022) 6(1):e28003. doi: 10.2196/28003
13. Fiske A, Henningsen P, Buyx A. Your robot therapist will see you now: ethical implications of embodied artificial intelligence in psychiatry, psychology, and psychotherapy. J Med Internet Res. (2019) 21(5):e13216. doi: 10.2196/13216
14. Sedlakova J, Trachsel M. Conversational artificial intelligence in psychotherapy: a new therapeutic tool or agent? Am J Bioethics. (2023) 23(5):4–13. doi: 10.1080/15265161.2022.2048739
15. Durden E, Pirner MC, Rapoport SJ, Williams A, Robinson A, Forman-Hoffman VL. Changes in stress, burnout, and resilience associated with an 8-week intervention with relational agent “woebot.”. Internet Interv. (2023) 33:100637. doi: 10.1016/j.invent.2023.100637
16. Fitzpatrick KK, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (woebot): a randomized controlled trial. JMIR Ment Health. (2017) 4(2):e19. doi: 10.2196/mental.7785
17. Darcy A, Daniels J, Salinger D, Wicks P, Robinson A. Evidence of human-level bonds established with a digital conversational agent: cross-sectional, retrospective observational study. JMIR Format Res. (2021) 5(5):e27868. doi: 10.2196/27868
18. Weizenbaum J. ELIZA—a computer program for the study of natural language communication between man and machine. Commun ACM. (1966) 9(1):36–45. doi: 10.1145/365153.365168
19. Liao L, Yang GH, Shah C. Proactive conversational agents. Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining (2023). p. 1244–7. doi: 10.1145/3539597.3572724
20. Lim SM, Shiau CWC, Cheng LJ, Lau Y. Chatbot-delivered psychotherapy for adults with depressive and anxiety symptoms: a systematic review and meta-regression. Behav Ther. (2022) 53(2):334–47. doi: 10.1016/j.beth.2021.09.007
21. Torous J, Lipschitz J, Ng M, Firth J. Dropout rates in clinical trials of smartphone apps for depressive symptoms: a systematic review and meta-analysis. J Affect Disord. (2020) 263:413–9. doi: 10.1016/j.jad.2019.11.167
22. Yasukawa S, Tanaka T, Yamane K, Kano R, Sakata M, Noma H, et al. A chatbot to improve adherence to internet-based cognitive–behavioural therapy among workers with subthreshold depression: a randomised controlled trial. BMJ Mental Health. (2024) 27(1):e300881. doi: 10.1136/bmjment-2023-300881
23. Stade EC, Stirman SW, Ungar LH, Boland CL, Schwartz HA, Yaden DB, et al. Large language models could change the future of behavioral healthcare: a proposal for responsible development and evaluation. Npj Ment Health Res. (2024) 3(1):12. doi: 10.1038/s44184-024-00056-z
24. Gabriel I, Manzini A, Keeling G, Hendricks LA, Rieser V, Iqbal H, et al. The ethics of advanced AI assistants. arXiv [Preprint]. arXiv:2404.16244 (2024). doi: 10.48550/arXiv.2404.16244
25. Mehta A, Niles AN, Vargas JH, Marafon T, Couto DD, Gross JJ. Acceptability and effectiveness of artificial intelligence therapy for anxiety and depression (youper): longitudinal observational study. J Med Internet Res. (2021) 23(6):e26771. doi: 10.2196/26771
26. Liao L, Yang GH, Shah C. Proactive conversational agents in the post-ChatGPT world. Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval (2023). p. 3452–5. doi: 10.1145/3539618.3594250
27. Eismann G. Therapie-Tools Emotionsregulation. Weinheim: Julius Beltz GmbH & Co. KG (2017). Available at: https://www.beltz.de/fachmedien/psychologie/produkte/details/34545-therapie-tools-emotionsregulation.html (Accessed June 03, 2025).
28. Neff KD. Self-Compassion: theory, method, research, and intervention. Annu Rev Psychol. (2023) 74:193–218. doi: 10.1146/annurev-psych-032420-031047
30. Jung S, Kneer J, Kruger THC. The German COVID-19 survey on mental health: primary results. medRxiv [Preprint]. (2020). doi: 10.1101/2020.05.06.20090340
31. Eichenberg C. Online-sexualberatung: wirksamkeit und wirkweise. Zeitschrift Für Sexualforschung. (2007) 20. doi: 10.1055/s-2007-981238
32. Schröder J, Sautier L, Kriston L, Berger T, Meyer B, Späth C, et al. Development of a questionnaire measuring attitudes towards psychological online interventions-the APOI. J Affect Disord. (2015) 187:136–41. doi: 10.1016/j.jad.2015.08.044
33. Kroenke K, Spitzer RL, Williams JBW, Löwe B. An ultra-brief screening scale for anxiety and depression: the PHQ-4. Psychosomatics. (2009) 50(6):613–21. doi: 10.1016/S0033-3182(09)70864-3
34. Hughes ME, Waite LJ, Hawkley LC, Cacioppo JT. A short scale for measuring loneliness in large surveys. Res Aging. (2004) 26(6):655–72. doi: 10.1177/0164027504268574
35. Löwe B, Spitzer RL, Zipfel S, Herzog W. Gesundheitsfragebogen für Patienten (PHQ-D) (Version 2. Auflage) [Manual Komplettversion und Kurzform] (2002). Available at: https://www.researchgate.net/publication/316586496_B_Lowe_RL_Spitzer_S_Zipfel_W_Herzog_PHQ-D_gesundheitsfragebogen_fur_patienten (Accessed March 01, 2025).
36. Connor KM, Davidson JR, Churchill LE, Sherwood A, Foa E, Weisler RH. Psychometric properties of the social phobia inventory (SPIN). new self-rating scale. Br J Psychiatr. (2000) 176:379–86. doi: 10.1192/bjp.176.4.379
37. Diener E, Emmons RA, Larsen RJ, Griffin S. The satisfaction with life scale. J Pers Assess. (1985) 49(1):71–5. doi: 10.1207/s15327752jpa4901_13
38. Hatcher RL, Gillaspy JA. Development and validation of a revised short version of the working alliance inventory. Psychother Res. (2006) 16(1):12–25. doi: 10.1080/10503300500352500
39. Bordin ES. The generalizability of the psychoanalytic concept of the working alliance. Psychother Theory Res Pract. (1979) 16(3):252–60. doi: 10.1037/h0085885
40. Hoffman V, Flom M, Mariano TY, Chiauzzi E, Williams A, Kirvin-Quamme A, et al. User engagement clusters of an 8-week digital mental health intervention guided by a relational agent (woebot): exploratory study. J Med Internet Res. (2023) 25(1):e47198. doi: 10.2196/47198
41. Jasper K, Weise C, Conrad I, Andersson G, Hiller W, Kleinstäuber M. The working alliance in a randomized controlled trial comparing internet-based self-help and face-to-face cognitive behavior therapy for chronic tinnitus. Internet Interv. (2014) 1(2):49–57. doi: 10.1016/j.invent.2014.04.002
42. Mukaka MM. Statistics corner: a guide to appropriate use of correlation coefficient in medical research. Malawi Med J. (2012) 24(3):69–71.23638278
43. Fulmer R, Joerin A, Gentile B, Lakerink L, Rauws M. Using psychological artificial intelligence (tess) to relieve symptoms of depression and anxiety: randomized controlled trial. JMIR Ment Health. (2018) 5(4):e64. doi: 10.2196/mental.9782
44. Tong F, Lederman R, D’Alfonso S, Berry K, Bucci S. Digital therapeutic alliance with fully automated mental health smartphone apps: a narrative review. Front Psychiatry. (2022) 13:819623. doi: 10.3389/fpsyt.2022.819623
45. Park SY, Do B, Yourell J, Hermer J, Huberty J. Digital methods for the spiritual and mental health of generation Z: scoping review. Interact J Med Res. (2024) 13(1):e48929. doi: 10.2196/48929
46. Jia R, Ayling K, Chalder T, Massey A, Gasteiger N, Broadbent E, et al. The prevalence, incidence, prognosis and risk factors for symptoms of depression and anxiety in a UK cohort during the COVID-19 pandemic. BJPsych Open. (2022) 8(2):e64. doi: 10.1192/bjo.2022.34
47. Kujawa A, Green H, Compas BE, Dickey L, Pegg S. Exposure to COVID-19 pandemic stress: associations with depression and anxiety in emerging adults in the United States. Depress Anxiety. (2020) 37(12):1280–8. doi: 10.1002/da.23109
48. Steffen A, Thom J, Jacobi F, Holstiege J, Bätzing J. Trends in prevalence of depression in Germany between 2009 and 2017 based on nationwide ambulatory claims data. J Affect Disord. (2020) 271:239–47. doi: 10.1016/j.jad.2020.03.082
49. Cacioppo JT, Cacioppo S. The growing problem of loneliness. Lancet. (2018) 391(10119):426. doi: 10.1016/s0140-6736(18)30142-9
50. Mann F, Wang J, Pearce E, Ma R, Schlief M, Lloyd-Evans B, et al. Loneliness and the onset of new mental health problems in the general population. Soc Psychiatry Psychiatr Epidemiol. (2022) 57(11):2161–78. doi: 10.1007/s00127-022-02261-7
51. Borghouts J, Eikey E, Mark G, De Leon C, Schueller SM, Schneider M, et al. Barriers to and facilitators of user engagement with digital mental health interventions: systematic review. J Med Internet Res. (2021) 23(3):e24387. doi: 10.2196/24387
52. Zhao R, Amarnath A, Karyotaki E, Struijs SY, Cuijpers P. Effects of psychological treatment for depression among people not actively seeking help: a meta-analysis. Psychol Med. (2022) 53(2):320–31. doi: 10.1017/S0033291722003518
53. Shin G, Feng Y, Jarrahi MH, Gafinowitz N. Beyond novelty effect: a mixed-methods exploration into the motivation for long-term activity tracker use. JAMIA Open. (2019) 2(1):62–72. doi: 10.1093/jamiaopen/ooy048
54. Cugelman B. Gamification: what it is and why it matters to digital health behavior change developers. JMIR Serious Games. (2013) 1(1):e3139. doi: 10.2196/games.3139
55. Schröder J, Berger T, Meyer B, Lutz W, Späth C, Michel P, et al. Impact and change of attitudes toward internet interventions within a randomized controlled trial on individuals with depression symptoms. Depress Anxiety. (2018) 35(5):421–30. doi: 10.1002/da.22727
56. Soroko E, Borzyszkowski M, Kleka P, Chmielak K, Pawlak Z, Skory M. Attitudes towards psychological online interventions (APOI) in relationship with personality and virtual social integration. Psychometric properties of the APOI questionnaire (2023). doi: 10.31234/osf.io/dmgsc
57. Kanani K, Regehr C. Clinical, ethical, and legal issues in E-therapy. Fam Soc. (2003) 84(2):155–62. doi: 10.1606/1044-3894.98
58. Sample I. Could AI help cure ‘downward spiral’ of human loneliness? The Guardian (2024). Available at: https://www.theguardian.com/technology/article/2024/may/27/could-ai-help-cure-downward-spiral-of-human-loneliness (Accessed March 01, 2025).
59. Haque MDR, Rubya S. An overview of chatbot-based mobile mental health apps: insights from app description and user reviews. JMIR Mhealth Uhealth. (2023) 11:e44838. doi: 10.2196/44838
60. Jacobs L. Shame in the therapeutic dialogue. In: The Voice of Shame. Sandwich, MA: Gestalt Press (1997).
61. Hahn WK. Shame: countertransference identifications in individual psychotherapy. Psychother Theory Res Pract Train. (2000) 37(1):10–21. doi: 10.1037/h0087670
62. DeCou CR, Lynch SM, Weber S, Richner D, Mozafari A, Huggins H, et al. On the association between trauma-related shame and symptoms of psychopathology: a meta-analysis. Trauma Violence Abuse. (2023) 24(3):1193–201. doi: 10.1177/15248380211053617
63. Matos M, Petrocchi N, Irons C, Steindl SR. Never underestimate fears, blocks, and resistances: the interplay between experiential practices, self-conscious emotions, and the therapeutic relationship in compassion focused therapy. J Clin Psychol. (2023) 79(7):1670–85. doi: 10.1002/jclp.23474
64. Romanovskyi O, Pidbutska N, Knysh A. Elomia Chatbot: The Effectiveness of Artificial Intelligence in the Fight for Mental Health (2021).
65. Croes EAJ, Antheunis ML, van der Lee C, de Wit JMS. Digital confessions: the willingness to disclose intimate information to a chatbot and its impact on emotional well-being. Interact Comput. (2024) 36(5):279–92. doi: 10.1093/iwc/iwae016
66. Deng Y, Liao L, Zheng Z, Yang GH, Chua T-S. Towards human-centered proactive conversational agents. Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (2024). p. 807–18. doi: 10.1145/3626772.3657843
67. Joinson AN. Self-disclosure in computer-mediated communication: the role of self-awareness and visual anonymity. Eur J Soc Psychol. (2001) 31(2):177–92. doi: 10.1002/ejsp.36
68. Lee J, Lee D. User perception and self-disclosure towards an AI psychotherapy chatbot according to the anthropomorphism of its profile picture. Telematics Inform. (2023) 85:102052. doi: 10.1016/j.tele.2023.102052
69. Baier AL, Kline AC, Feeny NC. Therapeutic alliance as a mediator of change: a systematic review and evaluation of research. Clin Psychol Rev. (2020) 82:101921. doi: 10.1016/j.cpr.2020.101921
70. Beatty C, Malik T, Meheli S, Sinha C. Evaluating the Therapeutic Alliance with a Free-Text CBT Conversational Agent (Wysa): A Mixed-Methods Study (2022).
71. Munder T, Wilmers F, Leonhart R, Linster HW, Barth J. Working alliance inventory-short revised (WAI-SR): psychometric properties in outpatients and inpatients. Clin Psychol Psychother. (2010) 17(3):231–9. doi: 10.1002/cpp.658
72. Clement S, Schauman O, Graham T, Maggioni F, Evans-Lacko S, Bezborodovs N, et al. What is the impact of mental health-related stigma on help-seeking? A systematic review of quantitative and qualitative studies. Psychol Med. (2015) 45(1):11–27. doi: 10.1017/S0033291714000129
73. Di Sarno M, Fanti E, Perry JC, Madeddu F, Di Pierro R. When psychotherapy runs into shame: a scoping review of empirical findings. J Psychother Integr. (2024) 34(4):463–83. doi: 10.1037/int0000337
74. Loveys K, Fricchione G, Kolappa K, Sagar M, Broadbent E. Reducing patient loneliness with artificial agents: design insights from evolutionary neuropsychiatry. J Med Internet Res. (2019) 21(7):e13664. doi: 10.2196/13664
75. Andersson G, Paxling B, Wiwe M, Vernmark K, Felix CB, Lundborg L, et al. Therapeutic alliance in guided internet-delivered cognitive behavioural treatment of depression, generalized anxiety disorder and social anxiety disorder. Behav Res Ther. (2012) 50(9):544–50. doi: 10.1016/j.brat.2012.05.003
76. Preschl B, Maercker A, Wagner B. The working alliance in a randomized controlled trial comparing online with face-to-face cognitive-behavioral therapy for depression. BMC Psychiatry. (2011) 11(1):189. doi: 10.1186/1471-244X-11-189
77. Hedman E, Andersson E, Lekander M, Ljótsson B. Predictors in internet-delivered cognitive behavior therapy and behavioral stress management for severe health anxiety. Behav Res Ther. (2015) 64:49–55. doi: 10.1016/j.brat.2014.11.009
78. Andersson E, Ljótsson B, Hedman E, Enander J, Kaldo V, Andersson G, et al. Predictors and moderators of internet-based cognitive behavior therapy for obsessive–compulsive disorder: results from a randomized trial. J Obsessive Compuls Relat Disord. (2015) 4:1–7. doi: 10.1016/j.jocrd.2014.10.003
79. Nordgren LB, Carlbring P, Linna E, Andersson G. Role of the working alliance on treatment outcome in tailored internet-based cognitive behavioural therapy for anxiety disorders: randomized controlled pilot trial. JMIR Res Protoc. (2013) 2(1):e2292. doi: 10.2196/resprot.2292
80. Starke C, Ventura A, Bersch C, Cha M, de Vreese C, Doebler P, et al. Risks and protective measures for synthetic relationships. Nat Hum Behav. (2024) 8(10):1834–6. doi: 10.1038/s41562-024-02005-4
81. Scheutz M. The inherent dangers of unidirectional emotional bonds between humans and social robots. (Vol. 50) (2012). Available at: http://choicereviews.org/review/10.5860/CHOICE.50-0844 (Accessed March 01, 2025).
82. Whitby B. Do you want a robot lover? University of Sussex (2011). Available at: https://sussex.figshare.com/articles/chapter/Do_you_want_a_robot_lover_/23343203/1 (Accessed March 01, 2025).
83. Laestadius L, Bishop A, Gonzalez M, Illenčík D, Campos-Castillo C. Too human and not human enough: a grounded theory analysis of mental health harms from emotional dependence on the social chatbot replika. New Media Soc. (2022) 24(1):3–21. doi: 10.1177/14614448221142007
84. Vinchhi N. A survey of emotionally manipulative risks of artificial intelligence (2022). doi: 10.13140/RG.2.2.28957.18402
85. Villongco C, Khan F. Sorry I didn’t hear you.” the ethics of voice computing and AI in high risk mental health populations. AJOB Neurosci. (2020) 11(2):105–12. doi: 10.1080/21507740.2020.1740355
86. Wells K. An eating disorders chatbot offered dieting advice, raising fears about AI in health. NPR (2023). Available at: https://www.npr.org/sections/health-shots/2023/06/08/1180838096/an-eating-disorders-chatbot-offered-dieting-advice-raising-fears-about-ai-in-hea (Accessed March 01, 2025).
87. Naderbagi A, Loblay V, Zahed IUM, Ekambareshwar M, Poulsen A, Song YJC, et al. Cultural and contextual adaptation of digital health interventions: narrative review. J Med Internet Res. (2024) 26:e55130. doi: 10.2196/55130
88. Eke DO, Wakunuma K, Akintoye S. Responsible AI in Africa: Challenges and Opportunities. Cham: Springer International Publishing (2023). doi: 10.1007/978-3-031-08215-3
89. De la Barrera U, Arrigoni F, Monserrat C, Montoya-Castilla I, Gil-Gómez J-A. Using ecological momentary assessment and machine learning techniques to predict depressive symptoms in emerging adults. Psychiatry Res. (2024) 332:115710. doi: 10.1016/j.psychres.2023.115710
Keywords: artificial intelligence, conversational agent, AI voicebot, feasibility, AI-supported psychotherapy, clinical large language model, help-seeking motives
Citation: Schäfer LM, Krause T and Köhler S (2025) Exploring user characteristics, motives, and expectations and the therapeutic alliance in the mental health conversational AI Clare®: a baseline study. Front. Digit. Health 7:1576135. doi: 10.3389/fdgth.2025.1576135
Received: 13 February 2025; Accepted: 19 May 2025;
Published: 13 June 2025.
Edited by:
Xiaomei Liu, Chinese Academy of Sciences (CAS), ChinaReviewed by:
Fangziyun Tong, University of Melbourne, AustraliaNanyu Kuang, National Institutes of Health (NIH), United States
Gagan Jain, Manipal University Jaipur, India
Copyright: © 2025 Schäfer, Krause and Köhler. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Lea Maria Schäfer, bGVhLnNjaGFlZmVyQGNoYXJpdGUuZGU=