Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Digit. Health, 02 December 2025

Sec. Digital Mental Health

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1586668

This article is part of the Research TopicTechnological Advances in Psychiatric Treatments: A Focused Exploration of Human-Computer Interaction (HCI) and Human Factors in Digital TherapeuticsView all 6 articles

Does human support add value to persuasive design-based digital mental health interventions? A propensity score matching study of a digital parenting program

  • Department of Community Mental Health, University of Haifa, Haifa, Israel

Background: While human support has been shown to increase user engagement with digital mental health interventions, it also increases managerial overhead, raises costs, and limits scalability. An alternative approach leverages persuasive design principles to potentially reduce the reliance on human support. Therapeutic persuasiveness (TP) is a concept for persuasive design that involves incorporating features that encourage users to make positive behavior changes in their lives. Prior research suggests that TP features can effectively improve both user engagement and intervention outcomes.

Objective: This study aimed to evaluate the added value of human support in a TP-enhanced digital parent training program (DPT) by comparing engagement and clinical outcomes between human-supported and self-directed intervention formats.

Methods: A propensity score matching approach was used to utilize data from two comparable studies, involving parents of children aged 3–7, all of whom received the same TP-enhanced DPT. One study included a self-directed condition (n = 38), while the other included a human-supported condition (n = 38). Human support was provided via chat and phone calls and included progress acknowledgments, personalized feedback, disengagement follow-up, and timely responses to parent-initiated messages. Engagement patterns and pre-to-post intervention changes in child behavior, parenting practices, and parental self-efficacy were compared between the two intervention formats.

Results: There were no significant differences between the self-directed and human-supported formats in program completion rates (89% vs. 92%, respectively; P = .51), the percentage of parents completing all the modules (81.6% vs. 76.3, P = .57) or total usage time (137 vs. 141 min, P = .14). Parents in the human-supported version logged in significantly more frequently than those in the self-directed group (Cohen's ds = 0.32, 0.34; Ps ≤ .04), which is attributed to parents’ additional engagement in messaging with the supporter. No significant differences were observed between groups in reported improvements in children's behavior problems, parenting practices, or parental self-efficacy (Ps ≥ .17).

Conclusions: These findings suggest that well-designed, technology-enabled intervention features may effectively support program adherence and therapeutic outcomes without requiring additional human support. This study highlights the importance of further research into the relative impact of human-supported vs. self-directed DMHIs and investigating how intervention quality might influence this impact.

Introduction

Digital mental health interventions (DMHI) have the potential to expand access to evidence-based mental health care at affordable rates (13). However, a major challenge in developing effective DMHIs lies in maintaining user engagement over time and ensuring adherence to the therapeutic process as intended (46).

One well-studied approach to improving engagement with DMHIs is the inclusion of human support, delivered by a trained clinician or a coach via chat, phone or email. The mode of support can vary, ranging from simple reminders to complete activities to full therapeutic sessions conducted alongside the DMHI (7). According to the Efficiency Model of Support (8), support encompasses three distinct areas designed to address specific failure points: technical support (addressing usability failures), use support (addressing engagement failures), and clinical support (ensuring that skills and knowledge are effectively applied in daily life). The supporter may receive information from the user or the system and tailor the guidance to meet the user's specific needs at a given time. While some reviews have concluded that human support improves engagement and outcomes [e.g., (3, 9, 10)], others have been less conclusive [e.g., (11)]. In either case, incorporating human support increases the costs of DMHIs delivery, ultimately limiting scalability and accessibility.

A promising approach to maintaining user engagement in self-directed DMHIs is enhancing intervention quality through persuasive system design, which aims to seamlessly tunnel users through the therapeutic process (12). This design approach has been shown to enhance user engagement (1315). One example of such design approach is enhancing the program's “therapeutic persuasiveness” (TP), which refers to the extent to which an intervention's features are designed to encourage users to make positive behavior changes in their lives (16). TP features, such as call to action prompts, task monitoring with feedback, and adaptation to user's state, have been found to predict program usage in real-world use of web-based behavior interventions, accounting for 11%–42% of the variance in program usage within the regression models (5, 17).

Building on these findings, empirical research has investigated whether directly manipulating TP levels impacts DHMI usage and efficacy. A pilot randomized controlled trial (RCT) (16) evaluated the impact of TP features on engagement and outcomes in a digital parent training program (DPT). Parents using a TP-enhanced version demonstrated significantly higher engagement (68.9% vs. 27.9% of parents completed all program modules) and greater improvements in child behavior (Cohen's ds = 0.43, 0.54) compared to those using a basic DPT version. These results raise the following question: Would adding human support to the TP-enhanced version further improve engagement and outcomes, or are TP features alone sufficient to facilitate the therapeutic process, potentially reducing the need for human support?

To address this question, the present study employed a propensity score matching paradigm to compare engagement and outcomes between parents using the TP-enhanced version with and without human support. The analysis utilized two datasets: data from parents who used the self-directed (unguided) version (U-DPT), collected in a previous pilot study (16), and data from parents using the human-supported version (S-DPT), collected in a subsequent follow-up study. Both studies implemented the same DPT, similar recruitment processes and parent populations, with the incorporation of human support being the primary distinguishing factor. The primary objectives of the current study were to estimate effect size differences and evaluate the added value of human support in enhancing engagement and outcomes.

Methods

Study design

This study utilized two datasets collected from separate trials examining a TP-enhanced DPT, implemented with and without human support. The first dataset originated from an RCT comparing a TP-enhanced to a basic version of the DPT (16). The second dataset was obtained from a subsequent study examining a comparable TP-enhanced version that included human support. Both studies employed automated system tracking for usage data collection and gathered outcome measures through self-reported questionnaires administered via Qualtrics at baseline and post-intervention (10 weeks from baseline).

Participants and recruitment procedure

Study protocols were approved by the institutional review board of the University of Haifa (approval numbers: 058/22 for the self-directed intervention; 418/23 for the human-supported intervention). No financial or material incentives were provided for parents to encourage their use of the digital program.

The eligibility criteria for participating parents were: having a child aged 3–7 with an elevated level of behavior problems, and access to a smartphone with an internet connection. Exclusion criteria were: child receiving treatment for behavioral/emotional difficulties or parent enrolled in another parent training program, and child diagnosed with an intellectual disability or developmental delay.

Participants for the first study were recruited between May and July 2022, and for the second study between January 2023 and January 2024, using similar Facebook advertising campaign, targeting parents of children aged 4–8 residing in Israel, without focusing on specific platform subgroups (e.g., parenting groups). The campaign targeted Hebrew-speaking parents, as the program and all study materials were provided in Hebrew. The advertisement consisted of a picture of a parent and children accompanied by a short paragraph outlining the study's aim and the eligibility criteria. Parents who expressed interest were directed to follow a link to the study's web page, where they provided their contact details and completed a preliminary eligibility questionnaire based on exclusion criteria and items related to child behavior. Parents who met the preliminary eligibility criteria were contacted by phone to confirm their eligibility, interest, and understanding of study terms. Interested parents signed a web-based consent form and completed a baseline assessment. Eligibility was then confirmed using Eyberg Child Behavior Inventory (ECBI) subscale scores (ECBI-Problem ≥ 15 or ECBI-Intensity ≥ 132). In the second study, eligibility was further assessed through a remote video interview with an independent clinician (OB), based on the Oppositional Defiant Disorder (ODD) criteria from the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (DSM-5). In total, 45 parents met the eligibility criteria in the first study and were provided access to the U-DPT, while 55 parents met the criteria in the second study and received access to the S-DPT. Six parents dropped out of the first study (6/45, 13.3%; four lost to follow-up, one discontinued participation, and one withdrew due to a medical reason), and seven parents dropped out of the second study (7/55, 12.73%; five lost to follow-up and two discontinued participation). A total of 39 parents who used the U-DPT and 48 parents who used the S-DPT completed the post-intervention assessment and were included in propensity score matching to balance ECBI levels at pre-intervention.

Overview of benevolent parenting intervention

The “Benevolent Parenting” program is based on the principle that parental behaviors and reactions significantly influence a child's behavior (18, 19). The program protocol, designed for completion within two months, was developed by Prof. Baumel, and integrates evidence-based components utilized in common parent training programs and includes seven modules: (1) introduction to parent training; (2) positive interactions and quality time; (3) parental emotion regulation; (4) effective routines and clear ground rules; (5) recognizing positive behaviors and ignoring minor negative behaviors; (6) overcoming disobedience; and (7) mindful parenting and communication between partners (16, 20). Each module began with a learning phase (10–25 min), presenting core concepts through videos, texts, and interactive features, including multiple-choice questions providing immediate feedback and responses to frequently asked questions. Each learning phase concluded with recommendations for practicing the acquired skills. To support skill acquisition, each learning phase was followed by a focusing phase (1–2 week) designed to increase the salience of therapeutic activities in parents' daily lives and facilitate skill development in a nonjudgmental manner. Accordingly, the following TP features were integrated into the focusing phase:

1. Call-to-action messages: Text messages with tips and motivating messages aligned with the relevant module were sent just before childcare pickup hours. For example, parents focusing on “positive interactions and quality time” received motivational messages with ideas for fun activities. Additionally, reminders were sent to re-engage parents who missed daily questionnaires or failed to complete a learning phase within a week. These digital triggers were designed and tailored based on the principles by Muench & Baumel (21).

2. Monitoring and ongoing feedback: Short daily questionnaires assessed parents’ application of newly acquired skills, with automatic, nonjudgmental feedback provided daily and weekly (see Figure 1). The questions were based on the relevant theme being practiced and were designed to help parents effortlessly integrate the associated techniques (22). For example, parents focusing on “effective routines” were asked if they had followed the relevant guidelines with their children that day.

3. Adaptation to user state: Parents could personalize the program based on their responses to questionnaires, with optional modules recommended to address specific needs.


Figure 1
a) A mobile app screen displaying an article titled \“Feedback on positive interactions\” in Hebrew. Below the title is an image of a man and two children blowing bubbles. The text describes ways to foster family time and shared experiences.\n\nb) A mobile survey screen in Hebrew titled \“Daily questionnaire on positive interactions.\” Above is an image of colorful balls. The survey asks how users spend quality time, with options like reading stories and playing games.

Figure 1. Mobile screenshots of the daily monitoring questionnaire (A) and the personalized positive feedback message [(B); text in Hebrew]. Images/icons reproduced from Pixabay (https://pixabay.com/): “Baby Dad Daughter” by ua_Bob_Dmyt_ua, “Dialog Tip” by OpenClipart-Vectors, “Balls Children's Playground” by palichka and “Square Shape” by GraphicMama-team, licensed under Content License.

Program content was created in a content management system that enabled delivery of web-based content, text messages, and email, all triggered by logic-based rules (if–then statements). The program was accessible through a mobile- and computer-compatible website and was delivered via MindTools, an open-source eHealth platform initially developed under the name Serafin. The platform was subsequently adapted and upgraded by the last author (AB) and is available on GitHub (23).

Human support

For the guided DPT (G-DPT), human support followed a coaching manual based on previous work (24, 25) as well as the supportive accountability model and the efficiency model of support (8, 26). The support framework aimed to enhance adherence to the therapeutic process through encouraging program utilization, facilitating skill practice, and providing guidance and problem-solving support.

The supporter was a licensed educational psychologist trained and supervised by the study's PI. Support was provided via a chat feature embedded in the DPT and through phone calls. Support included: (1) initiating contact after the first login to set expectations and offer an optional phone call; (2) sending messages after each learning phase to summarize progress and introduce the upcoming focusing phase; (3) acknowledging efforts during the focusing phase and providing tips and information based on daily questionnaires; (4) offering a phone call after the “overcoming disobedience” phase to clarify key points. To address disengagement (e.g., failing to complete a learning phase within a week or respond to the daily questionnaire for five consecutive days), the supporter contacted the parent through the chat to offer additional support. Furthermore, the supporter responded to all parent-initiated messages within one working day.

Data on user engagement with the human-supporter in the guided version of the program is reported extensively in another paper that is currently under review. Here, we provide only succinct details for brevity, as they may be helpful for the comparisons made in this paper: Parents sent an average of 6.18 messages (SD = 5.93) to the supporter and spent an average of 8.60 min (SD = 11.83) on phone calls with them.

Measures

Study data consisted of self-reported pre- and post-intervention (10 weeks) questionnaires from parents, as well as program usage metrics. Parents completed questionnaires using Qualtrics.

Eyberg child behavior inventory (ECBI)

Children's behavior problems were assessed using two subscales of the ECBI: Intensity and Problem (27). For each of the 36 behaviors, parents rated the intensity on a Likert scale from 1 (never) to 7 (always) and indicated whether the behavior was a problem (0 = no; 1 = yes) (28, 29). The internal consistency in this study (Cronbach's alpha) was .90 for ECBI-Problem and .91 for ECBI-Intensity.

The parenting scale (PS)

The PS (30) assesses parental responses to their child's misbehavior, offering both effective and ineffective options for each hypothetical situation. Parents rated their typical response on a 7-point Likert scale (7 = ineffective; 1 = effective). Two subscales of the PS were used: overreactivity (11 items) and laxness (10 items). Internal consistency was acceptable, with Cronbach's alpha coefficients of .80 for overreactivity, and .89 for laxness.

The parenting tasks checklist (PTC)

The PTC (31) evaluates task-specific self-efficacy, measuring parents' confidence in managing parenting tasks and situations. A shorter version was used for this study, comprising two subscales aligned with the full version: behavioral self-efficacy (6 items, e.g., “Throws a tantrum”) and setting self-efficacy (6 items, e.g., “Shopping with child”). Parents rated each item on a scale from 0 (“Certain I can't do it”) to 100 (“Certain I can do it”). Internal consistency was high with Cronbach's alphas scores of .89 for the setting self-efficacy, and .96 for the behavior self-efficacy.

Alabama parenting questionnaire (APQ) positive parenting practices

Positive parenting practices were evaluated using the APQ Positive Parenting Practices subscale (32), which includes 3 items (e.g., “You make sure your child knows when he or she has done something good”). Parents rated each item on a 5-point Likert scale from 1 (never) to 5 (always), reflecting the frequency of these behaviors at home. The internal consistency score in this study was .88.

Ehealth therapeutic alliance inventory (ETAI)

The ETAI (33) assesses the therapeutic alliance in eHealth interventions using 14 items describing parents' thoughts or feelings towards the DPT (e.g., “Benevolent Parenting focuses on goals that are important to me and helps me achieve them”). Parents rated their level of agreement with each statement on a 7-point Likert scale (1 = Strongly Disagree; 7 = Strongly Agree). The internal consistency score of the scale was .87.

Acceptance (satisfaction)

Parents' attitudes and satisfaction with the DPT were assessed post-intervention. Perceived usefulness (e.g., “The DPT helped me manage my child's misbehavior/become a better parent”) and satisfaction (“I would recommend the DPT to parents of children with behavior problems”) were evaluated using a 5-point Likert scale from 1 (I do not agree) to 5 (I agree). A similar scale has been successfully used in previous studies (1, 24).

Program usage and completion rate

Program usage metrics were collected automatically and included the number of login days, unique logins, total usage time (in minutes), and total time spent using the program (in weeks, from first to last login). Program completion metrics included the percentage of parents who completed the “overcoming disobedience” module (the main obligatory component of the program), percentage of parents who completed the entire program, and completion ratio (number of modules completed divided by the total number of modules in the program).

Statistical analyzes

Descriptive statistics were used to analyze demographic parameters, program usage metrics, and child and parent outcomes. Differences in continuous variables were analyzed using paired-sample t-tests. Differences in dichotomous variables were analyzed using McNemar's test (34).

Comparison of U-DPT vs. S-DPT using propensity score matching

As the U-DPT and the S-DPT studies were conducted in different time periods, it was necessary to control for potential confounders between the two groups. A propensity score matching approach was applied to balance the following covariates: child's age and gender, number of participating parents (e.g., single or both parents), family income, and ECBI scores at pre-assessment (35, 36). Propensity scores were calculated for U-DPT (n = 39) and the S-DPT (n = 48) participants. Pairs were matched using a recommended match tolerance (caliper) of 0.12 (37), selecting the most similar propensity score without replacement. Baseline variables balance was assessed using standardized bias, defined as the mean difference divided by the common standard deviation. Standardized biases for continuous variables were considered acceptable at ≤0.2 and for binary variables at ≤0.1 (37, 38). Statistical significance was defined as p < .05. Exact p-values are reported for all analyses to ensure transparency. All statistical analyses were conducted using IBM SPSS Statistics, version 27.

Results

Overall, 38 pairs of parents met the recommended match tolerance for the propensity scores matching paradigm and were included in the subsequent analysis. Descriptive statistics for parents' socio-demographic characteristics by DPT group are presented in Table 1. The average age of the children was 4.95 years; the average age of the parents was 36.19 years. On average, families had 2.62 children. To reflect the Israeli population and cultural context, household income was reported in Israeli Shekels [ILS; the average household monthly income was approx. 16,500 ILS in 2023, (39)], and religiosity, a salient cultural factor of Israeli population, was included among the demographic characteristics. No significant differences were found in demographic characteristics between the two study groups.

Table 1
www.frontiersin.org

Table 1. Participant demographic characteristics by intervention group at baseline.

Program usage and completion rates by study groups are presented in Table 2. The program completion ratio was .89 for the U-DPT group and .92 for the S-DPT group. No significant differences were observed in the percentage of parents completing the “overcoming disobedience” module (the critical module for addressing behavior problems) or in the percentage of parents completing all modules in their personalized program. However, significant differences were found in the number of login days and unique logins, with S-DPT users logging in more frequently.

Table 2
www.frontiersin.org

Table 2. Program completion and usage metrics.

Differences between self-directed and the human-supported DPT in reported changes following the intervention are reported in Table 3. Both intervention conditions showed significant improvements from pre- to post-intervention in ECBI metrics (Ps ≤ .001) and in parenting measures (.01 ≤ Ps ≤ .001), except for the APQ. However, no significant differences were observed between the U-DPT and S-DPT groups in reported changes.

Table 3
www.frontiersin.org

Table 3. Children and parents’ outcomes, between baseline and post-intervention time points.

Differences in parents' perceptions of the therapeutic alliance and program satisfaction are reported in Table 4. No significant differences were observed between the U-DPT and S-DPT groups.

Table 4
www.frontiersin.org

Table 4. Reported therapeutic alliance and program satisfaction at post-intervention.

Discussion

This study used propensity score matching to compare outcomes between two delivery formats of a DPT: a self-directed and a human-supported version, both incorporating TP features designed to support user adherence through the therapeutic process. The findings revealed comparable engagement patterns, including program completion rates and overall usage time, as well as similar improvements in children's behavior problems, parental behaviors, and parental self-efficacy. Additionally, therapeutic alliance and satisfaction levels were similar among both groups. These results align with prior research suggesting that DMHIs can achieve similar levels of adherence and efficacy regardless of whether they are delivered with or without human support (12, 4042).

However, a significant difference in login frequency was observed, with S-DPT users logging in more frequently than U-DPT users (27.87 vs. 23.10 login days, respectively). This difference may be attributed to parent engagement in messaging with the supporter (as S-DPT participants sent an average of 6.18 messages), which likely influenced their overall login patterns.

A key factor underlying these findings may be the integration of TP features in both versions of the DPT. These features might have influenced parents' adherence by enhancing accountability toward the therapeutic process (26, 43), and by addressing specific failure points, such as low engagement and inadequate implementation of knowledge and tools (8). According to Mohr et al. (26), accountability is strengthened when clear, process-oriented expectations are set, prompting users to justify their actions. In this study, TP features, including daily call-to-action messages and automated feedback following daily questionnaires, could have fostered accountability by building anticipation and ensuring continuous interaction between parents and the program while addressing common implementation barriers. The presence of clear expectations, monitoring, and feedback through TP features may explain the similar adherence patterns observed across both groups, suggesting that the human support protocol (as implemented in this study) may have been redundant. This finding also supports prior propositions that adequate product design can positively impact user adherence (5, 44).

Another related dimension is the therapeutic alliance, measured using the ETAI, which assesses application-induced accountability alongside sense of relatedness to the intervention, perceived emotional investment, and the traditional aspects of the therapeutic alliance nurtured in client–therapist collaboration (11, 45). Therapeutic alliance has been suggested to predict engagement in both human-supported and self-directed DMHIs (46). The similarity in therapeutic alliance scores across both conditions suggests that the limited human support protocol employed (e.g., mostly through chat messages and an average of less than nine minutes of phone calls per parent) did not uniquely enhanced therapeutic alliance, engagement, or outcomes. This finding further highlights the capacity of well-designed automated features to replicate some aspects of human interaction, thereby reducing the necessity for resource-intensive human support.

Limitations and future directions

This study has several limitations that should be noted. First, as a quasi-experiment study with a limited sample size, it lacks both the statistical power and the randomization of participants that a full scale RCT provides. Conducting an RCT of the TP-enhanced DPT, with and without human support, on a larger sample, would allow for a more nuanced examination of the effects of DMHI design quality on the added value of human support. Moreover, the nature of recruitment via social networks introduces the possibility of self-selection bias, which may limit the representativeness of the sample relative to the broader Israeli parent population. Nevertheless, previous research has shown that social media recruitment, particularly through Facebook, can actually reach broader and more diverse populations across socioeconomic backgrounds (47). In line with this, the current sample included participants from diverse income levels and varying degrees of religiosity.

Second, the specific human support protocol used in this study (e.g., its method, frequency, and intensity) may have influenced the results. Prior research suggests that variations in support protocols can significantly impact adherence and outcomes [e.g., (48, 49)]. To generalize the findings, future studies should investigate a broader range of support models to determine which protocols provide the greatest value for money in terms of adherence and outcomes, considering the target populations and therapeutic goals. Third, this study focused on a DMHI designed to support children through their parents. Therefore, further research is needed to determine whether similar findings apply to DMHIs that support adults with various mental health conditions. Finally, future studies could explore the integration of artificial intelligence (AI) and machine learning (ML) into self-directed DMHIs (50). These technologies could facilitate real-time, tailored, and personalized content delivery, addressing users' specific needs while eliminating the cost associated with human support.

Conclusions

This study underscores the importance of investigating the impact of human support vs. self-directed DMHIs on engagement and outcomes, as well as the role of DMHI quality in this context. The findings contribute to the growing body of literature suggesting that well-designed, technology-enabled features may facilitate adherence and enhance therapeutic outcomes. Future research should further investigate how intervention quality influences user engagement and the extent to which human support remains necessary in different contexts.

Data availability statement

The datasets presented in this article are not readily available because these datasets are part of a larger ongoing research project. They can be made available upon specific individual requests. Requests to access the datasets should be directed toYWJhdW1lbEB3ZWxmYXJlLmhhaWZhLmFjLmls.

Ethics statement

The studies involving humans were approved by the institutional review board of the University of Haifa (approval numbers: 058/22 and 418/23). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

CS: Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Validation, Writing – original draft, Writing – review & editing. OB: Conceptualization, Project administration, Writing – review & editing. AB: Conceptualization, Methodology, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was was funded by Israel Science Foundation and The Israel National Institute for Health Policy Research. The funders had no role in the design, data collection, analysis, or preparation of the manuscript.

Conflict of interest

AB has provided consultancy services to iFeel Online.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

The reviewer RCP declared a past co-authorship with the author AB to the handling editor.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Baumel A, Correll CU, Birnbaum M. Adaptation of a peer based online emotional support program as an adjunct to treatment for people with schizophrenia-spectrum disorders. Internet Interv. (2016) 4:35–42. doi: 10.1016/j.invent.2016.03.003

PubMed Abstract | Crossref Full Text | Google Scholar

2. Rogers MAM, Lemmen K, Kramer R, Mann J, Chopra V. Internet-delivered health interventions that work: systematic review of meta-analyses and evaluation of website availability. J Med Internet Res. (2017) 19(3):1–28. doi: 10.2196/jmir.7111

Crossref Full Text | Google Scholar

3. Linardon J, Cuijpers P, Carlbring P, Messer M, Fuller-Tyszkiewicz M. The efficacy of app-supported smartphone interventions for mental health problems: a meta-analysis of randomized controlled trials. World Psychiatry. (2019) 18(3):325–36. doi: 10.1002/wps.20673

PubMed Abstract | Crossref Full Text | Google Scholar

4. Arnold C, Villagonzalo KA, Meyer D, Farhall J, Foley F, Kyrios M, et al. Predicting engagement with an online psychosocial intervention for psychosis: exploring individual- and intervention-level predictors. Internet Interv. (2019) 18(July):100266. doi: 10.1016/j.invent.2019.100266

PubMed Abstract | Crossref Full Text | Google Scholar

5. Baumel A, Kane JM. Examining predictors of real-world user engagement with self-guided eHealth interventions: analysis of mobile apps and websites using a novel dataset. J Med Internet Res. (2018) 20(12):1–13. doi: 10.2196/11491

Crossref Full Text | Google Scholar

6. Baumel A. Therapeutic activities as a link between program usage and clinical outcomes in digital mental health interventions: a proposed research framework. J Technol Behav Sci. (2022) 7(2):234–9. doi: 10.1007/s41347-022-00245-7

Crossref Full Text | Google Scholar

7. Bernstein EE, Weingarden H, Wolfe EC, Hall MD, Snorrason I, Wilhelm S. Human support in app-based cognitive behavioral therapies for emotional disorders: scoping review. J Med Internet Res. (2022) 24(4):e33307. doi: 10.2196/33307

PubMed Abstract | Crossref Full Text | Google Scholar

8. Schueller SM, Tomasino KN, Mohr DC. Integrating human support into behavioral intervention technologies: the efficiency model of support. Clin Psychol Sci Pract. (2016) 24(1):27–45. doi: 10.1037/h0101740

Crossref Full Text | Google Scholar

9. Mohr DC, Azocar F, Bertagnolli A, Choudhury T, Chrisp P, Frank R, et al. Banbury forum consensus statement on the path forward for digital mental health treatment. Psychiatr Serv. (2021) 72(6):677–83. doi: 10.1176/appi.ps.202000561

PubMed Abstract | Crossref Full Text | Google Scholar

10. Day JJ, Sanders MR. Do parents benefit from help when completing a self-guided parenting program online? A randomized controlled trial comparing triple P online with and without telephone support. Behav Ther. (2018) 49(6):1020–38. doi: 10.1016/j.beth.2018.03.002

PubMed Abstract | Crossref Full Text | Google Scholar

11. Boucher EM, Raiker JS. Engagement and retention in digital mental health interventions: a narrative review. BMC Digital Health. (2024) 2(1):52. doi: 10.1186/s44247-024-00105-9

Crossref Full Text | Google Scholar

12. Kelders SM, Bohlmeijer ET, Pots WT, van Gemert-Pijnen JE. Comparing human and automated support for depression: fractional factorial randomized controlled trial. Behav Res Ther. (2015) 72:72–80. doi: 10.1016/j.brat.2015.06.014

PubMed Abstract | Crossref Full Text | Google Scholar

13. Jha S, Jewer S, Agyapong VIO, Orji R. Can your smartphone save a life? A systematic review of mobile-based interventions for suicide prevention. Int J Hum Comput Interact. (2024):1–19. doi: 10.1080/10447318.2024.2323274

Crossref Full Text | Google Scholar

14. Lipschitz JM, Pike CK, Hogan TP, Murphy SA, Burdick KE. The engagement problem: a review of engagement with digital mental health interventions and recommendations for a path forward. Curr Treat Options Psychiatry. (2023) 10(3):119–35. doi: 10.1007/s40501-023-00297-3

PubMed Abstract | Crossref Full Text | Google Scholar

15. Radomski AD, Wozney L, McGrath P, Huguet A, Hartling L, Dyson MP, et al. Design and delivery features that may improve the use of internet-based cognitive behavioral therapy for children and adolescents with anxiety: a realist synthesis with a persuasive systems design perspective. J Med Internet Res. (2019) 21(2):e11128. doi: 10.2196/11128

PubMed Abstract | Crossref Full Text | Google Scholar

16. Baumel A, Brandes O, Brendryen H, Muench F, Kane JM, Saar C. The impact of therapeutic persuasiveness on engagement and outcomes in unguided interventions: a randomized pilot trial of a digital parent training program for child behavior problems. Internet Interv. (2023) 34(October):100680. doi: 10.1016/j.invent.2023.100680

PubMed Abstract | Crossref Full Text | Google Scholar

17. Baumel A, Yom-Tov E. Predicting user adherence to behavioral eHealth interventions in the real world: examining which aspects of intervention design matter most. Transl Behav Med. (2018) 8(5):793–8. doi: 10.1093/tbm/ibx037

PubMed Abstract | Crossref Full Text | Google Scholar

18. Domhardt M, Engler S, Nowak H, Lutsch A, Baumel A, Baumeister H. Mechanisms of change in digital health interventions for mental disorders in youth: systematic review. J Med Internet Res. (2021) 23(11):1–21. doi: 10.2196/29742

Crossref Full Text | Google Scholar

19. Schmidt SJ, Schimmelmann BG. Mechanisms of change in psychotherapy for children and adolescents: current state, clinical implications, and methodological and conceptual recommendations for mediation analysis. Eur Child Adolesc Psychiatry. (2015) 24(3):249–53. doi: 10.1007/s00787-015-0698-0

PubMed Abstract | Crossref Full Text | Google Scholar

20. Saar CR, Brandes O, Baumel A. Identifying active ingredients that cause change in digital parent training programs for child behavior problems: a qualitative exploration. Child Psychiatry Hum Dev. (2024):1–13. doi: 10.1007/s10578-024-01803-1

Crossref Full Text | Google Scholar

21. Muench F, Baumel A. More than a text message: dismantling digital triggers to curate behavior change in patient-centered health interventions. J Med Internet Res. (2017) 19(5):e147. doi: 10.2196/jmir.7463

PubMed Abstract | Crossref Full Text | Google Scholar

22. Baumel A, Muench FJ. Effort-optimized intervention model: framework for building and analyzing digital interventions that require minimal effort for health-related gains. J Med Internet Res. (2021) 23(3):1. doi: 10.2196/24905

Crossref Full Text | Google Scholar

23. MindTools Israeli Upgraded Version of Serafin. Available online at: https://github.com/inonit/serafin/commits/feature/israeli-version (Accessed December 4, 2022).

Google Scholar

24. Baumel A, Tinkelman A, Mathur N, Kane JM. Digital peer-support platform (7Cups) as an adjunct treatment for women with postpartum depression: feasibility, acceptability, and preliminary efficacy study. JMIR Mhealth Uhealth. (2018) 6(2):e38. doi: 10.2196/mhealth.9482

PubMed Abstract | Crossref Full Text | Google Scholar

25. Baumel A, Correll CU, Hauser M, Brunette M, Rotondi A, Ben-Zeev D, et al. Health technology intervention after hospitalization for schizophrenia: service utilization and user satisfaction. Psychiatr Serv. (2016) 67(9):1035–8. doi: 10.1176/appi.ps.201500317

PubMed Abstract | Crossref Full Text | Google Scholar

26. Mohr DC, Cuijpers P, Lehman K. Supportive accountability: a model for providing human support to enhance adherence to eHealth interventions. J Med Internet Res. (2011) 13(1):e30. doi: 10.2196/jmir.1602

PubMed Abstract | Crossref Full Text | Google Scholar

27. Eyberg SM. Eyberg child behavior inventory and Sutter-Eyberg student behavior inventory-revised: professional manual. Phychol Assess Resour. (1999).

Google Scholar

28. Burns GL, Patterson DR, Nussbaum BR, Parker CM. Disruptive behaviors in an outpatient pediatric population: additional standardization data on the Eyberg child behavior inventory. Psychol Assess. (1991) 3(2):202–7. doi: 10.1037/1040-3590.3.2.202

Crossref Full Text | Google Scholar

29. Burns GL, Patterson DR. Conduct problem behaviors in a stratified random sample of children and adolescents: new standardization data on the Eyberg child behavior inventory. Psychol Assess. (1990) 2(4):391–7. doi: 10.1037/1040-3590.2.4.391

Crossref Full Text | Google Scholar

30. Arnold DS, O'Leary SG, Wolff LS, Acker MM. The parenting scale: a measure of dysfunctional parenting in discipline situations. Psychol Assess. (1993) 5(2):137–44. doi: 10.1037/1040-3590.5.2.137

Crossref Full Text | Google Scholar

31. Sanders MR, Woolley ML. Parenting Tasks Checklist. Brisbane: PFSC (2001).

Google Scholar

32. Frick PJ. Alabama Parenting Questionnaire: Psychometric. (1991).

Google Scholar

33. Ashur O, Muench F, Baumel A. Evaluating the contribution of new factors in the assessment of eHealth therapeutic alliance. J Technol Behav Sci. (2023) 9(2):294–305. doi: 10.1007/s41347-023-00336-z

Crossref Full Text | Google Scholar

34. Chen JW, Maldonado DR, Kowalski BL, Miecznikowski KB, Kyin C, Gornbein JA, et al. Best practice guidelines for propensity score methods in medical research: consideration on theory, implementation, and reporting. A review. Arthroscopy. (2022) 38(2):632–42. doi: 10.1016/j.arthro.2021.06.037

PubMed Abstract | Crossref Full Text | Google Scholar

35. Guo S, Barth RP, Gibbons C. Propensity score matching strategies for evaluating substance abuse services for child welfare clients. Child Youth Serv Rev. (2006) 28(4):357–83. doi: 10.1016/j.childyouth.2005.04.012

Crossref Full Text | Google Scholar

36. Rudner LM, Johnette P. Consider propensity scores to compare treatments. Pract Assess Res Eval. (2006) 11(1). doi: 10.7275/kqk5-wa12

Crossref Full Text | Google Scholar

37. Austin PC. An introduction to propensity score methods for reducing the effects of confounding in observational studies. Multivariate Behav Res. (2011) 46(3):399–424. doi: 10.1080/00273171.2011.568786

PubMed Abstract | Crossref Full Text | Google Scholar

38. Rosenbaum PR, Rubin DB. Constructing a control group using multivariate matched sampling methods that incorporate the propensity score. Am Stat. (1985) 39(1):33–8. doi: 10.1080/00031305.1985.10479383

Crossref Full Text | Google Scholar

39. Biton L, Yagur-Kroll A. Israel Figures Selected Data From the Statistical Abstract of Israel (2024). www.cbs.gov.il

Google Scholar

40. Mira A, Breton-Lopez J, Garcia-Palacios A, Quero S, Banos RM, Botella C. An internet-based program for depressive symptoms using human and automated support: a randomized controlled trial. Neuropsychiatr Dis Treat. (2017) 13:987–1006. doi: 10.2147/NDT.S130994

PubMed Abstract | Crossref Full Text | Google Scholar

41. Renfrew ME, Morton DP, Morton JK, Przybylko G. The influence of human support on the effectiveness of digital mental health promotion interventions for the general population. Front Psychol. (2021) 12:716106. doi: 10.3389/fpsyg.2021.716106

PubMed Abstract | Crossref Full Text | Google Scholar

42. Aardoom JJ, Dingemans AE, Spinhoven P, van Ginkel JR, de Rooij M, van Furth EF. Web-based fully automated self-help with different levels of therapist support for individuals with eating disorder symptoms: a randomized controlled trial. J Med Internet Res. (2016) 18(6):e159. doi: 10.2196/jmir.5709

PubMed Abstract | Crossref Full Text | Google Scholar

43. Christensen RE, Ranpariya V, Kammrath LK, Masicampo EJ, Roberson KB, Feldman SR. The presence of accountability in digital interventions targeting non-adherence: a review. Patient Educ Couns. (2022) 105(8):2637–44. doi: 10.1016/j.pec.2022.01.010

PubMed Abstract | Crossref Full Text | Google Scholar

44. Kelders SM, Kok RN, Ossebaard HC, Van Gemert-Pijnen JEWC. Persuasive system design does matter: a systematic review of adherence to web-based interventions. J Med Internet Res. (2012) 14(6):e152. doi: 10.2196/jmir.2104

PubMed Abstract | Crossref Full Text | Google Scholar

45. Beckner V, Vella L, Howard I, Mohr DC. Alliance in two telephone-administered treatments: relationship with depression and health outcomes. J Consult Clin Psychol. (2007) 75(3):508. doi: 10.1037/0022-006X.75.3.508

PubMed Abstract | Crossref Full Text | Google Scholar

46. Tremain H, McEnery C, Fletcher K, Murray G. The therapeutic alliance in digital mental health interventions for serious mental illnesses: narrative review. JMIR Ment Health. (2020) 7(8):e17204. doi: 10.2196/17204

PubMed Abstract | Crossref Full Text | Google Scholar

47. Thornton L, Batterham PJ, Fassnacht DB, Kay-Lambkin F, Calear AL, Hunt S. Recruiting for health, medical or psychosocial research using Facebook: systematic review. Internet Interv. (2016) 4:72–81. doi: 10.1016/j.invent.2016.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

48. Connell AM, Stormshak EA, Mauricio AM, Hails KA, Ramirez-Miranda J, Inyangson JI. A digital health model for school-based implementation to improve parent and child outcomes: comparison of active versus light-touch coaching effects. J Prev. (2024) 45(6):927–45. doi: 10.1007/s10935-024-00806-x

PubMed Abstract | Crossref Full Text | Google Scholar

49. Hadjistavropoulos HD, Schneider LH, Edmonds M, Karin E, Nugent MN, Dirkse D, et al. Randomized controlled trial of internet-delivered cognitive behaviour therapy comparing standard weekly versus optional weekly therapist support. J Anxiety Disord. (2017) 52:15–24. doi: 10.1016/j.janxdis.2017.09.006

PubMed Abstract | Crossref Full Text | Google Scholar

50. Cevasco KE, Morrison Brown RE, Woldeselassie R, Kaplan S. Patient engagement with conversational agents in health applications 2016–2022: a systematic review and meta-analysis. J Med Syst. (2024) 48(1):40. doi: 10.1007/s10916-024-02059-x

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: human support, persuasive design, user engagement, parent training, mental health, child behavior problems

Citation: Saar CR, Brandes O and Baumel A (2025) Does human support add value to persuasive design-based digital mental health interventions? A propensity score matching study of a digital parenting program. Front. Digit. Health 7:1586668. doi: 10.3389/fdgth.2025.1586668

Received: 3 March 2025; Accepted: 20 October 2025;
Published: 2 December 2025.

Edited by:

Khaldoon Dhou, Texas A&M University Central Texas, United States

Reviewed by:

Rüdiger Christoph Pryss, Julius Maximilian University of Würzburg, Germany
Jennifer Wright-Berryman, University of Cincinnati, United States

Copyright: © 2025 Saar, Brandes and Baumel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chen R. Saar, Y3NhYXJAY2FtcHVzLmhhaWZhLmFjLmls

ORCID:
Chen R. Saar
orcid.org/0009-0002-0527-0045

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.