Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Digit. Health, 18 June 2025

Sec. Connected Health

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1621293

Beyond the interface: benchmarking pediatric mobile health applications for monitoring child growth using the Mobile App Rating Scale


Anggi Septia Irawan
Anggi Septia Irawan1*Arie Dwi Alristina,Arie Dwi Alristina2,3Rizky Dzariyani LailiRizky Dzariyani Laili3Nuke AmaliaNuke Amalia4Arief Purnama MuharramArief Purnama Muharram5Adriana Viola MirandaAdriana Viola Miranda6Bence DbrssyBence Döbrössy1Edmond Girasek
Edmond Girasek1
  • 1Institute of Behavioral Sciences, Semmelweis University, Budapest, Hungary
  • 2Health Sciences Division, Doctoral College, Semmelweis University, Budapest, Hungary
  • 3Sekolah Tinggi Ilmu Kesehatan Hang Tuah, Surabaya, Indonesia
  • 4Safety and Health Engineering Study Program, Politeknik Perkapalan Negeri, Surabaya, Indonesia
  • 5HealthAI Indonesia, Jakarta, Indonesia
  • 61000 Days Fund, Denpasar, Indonesia

Introduction: As mHealth applications become increasingly adopted in Indonesia, it is crucial to assess their quality and usability for parents and healthcare professionals.

Aim: This study evaluated the quality of pediatric-related mobile health (mHealth) applications available in Indonesia, focusing on their ability to support child growth monitoring and provide educational resources for parents and caregivers.

Methodology: This is a cross-sectional study. From December 1, 2024, and January 31, 2025 we conducted systematic search for pediatric mHealth applications in Indonesian Google Play Store and Apple App Store using predetermined keywords. Inclusion criteria required the applications to be available in Bahasa Indonesia, focus on child health, and include growth tracking or stunting prevention features. We excluded applications that were not functioning during the testing period. Quality assessment was conducted by five healthcare professionals using the Mobile App Rating Scale (MARS). MARS assessed applications from multiple domains, including engagement, functionality, aesthetics, and information quality. Inter-rater reliability was ensured using the Intraclass Correlation Coefficient (ICC). The results were analyzed using descriptive statistics, Pearson's correlation, and T-tests. A p-value of <0.05 is considered to be statistically significant.

Findings: Nine applications were included in this study. Seven of the applications (77.78%) focused on tracking child growth and development and providing educational content. Less than half of the apps had built-in community features that enabled social support (n = 4, 44.44%) and features for feedback mechanisms & personalized guidance (n = 3, 33.33%) respectively. The majority were developed by commercial companies (n = 7, 77.78%). Quality assessment found significant variability across the apps, with high functionality and aesthetics scores but more variability in the domains of app engagement, quality of information, and subjective quality or perceived value.

Conclusion: This research underscored the need for the development of higher-quality, evidence-based mHealth apps for pediatric care in Indonesia, particularly in improving user engagement, feedback mechanisms and accessibility.

1 Introduction

The rapid advancement of mobile health (mHealth) technologies has significantly transformed healthcare delivery worldwide, particularly in pediatric care (1). MHealth applications offer innovative solutions for monitoring child growth, tracking developmental milestones, and providing health education to parents and caregivers (2). These digital tools have been shown to enhance parental knowledge (3), improve care seeking behavior (4), and support early childhood development interventions (4).

Smartphone penetration is projected to grow steadily between 2024 and 2029, with estimates suggesting it will reach 97 percent by 2029, an increase of over 15 percentage points (5). This consistent rise in smartphone adoption highlights the expanding potential of mobile platforms to support public health interventions (6). In this context, digital health solutions, particularly mobile applications, are increasingly recognized as valuable tools to facilitate early detection and intervention efforts, especially in areas such as child growth monitoring and early childhood development (7).

Studies suggested that mHealth apps can improve parental awareness, increase adherence to immunization schedules, and improve nutritional monitoring (8). Several pediatric-focused mHealth apps are available in Indonesia, including PrimaKu, Asianparent, and Tentang Anak, which offer features such as child growth monitoring, vaccination tracking, and parental education. However, the effectiveness of these applications depends on their usability, engagement, and the quality of the information they provide (9).

Assessing their quality using a standardized framework, such as the Mobile App Rating Scale (MARS), is essential to ensure they meet the needs of Indonesian parents and healthcare professionals (10). Despite the increasing adoption of mHealth applications, challenges remain in ensuring accessibility, credibility, and sustained user engagement (11). Many existing apps focus on tracking and information dissemination but lack of interactive features, such as feedback mechanisms, goal-setting, and social support, which are crucial for long-term user engagement (12). The user version of the MARS framework has been developed to incorporate user perspectives in the assessment process, which is essential for creating effective mHealth resources (13). Additionally, by integrating behavior change techniques into apps development can enhance alignment with users' desired health outcomes (14).

This study evaluated pediatric-related mHealth applications available in Indonesia, analyzing their quality using the MARS framework. By assessing key attributes such as engagement, functionality, aesthetics, and information quality, this research aimed to identify strengths and gaps in the current landscape of pediatric mHealth applications by MARS.

2 Methods

2.1 Study design

This study was designed as a cross-sectional analysis of mobile applications related to baby growth tracking, available in Indonesian App Stores. No regulatory approval was required for this study. The research was conducted under the Strengthening the Reporting of Observational studies in Epidemiology (STROBE) guidelines.

2.2 Raters selection

Five healthcare professionals were selected as raters (The data shown in Supplementary Table S1) based on the following criteria:

Inclusion criteria:

(i) Healthcare professionals or lecturers in the health sector, and/or (ii) Actively engaged in clinical practice in Indonesia.

Exclusion criteria:

(i) Not owning a mobile phone, (ii) Unable to download apps from the Apple App Store or Google Play Store, and (iii) Having hearing, visual, or motor impairments that could hinder participation.

2.3 Selection of the pediatric-related mHealth apps

Researchers selected mobile applications related to pediatric care and baby growth tracking between December 1, 2024, and January 31, 2025. The search was conducted in both the Indonesian App Store (iOS) and the Indonesian Google Play Store (Android). The keywords used for this search include: “pediatrik” (pediatric), “kesehatan bayi” (baby health), “stunting” (stunted growth), “pertumbuhan bayi” (baby growth), and “mengasuh anak” (parenting). Since the App Store and Google Play Store do not support the use of truncation or logical operators (AND, OR, NOT), each search term was entered separately.

To refine the selection, each researcher independently removed duplicate apps from the same app stores (iOS or Android). Subsequently, they compiled a unified list of apps available on both platforms to ensure accessibility for all users. After comparing their lists to verify completeness, the researchers downloaded the remaining apps to their devices and applied the inclusion criteria:

(1) The application must be available in Bahasa Indonesia,

(2) It must focus on pediatric care, and

(3) It must include at least one feature of baby growth tracking or stunting prevention features.

2.4 Evaluation of the pediatric-related mobile apps

2.4.1 The use of standardized rating scale for mobile applications

This study utilized the original English version of the MARS for evaluation. The first component of MARS, known as “App Classification,” was assessed by two academic researchers. MARS is specifically designed to evaluate health-related mobile applications and consists of a primary section with 23 items divided into five categories (A, B, C, D, and E), along with an additional section (F) containing six items (The data shown in Supplementary Table S2). Each item on the MARS scale is rated on a 5-point Likert scale, A score of 1 indicates poor quality, while a score of 5 signifies high quality (10, 15).

2.4.2 Evaluation of the apps by raters

The app evaluation was conducted by five academic health researchers. Before assessing pediatric care-related applications, the raters underwent training to familiarize themselves with the MARS. To ensure a consistent understanding of the MARS criteria, all raters participated in discussions to standardize their evaluation approach.

As part of the training, a test assessment was conducted using an app that was not included in the study sample. Each rater independently evaluated Halodoc, an app focused on general healthcare services rather than exclusively on pediatric care. The raters downloaded the app, tested its features for at least 15 min, and then completed the MARS assessment. Afterwards, they compared their scores. If any individual rating differed by 2 points or more, further discussion was held until a consensus was reached, ensuring uniformity in the evaluation process.

The formal assessment of pediatric care-related apps took place in February 2025. Each of the five raters downloaded and used each selected app for 15 min before completing the standardized online MARS questionnaire. During this evaluation, one application, Miki Anthropometri, was found to be unavailable and was subsequently excluded from the study.

2.5 Statistical methods

2.5.1 Intraclass Correlation Coefficient (ICC)-raters

Ensuring the consistency of Mobile Application Rating Scale (MARS) scores across different raters, studies have employed the Intraclass Correlation Coefficient (ICC) as the primary method for assessing inter-rater reliability. The ICC is widely used for evaluating the reliability of measurements involving multiple assessors, particularly when working with ordinal or continuous data (16, 17).

A two-way random effects model with absolute agreement was chosen because it is specifically designed for situations where multiple independent raters provide evaluations, and the focus is on achieving absolute agreement rather than just relative consistency. This model effectively accounts for systematic differences among raters as well as measurement error, ensuring a more accurate assessment of reliability (17).

Since MARS scores are measured using a five-point Likert scale, the ICC serves as a robust indicator of variability across different evaluators while maintaining statistical precision. The reliability values were interpreted based on Cicchetti's classification (18). Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. In this study, ICC values ranged from 0.80 to 0.95, indicating a high level of agreement among the raters. These results confirmed strong inter-rater reliability, ensuring that the app evaluations were consistent and reproducible (19).

2.5.2 Descriptive analysis

Descriptive analysis was used to describe the frequency distribution of app characteristics and research variables. The app characteristics are presented based on the theoretical background and strategy, affiliation, and technical aspects of the apps. The descriptive analysis in this study provided data on MARS mean scores and MARS scores for subcategories for each mobile health app, as well as boxplots to display the spread and distribution of scores.

2.5.3 Statistics methods analysis

Pearson's correlation was used to analyze the correlation between apps and MARS variables (engagement, functionality, aesthetics, and information) with a p value of <0.05. An unpaired t-test was conducted to determine whether there was a statistically significant difference between the means of two independent groups, with a p value of <0.05. In this case, the total MARS mean scores of commercial and non-commercial apps were compared to assess whether commercial apps significantly outperformed non-commercial ones. The t-test helped determine whether the observed differences in mean scores were due to a genuine underlying effect or merely random variation.

2.5.4 Heatmap visualization: method and justification

The patterns and variations in MARS were identified by scores across different applications and rating categories, a heatmap visualization was implemented (20). This method was selected for its ability to intuitively compare the relative performance of various applications across multiple dimensions, including engagement, functionality, aesthetics, information quality, and subjective quality.

The heatmap's color scheme followed a gradient from cool to warm tones, reflecting the magnitude of MARS scores. Cooler colors, such as blue and green, indicate lower scores, while warmer hues, like yellow, orange, and red, represent higher scores. Through this approach, key trends such as consistently high-performing apps, areas requiring improvement, and variations across different rating sections can be effectively identified.

Furthermore, the heatmap aided in recognizing consistency patterns across different evaluation criteria. For example, an application that scores consistently high across all categories will display predominantly warm tones, whereas one with mixed performance will exhibit a more varied color distribution.

3 Results

3.1 Selection of the pediatric care-related mobile apps

A total of 13 apps in the App Store and 32 apps in the Google Play Store were identified (The data shown in Figure 1). The duplicates were eliminated in each list. The two lists were checked, analyzing the name of the app and the developer. 10 apps were available and selected on both systems. After downloading, one app was excluded because it was not functioning during the assessment. And only nine apps were included after screening.

Figure 1
Flowchart detailing app selection process. Identification: 523 apps from Playstore (310) and App Store (213). After removing 450 duplicates, 73 apps screened for eligibility. Screening: 63 did not meet criteria. 10 apps screened via installation, 1 unavailable. Study: 9 apps included.

Figure 1. Flowchart of the pediatric care mobile apps selection.

3.2 General characteristics of the pediatric-related mobile apps

Among the nine analyzed mobile health applications (The data shown in Figure 1) for pediatric care, the majority (77.78%) incorporated monitoring and tracking features, highlighting their primary function as tools for continuous health data collection (The data shown in Supplementary Table S3). Similarly, from Figure 2 shown information and education components were present in 80% of the apps, indicating a strong emphasis on knowledge dissemination to caregivers and healthcare providers. Assessment capabilities were also observed in 80% of the apps, allowing for evaluations of child development and health status. In addition, Advice, tips, strategies, and skills training were also included in 60% of the apps, providing users with guidance on best practices for child health and nutrition. However, fewer apps included feedback mechanisms (40%) or goal-setting features (40%), suggesting limited interactive elements for user engagement and motivation.

Figure 2
Bar chart displaying the percentage usage of various strategies. Assessment is at 80%, Information/Education and Goal Setting are at 70%, Feedback at 60%, Monitoring/Tracking at 50%, and Advice/Tips/Strategies at 30%.

Figure 2. Apps characteristics based on theoretical background and strategies.

Based on the affiliation (The data shown in Figure 3), most of the applications (77.78%) were commercially affiliated, while only one (11.11%) was developed by a government entity, and another (11.11%) was linked to the academic sector. The dominance of commercial applications suggested that the market for pediatric mobile health tools is largely driven by private sector initiatives rather than public health or academic research.

Figure 3
A donut chart shows percentages of different sectors: Commercial at 77.8%, Government at 11.1%, and Academia at 11.1%. The Commercial sector is the largest segment, marked in dark blue.

Figure 3. Apps characteristics based on affiliation.

Regarding technical functionalities (The data shown in Figure 4), 70% of the apps provided reminder notifications, aiding users in maintaining consistent monitoring and engagement. However, only 20% allowed content sharing on social media platforms such as Facebook, which may limit peer support and community engagement. The majority (90%) required users to log in, possibly for data security and personalization purposes. Additionally, 40% featured an app-based community, which can enhance user engagement through peer interaction and support.

Figure 4
Bar chart showing four features with their respective percentages. \

Figure 4. Apps characteristics based on technical aspects of the apps.

3.3 Comparison of MARS score

Four raters evaluated and rated the 9 apps analyzed (The data shown in Supplementary Table S4). The inter-rater agreement between the two raters was considered good, with Kendall's coefficient of concordance value of 0.93 and a p-value of 0.03.

The heatmap visualization highlighted several key insights (The data shown in Figure 5). Among the high performers, Asianparent, and Tentang Anak stood out with consistently high scores across all sections. Their Engagement, Functionality, and Aesthetics scores were perfect (5.0), making them the most well-rounded apps. On the other hand, PSG Balita had a significantly low Engagement score of 1.40, despite achieving a high Functionality score of 5.00. In addition, Astuti had the lowest Subjective Quality score at 1.50, indicating notable shortcomings in user perception. Examining overall trends, Functionality appeared to be the most consistent category, with several apps scoring close to 5.0. However, Engagement scores varied widely, with some apps excelling while others struggling. Furthermore, Information scores were generally lower than other sections, suggesting a need for more credible and high-quality content.

Figure 5
Heat map displaying app scores across MARS sections: Engagement, Functionality, Aesthetics, Information, Total MARS Mean, and Subjective Quality. Apps include Primaku, Asianparent, PSG Balita, and others. Scores range from 1.4 to 5.0, with color gradients from blue (low) to red (high).

Figure 5. Heatmap visualization of comparison MARS mean score.

The descriptive statistics of the MARS scores revealed notable trends across different evaluation sections. The Engagement section had a mean score of 3.80, but with a high standard deviation of 1.17, indicating significant variability among apps. Some applications, such as Asianparent and Tentang Anak, achieved perfect scores of 5.0, whereas PSG Balita scored the lowest at 1.40, reflecting a lack of interactive and engaging features. In contrast, the Functionality section demonstrated the highest consistency, with an average score of 4.61 and a low standard deviation of 0.33. This has shown that most of the apps provide a well-structured user interface with reliable performance and ease of navigation.

The Aesthetics scores exhibited moderate variability, with a mean of 4.07 and a standard deviation of 0.62. While apps like Asianparent and Tentang Anak scored 5.0, others, such as PSG Balita, scored notably lower at 3.33, indicating some inconsistency in visual design and stylistic appeal. The Information section had an average score of 3.99, slightly lower than the other categories, with a standard deviation of 0.78. These findings suggested that while some apps provide high-quality, evidence-based information, others might require improvements in content credibility and comprehensiveness.

The Total MARS Mean Score across all applications was 4.12, reflecting generally positive performance; however, individual app scores ranged from 3.18 to 4.89, indicating some variability in overall quality. Lastly, the Subjective Quality category, which represents users' perceived value of the apps, had the lowest mean score of 3.33, with the highest standard deviation of 0.98. These findings suggested a wide range of user experiences, with some apps being highly rated while others, such as Astuti with 1.50, facing significant user-perceived shortcomings.

The boxplot (The data shown in Figure 6) visualization illustrated these variations, highlighting how Functionality remained consistently high across apps, while Engagement and Subjective Quality showed substantial fluctuations. The histogram distributions reinforced these findings, showing that while Functionality and Aesthetics tended to cluster around higher values, Engagement and Subjective Quality displayed more diverse patterns, suggesting areas where improvement is needed.

Figure 6
Box plot comparing scores of five categories: Engagement, Functionality, Aesthetics, Information, Total MARS Mean, and Subjective Quality. Scores range from approximately 3.5 to 4.5, with varied medians and distributions.

Figure 6. Boxplot of MARS scores across sections.

3.4 Correlation matrix MARS categories

The statistical analysis revealed key insights into the variability and relationships among different MARS categories (The data shown in Figure 7). Engagement showed the highest variability, with a standard deviation of 1.17, indicating that user experience differed significantly across apps in terms of interactivity and appeal. In contrast, Functionality had the lowest variability (0.33), suggesting that most apps performed consistently well in usability and navigation. Aesthetics scores exhibited moderate variation (0.62), reflecting differences in visual design quality, while Information Quality also showed moderate variability (0.77), indicating that credibility and the depth of health information vary across apps.

Figure 7
Correlation matrix heatmap with variables: Engagement, Functionality, Aesthetics, and Information. Colors range from blue (negative correlation) to red (positive correlation). Notable correlations include Engagement and Information at 0.89, Aesthetics and Engagement at 0.81, and Functionality and Aesthetics at 0.31.

Figure 7. Correlation matrix MARS categories.

The correlation analysis highlighted a strong positive correlation (0.89) between Engagement and Information Quality, suggesting that apps that successfully engage users also tend to provide high-quality and credible health information. However, the weakest correlation (−0.33) was observed between Functionality and Information Quality, indicating that a well-functioning app does not necessarily offer reliable medical content. These findings underscored the importance of balancing technical usability with medical accuracy, as strong functionality alone did not guarantee trustworthy health information. Developers should focus on enhancing both user experience and content credibility to ensure effective and reliable mobile health apps.

3.5 Behavioral outcomes result

Figure 8 showed the average score per behavioral category across all apps. Highest scoring categories included Awareness and Help Seeking (both averaging ∼3.78). Lowest scoring: Behavior Change (∼2.89), indicating that this was the area where most apps had the lowest scores (Data shown in Supplementary Table S5).

Figure 8
Radar chart comparing mobile applications in categories such as awareness, knowledge, attitudes, intention to change, help seeking, and behavior change. Applications include Primaku, Kms Bunda dan Balita, Asianparent: Kehamilan & Bayi, and others. Each application's performance in these categories is represented by different colored layers.

Figure 8. Average MARS behavioral outcomes by category.

The analysis of behavioral outcomes among mobile health apps highlighted significant variations in their effectiveness. The Asianparent and Tentang Anak consistently achieved the highest scores across all categories (The data shown in Figure 9), demonstrating strong performance in raising awareness, enhancing knowledge, shaping positive attitudes, encouraging intention to change, promoting help-seeking behavior, and driving actual behavior change. Imuni, Primaku, KMS Bunda dan Balita, and Teman Bumil showed relatively strong results, particularly in help-seeking and behavior change, but did not reach the top-tier performance as Asianparent and Tentang Anak. On the other hand, PSG Balita and Astuti exhibited the weakest influence, with lower scores in key areas such as attitudes, intention to change, and behavior change, suggesting limited effectiveness in fostering health-related improvements. These findings indicated that while certain apps successfully drive user engagement and behavioral change, others might require enhancements in content, usability, and intervention strategies to maximize their impact.

Figure 9
Bar chart showing percentage popularity of health and parenting apps. \

Figure 9. Average MARS behavioral scores by apps.

3.6 Commercial vs. non-commercial apps

An independent t-test was conducted to compare the total MARS Mean scores between Commercial and Non-Commercial apps (The data shown in Figure 10). The results showed a statistically significant difference between the two groups, with a t-statistic of 4.36 and a p-value of 0.012 (p < 0.05), indicating that commercial apps tended to perform significantly better than non-commercial apps. Commercial apps had a significantly higher mean MARS score (4.34) compared to Non-Commercial apps (3.34). The p-value (<0.05) suggested that this difference is statistically significant, meaning the likelihood that this result occurred by chance is low. These findings suggested that commercial apps might have better resources, design, and engagement strategies, leading to higher overall quality.

Figure 10
Box plot comparing Total MARS Mean Scores by affiliation. Non-commercial scores average around 3.25 to 3.5, while commercial scores range from 4.0 to 4.5, indicating higher ratings.

Figure 10. Comparison of MARS scores between commercial and non-commercial apps.

4 Discussion

The comprehensive analysis provided insights into their functionality, design quality, engagement features, behavioral outcomes, and credibility of information. The findings demonstrated a wide variation in app quality, with clear patterns emerging between commercial and non-commercial apps. The analysis revealed that functionality remains a strong aspect across most apps, with a high mean score (4.61) and low variability. This indicated developers, particularly in commercial sectors, prioritize user-friendly interfaces, efficient navigation, and minimal technical glitches. This also aligned with previous research, highlighting usability as a critical factor in apps adoption and sustained usage (2123).

However, usability alone does not equate to effectiveness in health communication or behavior change. One of the most striking findings was the high variability in engagement scores (mean = 3.80, SD = 1.17), which underscores disparities in how different apps maintain user interest. Apps such as Asianparent and Tentang Anak demonstrated strong engagement through features such as gamified content, push notifications, and community forums. This correlated with the findings by Stoyanov et al, which emphasized the significance of interactivity and tailored content in promoting continued use of health apps. On the opposite end, PSG Balita scored poorly, showing that technical performance without engaging elements was insufficient to keep users actively involved (10).

A closer look at information quality reveals another critical gap. Despite the high aesthetic and functional scores, several apps lacked well-sourced or expert-reviewed content. This confirms concerns raised by earlier studies that many commercially popular apps were not adhered to clinical guidelines or transparency regarding content sources (24, 25). The weak correlation (−0.33) between functionality and information quality in our data further underscored that a well-functioning app might still be inadequate in delivering trustworthy health information. This is concerning in the context of pediatric health, where misinformation can have serious consequences for child development and public health outcomes (26, 27).

Another important finding was the correlation between engagement and information quality (r = 0.89), suggesting the apps that were engaging also tended to present more credible information. This may be due to better resourcing or more holistic development strategies in commercially successful apps. It also supports theories from behavior change models, such as the COM-B framework (Capability, Opportunity, Motivation, and Behavior), which emphasized that both cognitive engagement and access to credible information are necessary to initiate and sustain behavior change (28, 29).

In terms of behavioral outcomes, apps performed moderately well in raising awareness and promoting help-seeking behavior but struggled with initiating long-term behavioral change. The average score for behavior change was the lowest among categories (mean = 2.89). This echoes similar findings by Zhao (30), who observed that many health apps fail to include evidence-based behavior change techniques (BCTs), such as goal setting, feedback loops, and rewards, that was critical for sustained impact. Given the burden of child malnutrition and stunting in Indonesia, the lack of BCTs significantly limited the apps' utility in public health interventions.

The comparative analysis between commercial and non-commercial apps revealed statistically significant differences in quality (p = 0.012), with commercial apps outperforming their counterparts in nearly all MARS categories. While this might be attributed to higher budgets, access to better design tools, and more aggressive user-testing, it raised ethical and equity concerns. Users from lower socioeconomic backgrounds might have limited access to premium features or might be exposed to advertisements and data privacy risks. This supported arguments that commercial models in child-focused mHealth might undermine public trust and dilute educational value (31, 32).

Our analysis indicated that applications with rich interactive features, including personalized dashboards, feedback systems, gamification, and real-time chat support, tend to receive higher user ratings and demonstrate stronger retention metrics. In contrast, applications lacking interactivity, such as those offering only static text or generalized health tips, are associated with lower engagement levels and higher dropout rates over time. This finding aligns with existing literature, which underscored the significance of user-centered design and engagement-driven features in sustaining digital health tool usage. Our benchmarking also showed that applications developed with active user input and designed around user-centered features are more likely to receive higher ratings and exhibit better retention metrics. This underscores the necessity for developers to prioritize not only the accuracy of health content but also the manner in which users interact with it. Incorporating engaging, responsive, and culturally appropriate features can improve the user experience and render digital tools more effective in supporting pediatric health outcomes. Conversely, applications that lack sufficient interactivity, such as those providing only static health information or basic notification systems, tend to experience lower user engagement. This is particularly evident in applications targeting parents in pediatric care, where usability and real-time feedback are essential. Low interactivity often results in reduced perceived value, limited user satisfaction, and ultimately, higher drop-off rates over time (33).

Existing studies revealed that, while mHealth applications share core features across different regions, their effectiveness is significantly influenced by local adaptation. Regional specificity in tool design, including the integration of local languages, is crucial, particularly in lower-middle-income countries, where numerous indigenous languages are spoken. For instance, a study conducted in Guangzhou, China revealed that 91.7% of the rural population was functionally illiterate and had not completed middle school, making it difficult for them to understand Mandarin or Cantonese (34).

This underscores the critical need for user-friendly interfaces and linguistic localization to enhance accessibility among the underserved populations. Moreover, in areas with limited Internet connectivity, such as parts of Sub-Saharan Africa, traditional communication methods, such as SMS and phone calls, are often preferred over internet-based applications. These platforms are more feasible for reaching users who lack smartphones or have stable Internet access. Therefore, to enhance the accessibility of mHealth in this group, two things need to be considered in designing and implementing the mHealth: the digital health literacy of the users and the supporting technological infrastructure (35).

Stronger collaboration between developers and policymakers is essential to address these challenges. Improving digital health literacy among decision makers can foster more effective engagement with technology developers and lead to more inclusive and sustainable solutions. Even when government staff possess technical expertise, low digital health literacy may hinder their ability to effectively evaluate and integrate new tools into health systems. Moreover, digital literacy skills in conjunction with the available supporting technological infrastructure can help in deciding which technology is appropriate for the users, such as SMS, phone calls, or mobile applications. Therefore, building capacity among public officials and ensuring their involvement in the design and scaling of digital health tools are critical steps toward improving accessibility for low-income and marginalized communities (3436).

Despite providing valuable insights, this study had some limitations. First, the MARS evaluation, while comprehensive, remained a subjective assessment that might not fully capture long-term behavioral outcomes To gain a comprehensive understanding of the long-term impact of mobile health (mHealth) applications on pediatric health behaviors, future research should employ longitudinal study designs. Such methodologies facilitate an in-depth examination of behavioral changes over time, particularly concerning the persistence of user engagement and adherence to health recommendations. Long-term studies are especially critical for evaluating outcomes such as vaccination completion, breastfeeding duration, and compliance with growth monitoring schedules. Furthermore, it is imperative to investigate how user demographics, including age, socioeconomic status, education level, and digital health literacy, influence the engagement and overall effectiveness of mHealth tools. Previous research has indicated that disparities in digital health literacy, particularly among healthcare professionals, policymakers, and government personnel, can impede the optimal adoption and integration of these tools within national health systems (37). Furthermore, future research should focus more closely on technical design factors, such as the user interface (UI) and user experience (UX). Examining how diverse user groups interact with different app features, especially within large-scale public health programs, can provide valuable insights into developing more inclusive, accessible, and culturally sensitive digital platforms. Evaluating usability across various social, geographic, and infrastructural contexts will enhance the likelihood of adoption and improve long-term retention, which are essential for sustained health improvements. It is also important for future research to assess a larger and more diverse sample of applications, including mHealth tools from platforms beyond Bahasa Indonesia app stores, and monitor real-world usage over extended periods. These strategies will contribute to a more robust understanding of how mHealth applications perform across different populations and contexts, ultimately supporting the development of more effective, equitable, and user-centered digital health solutions for pediatric care in low- and middle-income countries.

5 Conclusions

This study applied the MARS framework to assess the quality and behavioral potential of nine pediatric mHealth apps in Indonesia. While many apps demonstrated strengths in functionality and aesthetics, significant gaps remained in engagement, information credibility, and behavioral impact. Commercial apps outperformed non-commercial ones, yet their dominance also raised concerns about equity and public interest.

To optimize the role of mHealth in reducing child stunting and improving pediatric care in Indonesia, future interventions must strike a balance between user-centered design and evidence-based health content. Multi-sectoral collaboration between developers, public health authorities, and academic institutions will be essential to develop inclusive, high-impact apps that not only inform but transform caregiver behavior and health outcomes.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

AI: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. AA: Data curation, Formal analysis, Investigation, Methodology, Validation, Writing – original draft, Writing – review & editing. RL: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. NA: Data curation, Formal analysis, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. APM: Data curation, Formal analysis, Investigation, Writing – original draft, Writing – review & editing. AVM: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing. BD: Supervision, Validation, Writing – original draft, Writing – review & editing. EG: Funding acquisition, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. Stipendium Hungaricum Scholarship, Tempus Public Foundation, registration number SHE-44703-004/2021.

Acknowledgments

We would like to express our heartfelt gratitude to the Institute of Behavioral Sciences at Semmelweis University for their invaluable academic support, insightful discussions, and continuous encouragement throughout the course of this study. We are also grateful to the Tempus Foundation for providing the main funding for this research. Their contributions have significantly enriched the depth and quality of our work. Finally, we extend our sincere thanks to Sarah Eringaard for her careful proofreading and English language editing of this manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdgth.2025.1621293/full#supplementary-material

Abbreviations

mHealth, Mobile Health; MARS, Mobile Application Rating Scale; ICC, Intraclass Correlation Coefficient; STROBE, The Strengthening the Reporting of Observational Studies in Epidemiology Guidelines; COM-B, Capability, Opportunity, Motivation, and Behavior; BCTs, Based Behavior Change Techniques.

References

1. Wang J-W, Zhu Z, Shuling Z, Fan J, Jin Y, Gao Z-L, et al. Effectiveness of mHealth app–based interventions for increasing physical activity and improving physical fitness in children and adolescents. Systematic review and meta-analysis. JMIR Mhealth Uhealth. (2024) 12:e51478. doi: 10.2196/51478

PubMed Abstract | Crossref Full Text | Google Scholar

2. Lee J, Su Z, Chen Y. Mobile Apps for Children’s Health and Wellbeing: Design Features and Future Opportunities. (1942-597X (Electronic)).

Google Scholar

3. Caroline B, Sandi C, Shazima T, Viveca L. Parents’ perceptions about future digital parental support-A phenomenographic interview study. Front Digit Health. (2021) 3:729697. doi: 10.3389/fdgth.2021.729697

PubMed Abstract | Crossref Full Text | Google Scholar

4. Jaggi L, Aguilar L, Alvarado Llatance M, Castellanos A, Fink G, Hinckley K, et al. Digital tools to improve parenting behaviour in low-income settings: a mixed-methods feasibility study. Arch Dis Child. (2023) 108(6):433–9. doi: 10.1136/archdischild-2022-324964

PubMed Abstract | Crossref Full Text | Google Scholar

5. UNESCO. Global Education Monitoring Report 2023: Technology in Education – A Tool on Whose Terms? PARIS: Unesco (2023).

Google Scholar

6. Hicks JL, Boswell MA, Althoff T, Crum AJ, Ku JP, Landay JA, et al. Leveraging Mobile Technology for Public Health Promotion: A Multidisciplinary Perspective. (1545-2093 (Electronic)).

Google Scholar

7. Rinawan FR, Susanti AI, Amelia I, Ardisasmita MN, Widarti , Dewi RK, et al. Understanding mobile application development and implementation for monitoring posyandu data in Indonesia: a 3-year hybrid action study to build “a bridge” from the community to the national scale. BMC Public Health. (2021) 21(1):1024. doi: 10.1186/s12889-021-11035-w

PubMed Abstract | Crossref Full Text | Google Scholar

8. Kabongo EM, Mukumbang FC, Delobelle P, Nicol E. Explaining the impact of mHealth on maternal and child health care in low- and middle-income countries: a realist synthesis. BMC Pregnancy Childbirth. (2021) 21(1):196. doi: 10.1186/s12884-021-03684-x

PubMed Abstract | Crossref Full Text | Google Scholar

9. Nurhaeni N, Chodidjah S, Adawiyah R, Astuti . Using a mobile application (“PrimaKu”) to promote childhood immunization in Indonesia: a cross-sectional study. Belitung Nurs J. (2021) 7(4):329–35. doi: 10.33546/bnj.1524

PubMed Abstract | Crossref Full Text | Google Scholar

10. Stoyanov SA-O, Hides L, Kavanagh DJ, Zelenko O, Tjondronegoro D, Mani M. Mobile app rating scale: a new tool for assessing the quality of health mobile apps. (2291-5222 (Print)).

Google Scholar

11. Deniz-Garcia A, Fabelo H, Rodriguez-Almeida AJ, Zamora-Zamorano G, Castro-Fernandez M, Alberiche Ruano MDP, et al. Quality, usability, and effectiveness of mHealth apps and the role of artificial intelligence: current scenario and challenges. J Med Internet Res. (2023) 25:e44030. doi: 10.2196/44030

PubMed Abstract | Crossref Full Text | Google Scholar

12. Moungui HC, Nana-Djeunga HC, Anyiang CF, Cano M, Ruiz Postigo JA, Carrion C. Dissemination strategies for mHealth apps: systematic review. JMIR Mhealth Uhealth. (2024) 12:e50293. doi: 10.2196/50293

PubMed Abstract | Crossref Full Text | Google Scholar

13. Gonzales A, Custodio R, Lapitan MC, Ladia MA. End users’ perspectives on the quality and design of mHealth technologies during the COVID-19 pandemic in the Philippines: qualitative study. JMIR Form Res. (2023) 7:e41838. doi: 10.2196/41838

PubMed Abstract | Crossref Full Text | Google Scholar

14. Aungst T, Seed S, Gobin N, Jung R. The good, the bad, and the poorly designed: the mobile app stores are not a user-friendly experience for health and medical purposes. Digit Health. (2022) 8:20552076221090038. doi: 10.1177/20552076221090038

PubMed Abstract | Crossref Full Text | Google Scholar

15. Chan AA-O, Horne RA-O, Hankins M, Chisari C. The Medication Adherence Report Scale: A measurement tool for eliciting patients’ reports of nonadherence. (1365-2125 (Electronic)).

Google Scholar

16. Bartoš F, Martinková P, Brabec M. Testing Heterogeneity in Inter-Rater Reliability. Quantitative Psychology. Cham: Springer International Publishing (2020).

Google Scholar

17. Kimel M, Revicki D. Inter-rater reliability. In: Maggino F, editor. Encyclopedia of Quality of Life and Well-Being Research. Cham: Springer International Publishing (2023). p. 3626–8.

Google Scholar

18. Wilks CR, Chu C, Sim D, Lovell J, Gutierrez P, Joiner T, et al. User engagement and usability of suicide prevention apps: systematic search in app stores and content analysis. JMIR Form Res. (2021) 5(7):e27018. doi: 10.2196/27018

PubMed Abstract | Crossref Full Text | Google Scholar

19. Gmel AI, Gmel G, von Niederhäusern R, Weishaupt MA, Neuditschko M. Should we agree to disagree? An evaluation of the inter-rater reliability of gait quality traits in Franches-Montagnes stallions. J Equine Vet Sci. (2020) 88:102932. doi: 10.1016/j.jevs.2020.102932

PubMed Abstract | Crossref Full Text | Google Scholar

20. Terhorst Y, Philippi P, Sander LB, Schultchen D, Paganini S, Bardus M, et al. Validation of the mobile application rating scale (MARS). PLoS One. (2020) 15(11):e0241480. doi: 10.1371/journal.pone.0241480

PubMed Abstract | Crossref Full Text | Google Scholar

21. Palos-Sanchez P, Saura J, Rios Martin MÁ, Aguayo-Camacho M. Toward a Better Understanding of the Intention to Use mHealth Apps: Exploratory Study. (2291-5222 (Electronic)).

Google Scholar

22. Weichbroth P. Usability testing of mobile applications: a methodological framework. Appl Sci. (2024) 14(5):1792. doi: 10.3390/app14051792

Crossref Full Text | Google Scholar

23. Putra POH, Dewi R, Budi I. Usability factors that drive continued intention to use and loyalty of mobile travel application. (2405-8440 (Print)).

Google Scholar

24. Jake-Schoffman DE, Silfee VJ, Waring ME, Boudreaux ED, Sadasivam RS, Mullen SP, et al. Methods for evaluating the content, usability, and efficacy of commercial mobile health apps. JMIR Mhealth Uhealth. (2017) 5(12):e190. doi: 10.2196/mhealth.8758

PubMed Abstract | Crossref Full Text | Google Scholar

25. Rojas Mezarina L, Silva-Valencia J, Escobar-Agreda S, Espinoza Herrera DH, Egoavil MS, Maceda Kuljich M, et al. Need for the development of a specific regulatory framework for evaluation of mobile health apps in Peru: systematic search on app stores and content analysis. JMIR Mhealth Uhealth. (2020) 8(7):e16753. doi: 10.2196/16753

PubMed Abstract | Crossref Full Text | Google Scholar

26. Brammall BR, Hayman MJ, Harrison CL. Pregnancy mobile app use: a survey of health information practices and quality awareness among pregnant women in Australia. Womens Health. (2024) 20:17455057241281236. doi: 10.1177/17455057241281236

PubMed Abstract | Crossref Full Text | Google Scholar

27. Osude N, O'Brien E, Bosworth HB. The search for the missing link between health misinformation & health disparities. Patient Educ Couns. (2024) 129:108386. doi: 10.1016/j.pec.2024.108386

PubMed Abstract | Crossref Full Text | Google Scholar

28. Wu M, Wang W, He H, Bao L, Lv P. Mediating Effects of Health Literacy, Self-Efficacy, and Social Support on the Relationship Between Disease Knowledge and Patient Participation Behavior Among Chronic Ill Patients: A Cross-Sectional Study Based on the Capability-Opportunity-Motivation and Behavior (COM-B) Model. (1177-889X (Print)).

Google Scholar

29. Huang Z, Lum E, Car J. Medication Management Apps for Diabetes: Systematic Assessment of the Transparency and Reliability of Health Information Dissemination. (2291-5222 (Electronic)).

Google Scholar

30. Zhao J, Freeman B, Li M. Can Mobile Phone Apps Influence People’s Health Behavior Change? An Evidence Review. (1438-8871 (Electronic)).

Google Scholar

31. Meyer CL, Surmeli A, Hoeflin Hana C, Narla NP. Perceptions on a mobile health intervention to improve maternal child health for Syrian refugees in Turkey: opportunities and challenges for end-user acceptability. Front Public Health. (2022) 10:1025675. doi: 10.3389/fpubh.2022.1025675

PubMed Abstract | Crossref Full Text | Google Scholar

32. Chang BL, Bakken S, Brown SS, Houston TK, Kreps GL, Kukafka R, et al. Bridging the digital divide: reaching vulnerable populations. (1067-5027 (Print)).

Google Scholar

33. Li K, Magnuson KI, Beuley G, Davis L, Ryan-Pettes SR. Features, design, and adherence to evidence-based behavioral parenting principles in commercial mHealth parenting apps: systematic review. JMIR Pediatr Parent. (2023) 6:e43626. doi: 10.2196/43626

PubMed Abstract | Crossref Full Text | Google Scholar

34. Lestari HM, Miranda AV, Fuady A. Barriers to telemedicine adoption among rural communities in developing countries: a systematic review and proposed framework. Clin Epidemiol Glob Health. (2024) 28:101684. doi: 10.1016/j.cegh.2024.101684

Crossref Full Text | Google Scholar

35. Irawan AS, Döbrössy BM, Biresaw MS, Muharram AP, Kovács SD, Girasek E. Exploring characteristics and common features of digital health in pediatric care in developing countries: a systematic review. Front Digit Health. (2025) 7:1533788. doi: 10.3389/fdgth.2025.1533788

PubMed Abstract | Crossref Full Text | Google Scholar

36. Till S, Mkhize M, Farao J, Shandu L, Muthelo L, Coleman T, et al. Digital Health Technologies for Maternal and Child Health in Africa and Other Low- and Middle-Income Countries: Cross-disciplinary Scoping Review With Stakeholder Consultation. (1438-8871 (Electronic)).

Google Scholar

37. Cheng C, Gearon E, Hawkins M, McPhee C, Hanna L, Batterham R, et al. Digital health literacy as a predictor of awareness, engagement, and use of a national web-based personal health record: population-based survey study. J Med Internet Res. (2022) 24(9):e35772. doi: 10.2196/35772

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: pediatric care, stunting prevention, assessment, e-health, digital health, user experience (UX), user interface (UI)

Citation: Irawan AS, Alristina AD, Laili RD, Amalia N, Muharram AP, Miranda AV, Döbrössy B and Girasek E (2025) Beyond the interface: benchmarking pediatric mobile health applications for monitoring child growth using the Mobile App Rating Scale. Front. Digit. Health 7:1621293. doi: 10.3389/fdgth.2025.1621293

Received: 7 May 2025; Accepted: 3 June 2025;
Published: 18 June 2025.

Edited by:

Kirti Sundar Sahu, Canadian Red Cross, Canada

Reviewed by:

Asriadi Asriadi, University of Mega Buana Palopo, Indonesia
Henrique Yoshikazu Shishido, Federal University of Paraná, Brazil

Copyright: © 2025 Irawan, Alristina, Laili, Amalia, Muharram, Miranda, Döbrössy and Girasek. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Anggi Septia Irawan, aXJhd2FuLmFuZ2dpQHBoZC5zZW1tZWx3ZWlzLmh1

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.