- 1Graduate School, Beijing University of Chinese Medicine, Beijing, China
- 2Department of Pain Medicine, Beijing Hospital of Traditional Chinese Medicine, Capital Medical University, Beijing, China
Background: The proliferation of short video platforms has transformed public health communication, yet the quality of medical information shared on these platforms remains inconsistent. Osteoarthritis (OA), a prevalent and burdensome chronic condition, is frequently featured in online health content. However, the reliability of such information has not been systematically evaluated across major Chinese short video platforms. To assess and compare the quality and reliability of OA-related health information on TikTok and Bilibili, and to examine the influence of uploader type and user engagement metrics on content quality.
Methods: In this cross-sectional study, a total of 189 OA-related videos were collected from TikTok (n = 96) and Bilibili (n = 93) using a standardized search strategy. Four validated instruments—the Journal of the American Medical Association (JAMA) benchmarks, modified DISCERN (mDISCERN), Global Quality Score (GQS), and Health on the Net Code (HONcode)—were used for video assessment. Each video was independently rated by two trained reviewers. Differences in quality scores were compared across platforms and uploader types (health professionals vs. non-professionals). Spearman correlation analysis was conducted to explore associations between video quality and engagement metrics (likes, comments, shares, favorites).
Results: TikTok videos exhibited significantly higher median scores on JAMA (2.4 vs. 2.1, P = 0.001), GQS (3.0 vs. 3.0, P = 0.006), and HONcode (11.0 vs. 9.3, P = 0.005) compared to Bilibili. No significant difference was observed for mDISCERN scores. Videos uploaded by healthcare professionals had significantly higher GQS (P = 0.004) and HONcode scores (P = 0.010) than those from non-professionals. User engagement metrics were positively correlated with content quality, particularly on TikTok (e.g., likes vs. JAMA, r = 0.732, P < 0.001).
Conclusions: OA-related videos on TikTok demonstrate higher overall quality and reliability compared to Bilibili, especially when created by healthcare professionals. User engagement metrics are positively associated with information quality, underscoring the importance of expert-led digital health communication. These findings highlight the need for platform-level interventions to promote trustworthy content and improve the digital health information ecosystem.
1 Introduction
Osteoarthritis (OA) is among the most prevalent chronic joint diseases worldwide, with its incidence steadily increasing due to global population aging (1). It primarily affects the knee, hip, and shoulder joints, significantly impairing mobility and reducing quality of life in older adults (2). The prolonged disease course and recurrent symptoms contribute to a substantial burden on both individual health and public healthcare systems (3, 4). Epidemiological studies demonstrate considerable regional variation in OA prevalence, with higher rates observed among older populations in developed countries (5). Beyond its direct physical impact, OA is associated with increased healthcare utilization and psychological distress (6, 7). Although available treatments—including pharmacological therapy, physical rehabilitation, and surgical interventions—offer symptom relief, their effectiveness varies, and some carry risks of adverse effects (8). Therefore, early screening, lifestyle modification, and enhanced patient self-management are critical components in OA care (9, 10).Research indicates that digital education has already played a role in osteoarthritis treatment, and some online education even outperforms patient self-management (11, 12).As OA requires long-term self-management, reliable online health information is particularly important for patient education.
The proliferation of internet access has transformed health information-seeking behaviors, with online platforms increasingly serving as primary sources of medical knowledge (13–15). Short video platforms such as TikTok and Bilibili have rapidly emerged as popular mediums for disseminating health information, owing to their intuitive, engaging, and easily shareable formats. Content on these platforms encompasses disease education, treatment options, patient experiences, expert commentary, and lifestyle guidance, offering opportunities for public health promotion (16).
Nevertheless, the quality and reliability of health information available on social media platforms vary widely (17–19). Algorithm-driven recommendation systems, favoring content with high engagement metrics, may inadvertently amplify low-quality or misleading medical information (20–22).Given the potential influence of such content on patient decision-making and health behaviors (23), evaluating the accuracy, credibility, and educational value of OA-related videos is imperative. Prior studies have assessed similar issues for other diseases (20, 24), but OA-related content on Chinese short video platforms remains underexplored.
Accordingly, this study aimed to systematically assess the quality and reliability of OA-related health information disseminated via TikTok and Bilibili. Specifically, we sought to (1) compare content quality between platforms, (2) examine the impact of uploader background (health professionals vs. laypersons) on video quality, and (3) investigate the relationship between content quality and user engagement metrics. To achieve this, we employed multiple validated evaluation instruments, including the Journal of the American Medical Association (JAMA) benchmarks, the modified DISCERN (mDISCERN) scale, the Global Quality Score (GQS) (25–27), and the Health On the Net Foundation's Code of Conduct (HONcode) (28).
2 Methods
This cross-sectional study aimed to assess the quality and reliability of OA-related health information disseminated via two major short video platforms, TikTok and Bilibili. A systematic search strategy, standardized video characterization, validated quality assessment instruments, and rigorous statistical analyses were employed to ensure comprehensive and objective evaluation.
2.1 Ethical considerations
This study analyzed publicly available video content from TikTok and Bilibili without collecting personal data or interacting with users. According to institutional and international guidelines, the study did not constitute human subjects research. Nonetheless, the protocol was reviewed and approved by the Institutional Review Board of Beijing Hospital of Traditional Chinese Medicine. The authors confirm compliance with TikTok's Terms of Use and Bilibili's platform policies.
2.2 Search strategy
To ensure the breadth and representativeness of the video sample, TikTok and Bilibili were selected as the primary data sources based on their large user bases and popularity in disseminating health-related content in China.
A new user account was registered on each platform to minimize the influence of personalized recommendation algorithms. The keyword “Osteoarthritis” (in Chinese: “骨关节炎”) was used to search for relevant videos. Searches were conducted in February 2025, without geographic restrictions, and included publicly available videos only. No filters were applied to sort the search results, thereby preserving the platforms' default ranking algorithms.
Although the study focused on the top-ranked search results to reflect real user exposure, the potential effect of algorithmic recommendation bias is acknowledged as a limitation.
2.3 Screening criteria
Videos were eligible if they met the following criteria: (1) explicitly related to osteoarthritis health education; (2) publicly accessible without privacy restrictions; and (3) duration between 0 and 60 min. Exclusion criteria included: (1) videos primarily addressing other diseases or pediatric populations; (2) advertisements unrelated to OA; and (3) duplicate videos or content lacking substantive OA relevance. An initial pool of 200 videos (100 from TikTok and 100 from Bilibili) was retrieved. After screening based on the predefined criteria, 189 videos were included in the final analysis, comprising 96 videos from TikTok and 93 videos from Bilibili. All inclusion and exclusion decisions were independently cross-checked by two reviewers to ensure consistency and transparency.
2.4 Video characterization
For each included video, a standardized set of characteristics was recorded, including basic information (video duration, main topic, and uploader identity), uploader type (categorized as healthcare professionals, such as osteopaths and rehabilitation therapists, or non-professionals, such as patients and health bloggers), content type (classified into disease knowledge, treatment methods, or lifestyle interventions, allowing for multiple categorizations per video), and user engagement metrics (number of likes, comments, favorites, and shares).
2.5 Video quality and reliability assessments
To systematically evaluate video quality, four validated instruments were employed: the Journal of the American Medical Association (JAMA) benchmarks, the modified DISCERN (mDISCERN) scale, the Global Quality Score (GQS), and the Health On the Net Foundation's Code of Conduct (HONcode).
The Journal of the American Medical Association (JAMA) benchmarks assessed the reliability of information sources and disclosure of conflicts of interest, with scores ranging from 0 to 4, where higher scores indicated greater credibility (Table 1) (20). The modified DISCERN (mDISCERN) scale evaluated clarity, reliability, balance, citation of external sources, and acknowledgment of uncertainties, using a binary scoring system (0 or 1 per item, total score 0–5) (Table 2) (29). The Global Quality Score (GQS) provided an overall assessment of educational value, rated from 1 (poor) to 5 (excellent) (Table 3) (25). The Health On the Net Foundation Code of Conduct (HONcode) assessed adherence to eight ethical principles for online health information, with a cumulative score ranging from 0 to 16 (Table 4) (30).
Table 1. Description of the JAMA benchmark criteria (4-point scale). Each criterion is scored as 1 if met, 0 if not. Total scores range from 0 to 4.
Table 2. Modified DISCERN (mDISCERN) scale criteria (5-point scale) used to assess the quality of OA-related video content.
Table 3. Description of the Global Quality Score (GQS) used to evaluate the overall quality and patient usefulness of OA-related video content.
Table 4. Description of the health on the Net (HONcode) criteria used to evaluate the ethical quality and transparency of OA-related video content.
Video evaluations were performed independently by four trained experts following a randomized allocation of videos. A computer-generated randomization sequence assigned 189 videos into two groups of approximately equal size. Each group was evaluated by two independent reviewers (Group I: Shuming Li and Ren Guo; Group II: Ping Chen and Chenyu Zhang), who were blinded to the video source and uploader identity. In cases of substantial discrepancy between raters, a third independent clinical expert adjudicated the scores, with the final rating determined by the median value of the three evaluations. Prior to assessment, all reviewers underwent calibration training based on standardized scoring guidelines to ensure consistency and minimize subjective bias.
2.6 Statistical analysis
Continuous variables were presented as mean ± standard deviation (SD) or median and interquartile range (IQR), depending on the distribution. Categorical variables were expressed as frequencies and percentages. The Shapiro–Wilk test was used to assess the normality of continuous variables. Independent sample t-tests or Mann–Whitney U tests were applied to compare video quality scores (JAMA, mDISCERN, GQS, and HONcode) between platforms and uploader types. Spearman's rank correlation analysis was conducted to evaluate the associations between video quality scores and user engagement metrics. A two-sided P-value < 0.05 was considered statistically significant. All analyses were performed using SPSS version 27.0 (IBM Corp, Armonk, NY, USA).All analyses were independently verified by two statisticians to ensure accuracy and reproducibility.
3 Results
3.1 Video characterization
A total of 200 osteoarthritis-related videos were initially retrieved, with 100 from TikTok and 100 from Bilibili. After applying predefined inclusion and exclusion criteria, 189 videos were included in the final analysis, comprising 96 TikTok and 93 Bilibili videos (Figure 1). Among these, 6 videos (6.3%) from TikTok and 28 videos (30.1%) from Bilibili were uploaded by non-professionals.
Figure 2 summarizes the basic characteristics of the included videos. TikTok videos had a median duration of 75.5 s (IQR 49–108), compared to 152 s (IQR 72.5–332.5) on Bilibili (P < .001). The number of comments was significantly higher on Bilibili (median 128; IQR 15.5–424) than on TikTok (median 27; IQR 4–108.5; P = .002), as were favorites (median 697 vs. 179.5; P = .001). No statistically significant differences were observed in the number of likes (median 898 vs. 662.5; P = .351) or shares (median 297 vs. 148; P = .085) between the two platforms. The number of followers of each uploader was also recorded to provide additional context for engagement differences, although this variable was not included in inferential analyses.
Figure 2. Distribution of video characteristics on TikTok and Bilibili. Violin plots show engagement metrics (likes, comments, favorites, shares; measured in counts) and video duration (measured in seconds).
3.2 Video quality scoring
Video quality was assessed using four established scoring tools (Figure 3). TikTok videos had significantly higher JAMA scores (median 2.4; IQR 2.2–2.6) than Bilibili videos (median 2.1; IQR 2.0–2.6; P = .001). GQS scores were also higher on TikTok (median 3; IQR 2–3) compared to Bilibili (median 3; IQR 2–3; P = .006). HONcode scores were higher for TikTok (median 11; IQR 9–11.75) than for Bilibili (mean 9.33 ± 2.59; P = .005). No statistically significant difference was found in mDISCERN scores (median 3; IQR 2–3 for both platforms; P = .196).
Figure 3. Distribution of video quality scores on tikTok and bilibili. Violin plots show results from four evaluation tools: JAMA (0–4), mDISCERN (0–5), Global Quality Score (1–5), and HONcode (0–16).
3.3 Correlation analysis
Spearman correlation analysis was conducted to examine the associations between user engagement metrics and video quality scores on TikTok and Bilibili. Results are presented in Figure 4.
Figure 4. Correlation between user engagement metrics and video quality scores on tikTok and bilibili. Presented as a ring plot, outer segments represent engagement metrics, and inner rings display Spearman correlation coefficients (r) and P values for each quality score (JAMA, mDISCERN, GQS, HONcode).
On TikTok, all engagement metrics were significantly positively correlated with JAMA scores, with the highest correlation observed for likes (r = 0.732, P < .001), followed by comments (r = 0.704), favorites (r = 0.684), and shares (r = 0.617). mDISCERN scores were also positively correlated with likes (r = 0.611), comments (r = 0.592), favorites (r = 0.561), and shares (r = 0.590; all P < .001). For GQS, correlation coefficients ranged from r = 0.467 (shares) to r = 0.578 (likes), and for HONcode from r = 0.483 (shares) to r = 0.618 (likes). Video duration was not significantly correlated with any score (P > .05).
On Bilibili, likes were moderately correlated with JAMA (r = 0.470, P < .001), mDISCERN (r = 0.399), GQS (r = 0.390), and HONcode (r = 0.449). Similar correlations were found for favorites (e.g., r = 0.480 for GQS; r = 0.495 for HONcode) and shares (r = 0.414 for HONcode). Comments also showed positive but weaker correlations with most scores (r = 0.309–0.430). Unlike TikTok, video duration was significantly correlated with JAMA (r = 0.344, P = .001), mDISCERN (r = 0.382, P < .001), and HONcode (r = 0.266, P = .010), but not with GQS (P = .071).
3.4 Comparison between groups
Comparison by uploader type revealed that videos from healthcare professionals had significantly higher GQS scores (median 3; IQR 2–3) compared to those from non-professionals (median 2; IQR 2–3; P = .004; Figure 5). HONcode scores were also significantly higher for professional uploaders (median 10; IQR 8–11) than for non-professionals (mean 8.76 ± 2.82; P = .010). No significant differences were found in JAMA scores (median 2.3; IQR 2.0–2.6 vs. mean 2.16 ± 0.76; P = .067) or mDISCERN scores (median 3; IQR 2–3 for both groups; P = .473).Subgroup analysis by content theme (disease knowledge, treatment, lifestyle) suggested a similar trend, though differences were not statistically significant.
Figure 5. Distribution of video quality scores by uploader type. Violin plots show scores from four evaluation tools—JAMA (0–4), mDISCERN (0–5), Global Quality Score (1–5), and HONcode (0–16)—for videos uploaded by medical and non-medical professionals.
4 Discussion
4.1 Platform differences and potential mechanisms
This study systematically evaluated the quality and reliability of OA-related health information on TikTok and Bilibili, revealing significant differences between the platforms. TikTok videos demonstrated higher quality across JAMA, GQS, and HONcode assessments, while mDISCERN scores showed no statistically significant difference. These findings suggest that, overall, TikTok may provide more credible and educational OA-related content than Bilibili. These results indicate an association rather than a causal relationship, given the cross-sectional naturse of the data.
Several factors may explain this disparity. First, platform algorithms likely influence the visibility of high-quality content (31). TikTok's algorithm is hypothesized to favor professionally produced, high-engagement videos, aligning with prior evidence showing that platforms such as YouTube prioritize trustworthy health information (25). In contrast, Bilibili's more decentralized and community-driven recommendation system may offer less curation, allowing content of variable quality to proliferate.
Second, differences in user demographics and video formats may shape content quality. TikTok's short-form, attention-driven design incentivizes concise, information-dense videos, while Bilibili's long-form culture may foster verbose or diluted messaging. Additionally, TikTok's recent moderation efforts targeting health misinformation may have elevated the baseline quality of content on the platform, while Bilibili continues to host a larger proportion of videos created by non-professionals, as reflected in our sample.
Importantly, while TikTok outperformed Bilibili in overall quality metrics, it remains susceptible to health misinformation (32). Prior research has demonstrated the presence of inaccurate or misleading videos even on well-regulated platforms. For example, in a study of sinusitis-related TikTok content, the majority of non-medical influencer videos contained inaccurate claims (33).This underscores the persistent need for quality assurance, regardless of platform. This highlights that algorithmic advantages do not necessarily eliminate misinformation risks, underscoring the need for ongoing monitoring and quality regulation.
4.2 Comparison with existing studies
Our results are consistent with previous studies demonstrating that healthcare professionals produce higher-quality content than non-professionals (34). Liang et al. similarly found that TikTok outperformed Bilibili in content reliability and educational value for videos about gastroesophageal reflux disease (20). This cross-condition consistency suggests structural platform advantages. However, such consistency should be interpreted as an observed pattern rather than a causal conclusion about platform mechanisms.
The observed superiority of professional content also echoes findings from YouTube-based studies, where professional authors achieved significantly higher GQS and mDISCERN scores than lay creators (25, 35).These results reinforce the importance of content source credibility in health communication.
Nonetheless, platform performance can vary by disease topic and cultural context. One study found that Chinese-language videos on Bilibili and TikTok outperformed English-language content on YouTube for gastric cancer, suggesting that public familiarity and topic sensitivity may influence content quality (24).
Regarding the relationship between content quality and user interaction, prior research has been mixed. Some studies have reported no significant correlation between engagement metrics and quality, while our findings demonstrate a strong positive association, particularly on TikTok. This may reflect algorithmic designs that amplify credible content, or user behavior that favors well-produced and informative videos (36, 37). Nevertheless, these correlations cannot confirm directional causality, as engagement may also drive algorithmic exposure or creator adaptation over time. In addition, We did not perform comment coding in this study, Further work may classify interaction types (e.g., medically relevant vs. casual comments) to assess “effective engagement” and its link to video quality.
4.3 Clinical and public health significance
This study underscores the critical role that short video platforms can play in patient education and chronic disease management. For individuals with osteoarthritis, access to accurate, high-quality online content can support informed decision-making, improve adherence to self-care practices, and potentially enhance long-term outcomes. The higher prevalence of reliable, professionally produced videos on TikTok suggests that certain platforms may offer more trustworthy health information environments than others (38). Nevertheless, these findings reflect associations observed within the current dataset and should not be generalized as an inherent superiority of one platform over another, since audience demographics and algorithmic mechanisms differ.
Given the chronic and self-managed nature of osteoarthritis, continuous access to concise and evidence-based content—particularly outside clinical settings—is essential. Videos produced by healthcare professionals may be especially effective due to their clarity, scientific accuracy, and ability to foster trust. Such content can serve as an extension of clinical care, reinforcing key messages and promoting sustained patient engagement (39, 40).
From a public health perspective, these findings suggest that social media platforms should be integrated into broader health communication strategies (41). To maximize public health benefits, platform operators should consider prioritizing evidence-based content in recommendation systems, introducing visible quality indicators (e.g., verification badges), and implementing mechanisms to flag or demote misleading material. Enhancing the discoverability of credible content—particularly from verified medical professionals—can improve the informational environment and reduce exposure to misinformation (42).
Furthermore, the observed association between video quality and user engagement supports the potential for high-quality content to achieve broader reach and impact. Public health agencies and healthcare institutions may benefit from partnering with digital creators or investing in content development to amplify accurate messaging at scale.
Finally, this study highlights the dual responsibility in digital health education: clinicians and public health professionals must actively engage in content dissemination while also equipping patients with critical appraisal skills to navigate an increasingly complex information landscape. Sustained improvement in online health information quality will thus require coordinated efforts among platforms, professionals, and users to balance accessibility with accuracy.
5 Limitations
This study has several limitations. First, the sample size of 189 OA-related videos may not fully capture the breadth and variability of content on TikTok and Bilibili. Future studies should consider larger and more diverse samples to improve generalizability. Second, the focus on a single disease—osteoarthritis—limits the applicability of findings to other medical domains. Expanding the scope to include additional health topics could yield a more comprehensive understanding of platform dynamics.
Third, although we employed four validated assessment tools (JAMA, mDISCERN, GQS, HONcode), the evaluation of video quality involves inherent subjectivity and does not directly assess clinical accuracy. Fourth, data access limitations precluded analysis of certain engagement metrics, such as view duration or algorithmic amplification, which may influence content visibility and user behavior. These constraints mean that platform-level mechanisms, such as algorithmic ranking and exposure bias, could not be fully captured, which may partly account for observed variations in engagement patterns. Finally, the cross-sectional design reflects a single time point, limiting the ability to assess temporal changes or causality. Therefore, the observed associations should be interpreted as descriptive rather than explanatory of underlying mechanisms.
Future research should incorporate longitudinal analyses, multi-language and multi-platform comparisons (43), and explore the effectiveness of content moderation strategies such as health labeling, verification badges, and user education initiatives to strengthen digital health literacy and content quality.Collaborative efforts among researchers, health authorities, and platform developers could further enhance the governance and transparency of online health information ecosystems.
6 Conclusion
This cross-sectional analysis revealed substantial variability in the quality and reliability of osteoarthritis-related health information across two major Chinese short video platforms. Overall, content on TikTok demonstrated higher credibility and educational value than that on Bilibili, particularly when produced by healthcare professionals. These observations reflect associations within the studied sample and should not be interpreted as evidence of inherent platform superiority. Importantly, video quality was positively associated with user engagement metrics, suggesting that well-crafted, accurate health content may also achieve greater visibility and influence. This correlation highlights the opportunity—but not the certainty—for credible content to reach wider audiences under current algorithmic dynamics. These findings underscore the critical role of platform governance and professional participation in shaping the digital health information environment. Future efforts should prioritize the integration of quality assurance mechanisms and the amplification of expert-led content to support informed health decision-making in the era of social media. By providing comparative data within the Chinese digital health context, this study contributes a localized perspective that can inform future international research on short-video health communication. Future research may also consider incorporating review mechanisms and algorithmic logic as mediating factors to better elucidate the pathways through which platform governance influences content quality and user engagement.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
This study analyzed publicly available, anonymized video content from TikTok and Bilibili, without collecting personal data or engaging with platform users. In accordance with institutional and international ethical standards, this research did not constitute human subjects research. The study protocol was reviewed and approved by the Institutional Review Board of Beijing Hospital of Traditional Chinese Medicine, Capital Medical University. The authors confirm that all analyses complied with TikTok’s Terms of Service and Bilibili’s platform policies, particularly regarding the use of publicly accessible data for academic research. The ethics committee determined that this study did not constitute human subjects research and that written informed consent was not required due to the use of anonymized, public data.
Author contributions
Q-HZ: Data curation, Visualization, Conceptualization, Writing – original draft, Methodology, Investigation, Writing – review & editing. KD: Writing – review & editing, Methodology, Investigation, Writing – original draft, Conceptualization, Data curation. AL: Investigation, Writing – review & editing, Data curation. C-YZ: Investigation, Data curation, Writing – review & editing. RG: Methodology, Data curation, Writing – review & editing. PC: Methodology, Data curation, Writing – review & editing. W-SD: Data curation, Methodology, Writing – review & editing. Y-LZ: Writing – review & editing, Methodology, Data curation. S-ML: Project administration, Supervision, Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Acknowledgements
The authors would like to thank those who uploaded videos on both platforms.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Deyle GD, Allen CS, Allison SC, Gill NW, Hando BR, Petersen EJ, et al. Physical therapy versus glucocorticoid injection for osteoarthritis of the knee. N Engl J Med. (2020) 382:1420–9. doi: 10.1056/NEJMoa1905877
2. Katz JN, Arant KR, Loeser RF. Diagnosis and treatment of hip and knee osteoarthritis A review. JAMA-J Am Med Assoc. (2021) 325:568–78. doi: 10.1001/jama.2020.22171
3. Cross M, Smith E, Hoy D, Nolte S, Ackerman I, Fransen M, et al. The global burden of hip and knee osteoarthritis: estimates from the global burden of disease 2010 study. Ann Rheum Dis. (2014) 73:1323–30. doi: 10.1136/annrheumdis-2013-204763
4. Lv Y, Sui L, Lv H, Zheng J, Feng H, Jing F. Burden of knee osteoarthritis in China and globally from 1992 to 2021, and projections to 2030: a systematic analysis from the global burden of disease study 2021. Front Public Health. (2025) 13:1543180. doi: 10.3389/fpubh.2025.1543180
5. Fu M, Zhou H, Li Y, Jin H, Liu X. Global, regional, and national burdens of hip osteoarthritis from 1990 to 2019: estimates from the 2019 global burden of disease study. Arthritis Res Ther. (2022) 24:8. doi: 10.1186/s13075-021-02705-6
6. Rathbun AM, Ryan AS, Shardell MD, Dong Y, Beamer BA, Luborsky M, et al. Association between depressive symptoms and self-reported physical activity in persons with knee osteoarthritis. J Rheumatol. (2025) 52:505–11. doi: 10.3899/jrheum.2024-0940
7. Nalamachu S, Robinson RL, Viktrup L, Cappelleri JC, Bushmakin AG, Tive L, et al. Pain severity and healthcare resource utilization in patients with osteoarthritis in the United States. Postgrad Med. (2021) 133:10–9. doi: 10.1080/00325481.2020.1841988
8. Abramoff B, Caldera FE. Osteoarthritis pathology, diagnosis, and treatment options. Med Clin North Am (2020) 104:293. doi: 10.1016/j.mcna.2019.10.007
10. Duong V, Oo WM, Ding C, Culvenor AG, Hunter DJ. Evaluation and treatment of knee pain A review. JAMA. (2023) 330:1568–80. doi: 10.1001/jama.2023.19675
11. Tore NG, Oskay D, Haznedaroglu S. The quality of physiotherapy and rehabilitation program and the effect of telerehabilitation on patients with knee osteoarthritis. Clin Rheumatol. (2023) 42:903–15. doi: 10.1007/s10067-022-06417-3
12. Bennell KL, Lawford BJ, Keating C, Brown C, Kasza J, Mackenzie D, et al. Comparing video-based, telehealth-delivered exercise and weight loss programs with online education on outcomes of knee osteoarthritis: a randomized trial. Ann Intern Med (2022) 175:198–209. doi: 10.7326/M21-2388
13. Jia X, Pang Y, Liu LS. Online health information seeking behavior: a systematic review. Healthc Basel Switz. (2021) 9:1740. doi: 10.3390/healthcare9121740
14. Zhao YC, Zhao M, Song S. Online health information seeking behaviors among older adults: systematic scoping review. J Med Internet Res (2022) 24:e34790. doi: 10.2196/34790
15. Ye C, Wang Z, Wu M, Kang R, Yuan F, Chen C. Behavioral drivers of AI nursing acceptance in the greater bay area: a family-caregiver perspective on trust and risk. Front Public Health. (2025) 13:1650804. doi: 10.3389/fpubh.2025.1650804
16. Li H, Li D, Zhai M, Lin L, Cao Z. Associations among online health information seeking behavior, online health information perception, and health service utilization: cross-sectional study. J Med Internet Res. (2025) 27:e66683. doi: 10.2196/66683
17. Goff AJ, Barton CJ, Merolli M, Zhang Quah AS, Ki-Cheong Hoe C, De Oliveira Silva D. Comprehensiveness, accuracy, quality, credibility and readability of online information about knee osteoarthritis. Health Inf Manag J (2023) 52:185–93. doi: 10.1177/18333583221090579
18. Kong W, Song S, Zhao YC, Zhu Q, Sha L. Tiktok as a health information source: assessment of the quality of information in diabetes-related videos. J Med Internet Res. (2021) 23:e30409. doi: 10.2196/30409
19. Suarez-Lledo V, Alvarez-Galvez J. Prevalence of health misinformation on social Media: systematic review. J Med Internet Res. (2021) 23:e17187. doi: 10.2196/17187
20. Liang Y, Xia J, Huo W, Liu B, Wang Z, Ding Y, et al. Video quality assessment and analysis of gastroesophageal reflux disease on TikTok and Bilibili: cross-sectional study. J Multidiscip Healthc. (2024) 17:5927–39. doi: 10.2147/JMDH.S485781
21. Denniss E, Lindberg R, McNaughton SA. Quality and accuracy of online nutrition-related information: a systematic review of content analysis studies. Public Health Nutr. (2023) 26:1345–57. doi: 10.1017/S1368980023000873
22. Segado Fernández S, Jiménez Gómez B, Jiménez Hidalgo P, Lozano-Estevan MDC, Herrera Peco I. Disinformation about diet and nutrition on social networks: a review of the literature. Nutr Hosp. (2025) 42:366–75. doi: 10.20960/nh.05533
23. John JN, Gorman S, Scales D, Gorman J. Online misleading information about women’s reproductive health: a narrative review. J Gen Intern Med. (2025) 40:1123–31. doi: 10.1007/s11606-024-09118-6
24. Wang M, Yao N, Wang J, Chen W, Ouyang Y, Xie C. Bilibili, Tiktok, and YouTube as sources of information on gastric cancer: assessment and analysis of the content and quality. BMC Public Health. (2024) 24:57. doi: 10.1186/s12889-023-17323-x
25. Liu Z, Chen Y, Lin Y, Ai M, Lian D, Zhang Y, et al. YouTube/Bilibili/Tiktok videos as sources of medical information on laryngeal carcinoma: cross-sectional content analysis study. BMC Public Health. (2024) 24:1594. doi: 10.1186/s12889-024-19077-6
26. Huang J, Zhang S, Xiao Q, Cao Y, Li B. YouTubeTM as a source of information for Candida auris infection: a systematic review. BMC Public Health. (2020) 20:832. doi: 10.1186/s12889-020-08731-4
27. Morena N, Ahisar Y, Wang X, Diana N, Rentschler CA, Meguerditchian AN. Are YouTube videos a reliable information source for young women with metastatic breast cancer? J Clin Oncol. (2022) 40:244. doi: 10.1200/JCO.2022.40.28_suppl.244
28. Boyer C, Selby M, Scherrer J-R, Appel RD. The health on the net code of conduct for medical and health websites. Comput Biol Med. (1998) 28:603–10. doi: 10.1016/S0010-4825(98)00037-7
29. Cetinavci D, Yasar V, Yucel A, Elbe H. Evaluation of the usage of YouTube videos about histology and embryology as an educational material. Anat Histol Embryol. (2022) 51:810–7. doi: 10.1111/ahe.12859
30. Kayabaşı M, Köksaldı S, Durmaz Engin C. Evaluating the quality and reliability of YouTube videos on myopia: a video content analysis. Int Ophthalmol. (2024) 44:329. doi: 10.1007/s10792-024-03250-2
31. Pennycook G, Rand DG. Fighting misinformation on social media using crowdsourced judgments of news source quality. Proc Natl Acad Sci U S A. (2019) 116:2521–6. doi: 10.1073/pnas.1806781116
32. Yeung A, Ng E, Abi-Jaoude E. Tiktok and attention-deficit/hyperactivity disorder: a cross-sectional study of social Media content quality. Can J Psychiatry Rev Can Psychiatr. (2022) 67:899–906. doi: 10.1177/07067437221082854
33. Dimitroyannis R, Fenton D, Cho S, Nordgren R, Pinto JM, Roxbury CR. A social media quality review of popular sinusitis videos on TikTok. Otolaryngol Head Neck Surg. (2024) 170:1456–66. doi: 10.1002/ohn.688
34. Niu Z, Hao Y, Yang F, Jiang Q, Jiang Y, Zhang S, et al. Quality of pancreatic neuroendocrine tumor videos available on TikTok and Bilibili: content analysis. JMIR Form Res. (2024) 8:e60033. doi: 10.2196/60033
35. Etumuse B, Greer M, Onyemachi J, El-Abed Y, Kamma S, Shah JD, et al. Medical misinformation and quality of public video content on Cannabis for chronic pain management: a cross-sectional analysis of the YouTube platform. J Pain Res. (2024) 17:3577–86. doi: 10.2147/JPR.S479200
36. Osman W, Mohamed F, Elhassan M, Shoufan A. Is YouTube a reliable source of health-related information? A systematic review. BMC Med Educ. (2022) 22:382. doi: 10.1186/s12909-022-03446-z
37. Izquierdo-Condoy JS, Arias-Intriago M, Mosquera-Quiñónez M, Melgar Muñoz FP, Jiménez-Ascanio M, Loaiza-Guevara V, et al. Assessing the educational impact and quality of medical microvideos on TikTok: the case of Latin America. Med Educ Online. (2025) 30:2474129. doi: 10.1080/10872981.2025.2474129
38. Lu X, Zhang R, Wu W, Shang X, Liu M. Relationship between internet health information and patient compliance based on trust: empirical study. J Med Internet Res. (2018) 20:e253. doi: 10.2196/jmir.9364
39. Zhang Z, Yang H, He J, Lu X, Zhang R. The impact of treatment-related internet health information seeking on patient compliance. Telemed J E-Health. (2021) 27:513–24. doi: 10.1089/tmj.2020.0081
40. Moseng T, Vliet Vlieland TPM, Battista S, Beckwée D, Boyadzhieva V, Conaghan PG, et al. EULAR recommendations for the non-pharmacological core management of hip and knee osteoarthritis: 2023 update. Ann Rheum Dis. (2024) 83:730–40. doi: 10.1136/ard-2023-225041
41. Gao Y, Liu F, Gao L. Echo chamber effects on short video platforms. Sci Rep. (2023) 13:6282. doi: 10.1038/s41598-023-33370-1
42. Christiansen A, Craythorne S-L, Crawford P, Larkin M, Gohil A, Strutt S, et al. Multimodal analysis of stories told by mental health influencers on TikTok. Health Expect Int J Public Particip Health Care Health Policy. (2025) 28:e70226. doi: 10.1111/hex.70226
Keywords: osteoarthritis, social media, quality of health information, TikTok, Bilibili, health communication, patient education
Citation: Zuo Q-H, Du K, Li A, Zhang C-Y, Guo R, Chen P, Du W-S, Zuo Y-L and Li S-M (2025) Quality and reliability of osteoarthritis-related health information on short video platforms: a cross-sectional comparative study of TikTok and Bilibili. Front. Digit. Health 7:1623247. doi: 10.3389/fdgth.2025.1623247
Received: 5 May 2025; Accepted: 22 October 2025;
Published: 7 November 2025.
Edited by:
Aikaterini Kassavou, University of Bedfordshire, United KingdomReviewed by:
Liu Yilong, University of Modena and Reggio Emilia, ItalyChenxi Ye, United International College, China
Copyright: © 2025 Zuo, Du, Li, Zhang, Guo, Chen, Du, Zuo and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Shu-Ming Li, bGlzaHVtaW5nQGJqemhvbmd5aS5jb20=
†These authors have contributed equally to this work and share first authorship
Qi-Heng Zuo1,2,†