Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Public Health

Sec. Digital Public Health

This article is part of the Research TopicAdvancing Healthcare AI: Evaluating Accuracy and Future DirectionsView all 19 articles

Implicit Bias in Digital Health: Systematic Biases in Large Language Models' Representation of Global Public Health Attitudes and Challenges to Health Equity

Provisionally accepted
Yuan  GaoYuan Gao1Yican  FengYican Feng2*Surng  Gahb JahngSurng Gahb Jahng1
  • 1Chung-Ang University, Dongjak-gu, Republic of Korea
  • 2Zhejiang Fashion Institute of Technology, Ningbo, China

The final, formatted version of the article will be published soon.

As emerging instruments in digital health, large language models (LLMs) assimilate values and attitudes from human-generated data, thereby possessing the latent capacity to reflect public health perspectives. This study investigates into the representational biases of LLMs through the lens of health equity. We propose and empirically validate a three-dimensional explanatory framework encompassing Data Resources, Opinion Distribution, and Prompt Language, positing that prompts are not just communicative media but critical conduits that embed cultural context. Utilizing a selection of prominent LLMs from the United States and China-namely Gemini 2.5 Pro, GPT-5, DeepSeek-V3, and Qwen 3. We conduct a systematic empirical analysis of their performance in representing health attitudes across diverse nations and demographic strata. Our findings demonstrate that: first, the accessibility of data resources is a primary determinant of an LLM's representational fidelity for internet users and nations with high internet penetration. Second, a greater consensus in public health opinion correlates with an increased propensity for the models to replicate the dominant viewpoint. Third, a significant "native language association" is observed, wherein Gemini 2.5 Pro and DeepSeek-V3 exhibit superior performance when prompted in their respective native languages. Conversely, models with enhanced multilingual proficiencies, such as GPT-5.0 and Qwen 3, display greater cross-lingual consistency. This paper not only quantifies the degree to which these leading LLMs reflect public health attitudes but also furnishes a robust analytical pathway for dissecting the underlying mechanisms of their representational biases. These findings bear profound implications for the advancement of health equity in the artificial intelligence era.

Keywords: Large language models, representational bias, Digital Health, health equity, algorithmic audit

Received: 14 Sep 2025; Accepted: 17 Nov 2025.

Copyright: © 2025 Gao, Feng and Jahng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Yican Feng, 2024700038@zjff.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.