- 1Pediatric Pain Pathway Lab, Department of Anesthesia, Critical Care and Pain Medicine, Boston Children’s Hospital/Harvard Medical School, Boston, MA, United States
- 2Department of Anesthesia, Critical Care and Pain Medicine, Harvard Medical School, Boston Children’s Hospital, Boston, MA, United States
- 3Department of Psychology, Uppsala University, Uppsala, Sweden
- 4Brain and Eye Pain Imaging Lab, Department of Anesthesia, Critical Care and Pain Medicine, Boston Children’s Hospital/Harvard Medical School, Boston, MA, United States
- 5Department of Ophthalmology, Boston Children's Hospital/Harvard Medical School, Boston, MA, United States
Clinical research is no longer a monopolistic environment wherein patients and participants are the sole voice of information. The introduction and acceleration of AI-based methods in healthcare is creating a complex environment where human-derived data is no longer the sole mechanism through which researchers and clinicians explore and test their hypotheses. The concept of self-agency is intimately tied into this, as generative data does not encompass the same person-lived experiences as human-derived data. The lack of accountability and transparency in recognizing data sources supporting medical and research decisions has the potential to immediately and negatively impact patient care. This commentary considers how self-agency is being confronted by the introduction and proliferation of generative AI, and discusses future directions to improve, rather than undermine AI-fueled healthcare progress.
Introduction
The integration of generative AI technologies into clinical research creates a chimeric environment that can distort the concept of self-agency from the perspective of the participant/patient and the experimenter/clinician. A chimeric environment refers to a hybrid environment wherein both human derived and synthetically created data are placed in equal weighting to support algorithm development. This is performed on occasion during algorithm training which has been shown to be an efficient way of improving artificial intelligence models, relative to using only non-synthetic data (1, 2). Artificial intelligence reflects the ability of a machine to perform beyond its explicit programming. Algorithms act to: (1) learn the core features that represent an input data set and (2) leverage this reduced set of features and latent space to generate novel products (e.g., text or images). These algorithms have the capacity to deviate regenerated data to extend beyond the original source document(s), to produce novel items that are changed in some fundamental supervised, or unsupervised way. In the context of generative AI, patient agency is a central concern in so much as it relates to the extent to which individuals can make autonomous, informed decisions about their health. Patient agency is being redefined and shaped not only by the accessibility, and interpretability of generative AI models, but also by the degree to which patients can trust, question, and act upon the information provided.
The recent growth of AI technologies in healthcare, based largely on the advent of more capable graphic processor units (GPU) and the digitization of healthcare data, has yielded a double edge sword; at once both enabling models that improve patient care and mitigate medical resource limitations (3), while also decoupling the patient-clinician relationship. This previously transparent interaction has evolved opaque characteristics, directly challenging the concept of patient agency (see (4–6) for review). It is at this point that persons working with generative AI seem to have the largest impact. As we begin to rely more on elements like telemedicine, and remote survey work, factors such as agent authentication are becoming a credible concern towards medical accuracy and accountability. The use of techniques such as generative adversarial networks to produce diagnostic images can create valuable training environments, at times surpassing the use of non-synthetic data (7); however, they may also create environments that extend past healthy human constraints and promote unhealthy lifestyles. We outline relevant concerns regarding the integration of generative AI methods with healthcare regarding patient agency.
To whom am I speaking with?
The power and capacity of generative AI in terms of large language models is growing rapidly. At the writing of this commentary, Google continues to deploy new iterations of its Gemini program that rivals those of competitors at OpenAI (8) and Claude who also continue the release of revised versions of their core programs. Having been trained on the entirety of text from the internet, these models are highly capable of text-based interface, including being able to provide written prompts, and generating answers, either to describe their response (e.g., explaining a script of code it generated) or for the response itself (e.g., writing an essay). As output is derived from human-based sources, researchers have begun to question, can algorithms such as large language models replace the use of human participants in research (9). Considering that generative models can adapt to text, images, and auditory modalities, the implications for healthcare are immense.
In clinical research, the agent under interest is the person for whom we want to explore a research question and develop our knowledge. This person can be an individual or part of a larger cohort; however, they have unique attributes that address how they experience their condition including their (medical) history, their personal experiences, and how their condition has shaped their perceptions on how their thoughts and actions are self-generated. This, in effect, represents their agency. The application of generative models directly undermines this element of agency by removing the person, or agent, and promoting the viewpoints or biases from their training data. Generative AI methods have the capacity to propagate ethical and gender stereotypes in text-based materials including patient reports, education materials, and patient communication [see (10)]. It has also been argued that generative AI methods could trigger, perpetuate, or exacerbate experiences in vulnerable populations, especially those reliant on such healthcare devices for information (11). Even before the proliferation of generative AI, a cottage industry of survey takers in economically-disadvantaged countries presented concerns regarding the validity of crowd-sourced data from internet-based surveys (12). One study found that approximately 33%–46% of survey takers would leverage the use of large language models in their responses (13) while other more recent research has found issues with up to 96% of their online survey findings (14). This concern has already led to a market of online survey tools meant to counter this influence, such as the Mturk Toolkit from CloudResearch (15). Medical doctors and researchers are turning more towards big data to fuel efforts at precision medicine (16) using online data gathering with the use of surveys for medical research; however, with the proliferation of generative AI this voice may not always be authentic. This remote disconnection produces a void wherein LLM-based approaches can be integrated with existing data collection strategies and begin to control the trajectory of research findings. That is to say, if a research survey was undertaken to address a target cohort (e.g., pediatrics), the findings of such a survey may reflect training data on an established cohort (e.g., adults), therein misleading treatment efforts aimed at integrating the agency of the target cohort. In effect, what we as researchers and clinicians are observing when interfacing with such artificial agents, is a mosaic of integrated and echoed perspectives sewn together through layered neural networks.
Shifting our diagnostic basis
Generative products are becoming ubiquitous, with access to the engines responsible for their creation coming standard on consumer products such as laptops and phones. The basis of generative AI techniques is to use the knowledge we have gathered to date to train a model that can create new information; information that has to date not existed based on the rules that are composed in the model itself. Intrinsic to this process are the mechanisms through which generative AI algorithms use to extrapolate existing information into novel productions. There is growing use of generative AI methods in healthcare for the creation of novel images of artificial bones (17), and brains (18), though their clinical integration has yet to be formally established. Importantly, as current diagnostic markers are based on normative data sets, the means through which we integrate this information into clinical workflow is imperative to consider.
Normative data sets underlie our understanding of deviance from a healthy standard. It is how we judge neurodevelopment, psychological health, and even brain health. To date, these databases, which have been composed of data collected from real world persons, are largely publicly available and many of us may have even interacted with them at our doctor's office [e.g., DSM-5 (19)]. However, we now face the growing realization that such databases could either be mixed with, or be purely generated from synthetic data sources [see (20)]. As an example, based on financial reasons, it may make more sense to train an algorithm aimed at detecting brain cancer from synthetic databases of MRI, or hybrid databases; however, the origin of such synthetic databases may be from only one sex, or the product of such algorithms may not represent cancerous growths, but rogue attempts at mimicking natural cortical folding patterns in the brain. This proposition is unique from using ML/AI to predict growth curves, for example, and refers rather to generating synthetic text responses or bone x-rays that would then be fed into growth curve models alongside human-derived data. The mere presence of a single data point that was not obtained from a verified human means that such models are hybrid-based and should be interpreted as such. Because such generative algorithms operate on a probability level, they have the inherent risk, for example, of generating data that is not theoretically possible in a human state, or data that does not consider factors such as race (21). Without explicit knowledge of how such algorithms were defined or trained, a clinician could easily be coerced into an improper diagnosis, much as if improper reagents were used to perform a standard blood test. As such, the basis from which we judge informed decisions can be manipulated, in a direction that hinders the clinical process. This chimeric integration of data sources could mean that therapeutic goals are shifted, and disease states are missed.
Future directions
To date, we have not been labeling data as AI produced vs. “uniquely human.” And why would we? This has not been a concern until recently. Attempts at chimeric databases have largely focused on non-healthcare data, producing images such as flowers. As the capacity of such algorithms grows alongside data storage capacity, it will be critical to isolate synthetic from non-synthetic databases and regulate their clinical integration. In terms of large language models, there are notable modern attempts to evaluate their capacity in clinical settings using evaluations such as CRAFT-MD (Conversational Reasoning Assessment Framework for Testing in Medicine) that highlight the exceptional capacity, but continued deficiencies (e.g., GPT-4 and LlaMA02-7b) in real world diagnostics (22). Just as it is next-to impossible to differentiate a piece of text written through AI from that of a human agent, it will be soon next to impossible to differentiate a synthetically created MRI from that of a real person. Proper regulation of this domain will require participation from funding agencies, ethics review boards, and publishers to ensure we are not marching towards a false flag.
It is clear now that generative AI models are reaching greater complexity and can interact with complex psychological and physiological data and generate a dataset equivalent to modern normative value datasets in a fraction of the time. Their potential for changing the clinical landscape is undoubtably immense and should be pursued. As we attempt to satisfy the immense hunger of such algorithms for data, we need to be mindful of the data being fed through iterations of such models and how that defines the outcomes we observe. Context is everything. Clinical research needs to leverage the power and speed of AI particularly as it relates to vulnerable populations; however, proceeding without caution could lead us to lose the agent and with it, their agency.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.
Author contributions
SH: Conceptualization, Investigation, Writing – original draft, Writing – review & editing. VF: Conceptualization, Investigation, Writing – original draft, Writing – review & editing. EM: Conceptualization, Investigation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Acknowledgments
We would like to acknowledge the support of the Department of Anesthesia, Critical Care and Pain Medicine at Boston Children's Hospital for financial support of this publication (SH and EM).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Kim J, Kim D, Lee S, Chi S. Hybrid DNN training using both synthetic and real construction images to overcome training data shortage. Autom Constr. (2023) 149. doi: 10.1016/j.autcon.2023.104771
2. Zhezherau A, Yanockin A. Hybrid Training Approaches for LLMs: Leveraging Real and Synthetic Data to Enhance Model Performance in Domain-Specific Applications. (2024). Available at: https://doi.org/10.48550/arXiv.2410.09168 (Accessed January 15, 2025).
3. Meskó B, Hetényi G, Győrffy Z. Will artificial intelligence solve the human resource crisis in healthcare? BMC Health Serv Res. (2018) 18:545. doi: 10.1186/s12913-018-3359-4
4. Limerick H, Coyle D, Moore JW. The experience of agency in human-computer interactions: a review. Front Hum Neurosci. (2014) 8:1–10. doi: 10.3389/fnhum.2014.00643
5. Madary M. The illusion of agency in human–computer interaction. Neuroethics. (2022) 15:16. doi: 10.1007/s12152-022-09491-1
6. Yu H, Du S, Kurien A, Van Wyk BJ, Liu Q. The sense of agency in human–machine interaction systems. Appl Sci. (2024) 14:7327. doi: 10.3390/app14167327
7. Natarajan SA, Madden M. Hybrid synthetic data generation pipeline that outperforms real data. J Electron Imaging. (2023) 32:1–20. doi: 10.1117/1.JEI.32.2.023011
9. Dillion D, Tandon N, Gu Y, Gray K. Can AI language models replace human participants? Trends Cogn Sci. (2023) 27:597–600. doi: 10.1016/j.tics.2023.04.008
10. Currie G, John G, Hewis J. Gender and ethnicity bias in generative artificial intelligence text-to-image depiction of pharmacists. Int J Pharm Pract. (2024) 32(6):524–31. doi: 10.1093/ijpp/riae049
11. Abdulai A. Is generative AI increasing the risk for technology-mediated trauma among vulnerable populations? Nurs Inq. (2025) 32:e12686. doi: 10.1111/nin.12686
12. Burnette CB, Luzier JL, Bennett BL, Weisenmuller CM, Kerr P, Martin S, et al. Concerns and recommendations for using Amazon MTurk for eating disorder research. Int J Eat Disord. (2022) 55:263–72. doi: 10.1002/eat.23614
13. Veselovsky V, Ribeiro MH, West R. Artificial Artificial Artificial Intelligence: Crowd Workers Widely Use Large Language Models for Text Production Tasks. (2023). Available at: https://doi.org/10.48550/arXiv.2306.07899 (Accessed January 15, 2025).
14. Kay CS. Why you shouldn’t trust data collected on MTurk. (2024). Available at: https://doi.org/10.31234/osf.io/zs6pk (Accessed January 15, 2025).
15. Litman L, Robinson J, Abberbock T. Turkprime.com: a versatile crowdsourcing data acquisition platform for the behavioral sciences. Behav Res Methods. (2017) 49:433–42. doi: 10.3758/s13428-016-0727-z
16. Roberts MC, Holt KE, Del Fiol G, Baccarelli AA, Allen CG. Precision public health in the era of genomics and big data. Nat Med. (2024) 30:1865–73. doi: 10.1038/s41591-024-03098-0
17. Prezja F, Paloneva J, Pölönen I, Niinimäki E, Äyrämö S. Deepfake knee osteoarthritis x-rays from generative adversarial neural networks deceive medical experts and offer augmentation potential to automatic classification. Sci Rep. (2022) 12:18573. doi: 10.1038/s41598-022-23081-4
18. Gong C, Jing C, Chen X, Pun CM, Huang G, Saha A, et al. Generative AI for brain image computing and brain network computing: a review. Front Neurosci. (2023) 17:1203104. doi: 10.3389/fnins.2023.1203104
19. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders, 5th ed. Washington, DC: American Psychiatric Association (2022).
20. Waqas A, Bui MM, Glassy EF, Naqa IE, Borkowski P, Borkowski AA, et al. Revolutionizing digital pathology with the power of generative artificial intelligence and foundation models. Lab Invest. (2023) 103:100255. doi: 10.1016/j.labinv.2023.100255
21. Dancy CL. How to use generative AI more responsibly. Nat Rev Psychol. (2024) 3:501–2. doi: 10.1038/s44159-024-00339-4
Keywords: agency, artificial intelligence, brain imaging, survey, database
Citation: Holmes SA, Faria V and Moulton EA (2025) Generative AI in healthcare: challenges to patient agency and ethical implications. Front. Digit. Health 7:1524553. doi: 10.3389/fdgth.2025.1524553
Received: 3 February 2025; Accepted: 27 May 2025;
Published: 18 June 2025.
Edited by:
Filippo Gibelli, University of Camerino, ItalyReviewed by:
Adewunmi Akingbola, University of Cambridge, United KingdomCopyright: © 2025 Holmes, Faria and Moulton. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Scott A. Holmes, c2NvdHQuaG9sbWVzQGNoaWxkcmVucy5oYXJ2YXJkLmVkdQ==