Your new experience awaits. Try the new design now and help us make it even better

PERSPECTIVE article

Front. Med., 26 August 2025

Sec. Regulatory Science

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1651934

Who moved my scan? Early adopter experiences with pre- and post-market healthcare AI regulation challenges


Aviad Raz*Aviad Raz1*Yael InbarYael Inbar2Netta AvnoonNetta Avnoon3Liat Bela Lifshitz-MilwidskyLiat Bela Lifshitz-Milwidsky1Barkan HofmanBarkan Hofman1Asaf HonigAsaf Honig4Ziv PazZiv Paz4
  • 1Department of Sociology and Anthropology, Ben-Gurion University of the Negev, Be’ersheba, Israel
  • 2Technology Management and Information Systems, Coller School of Management, Tel Aviv University, Tel Aviv-Yafo, Israel
  • 3Western University, London, ON, Canada
  • 4Clalit Health Services, Tel Aviv-Yafo, Israel

Real-world clinical experience provides a much-needed opportunity for deep learning AI algorithms to evolve and improve. Yet, it also constitutes a regulatory challenge, since such potential for learning may essentially change the algorithm and introduce new biases. We focus on the gaps between “lifecycle” regulation and implementation from the perspective of the deployers, addressing three interconnected dimensions: (a) How precautionary regulation affects AI deployment in healthcare, (b) How healthcare providers view explainable AI (XAI), and (c) How AI deployment influences, and is influenced by, team routines in clinical settings. We conclude by suggesting ways in which the ends of healthcare AI regulation and deployment can successfully meet.

1 Introduction

The advent of potential artificial intelligence (AI) automation in healthcare paradoxically leads to a growing need for human supervision both before and after deployment (1). In contrast to conventional, static medical devices, the unique power of AI and specifically machine learning (ML)-based algorithms lies in their ability for dynamic, continuous learning, before and after deployment. Real-world experience evidently provides an important opportunity for evolving and improving for such deep learning algorithms. Yet, it also constitutes a regulatory challenge, since such potential for learning may essentially change the algorithm and introduce new biases (2). To avoid re-approval by the US Food and Drug Administration (FDA), the learning capability of many healthcare AI systems is shut down upon deployment, thus removing the risk along with the potency. The challenge is, therefore, how to match such adaptive systems with an equally adaptive regulation? There is little theorization, let alone research, addressing the gap between the design of AI “lifecycle” regulation and its practical application in clinical settings, including the role of healthcare providers as early AI deployers. This Perspective Article sets out to fill some of these literature voids.

According to the new EU AI Act (EU AIA), healthcare AI falls under the “high-risk” category and requires oversight by deployers. Yet, clarification is needed regarding how exactly should deployers oversee such “adaptive” or “self-learning” systems. Additionally, as deployers encompass both healthcare organizations and individual clinicians, clarifying this distinction is important as the challenges and motivations for each can differ. Recognizing this challenge, the US FDA is in the process of shifting from the traditional medical device regulatory framework and has introduced the AI/ML-based software as a medical device action plan in 2021, employing a “total product lifecycle” approach to include post-market phases–covering data quality, algorithm transparency, and robust change management (3). In 2024, the FDA further issued guidelines for an adaptive regulatory framework of AI/ML-enabled device software (4, 5). This framework potentially lets developers address AI’s continuous evolution without repeatedly seeking new FDA approvals. However, on-site post-deployment evaluations of AI in diverse clinical settings, requiring the implementation of robust, real-time monitoring mechanisms, remain inconsistent and underdeveloped (2, 6). This often means that healthcare providers-cum-deployers must step in to develop local protocols for quality assurance and recalibration.

To address these challenges of pre- and post-market regulation we focus here on the gaps from the perspective of the deployers. Looking at AI regulation in healthcare from the point of view of deployers, we focus on three interconnected dimensions: (a) how precautionary regulation affects AI deployment in healthcare, (b) how healthcare providers’ assumptions about explainable AI (XAI) relate to clinical trust, and (c) how AI deployment influences, and is influenced by, team routines in clinical settings. We conclude by suggesting ways in which the ends of healthcare AI regulation and deployment can successfully meet.

2 How precautionary regulation affects AI deployment in healthcare

Food and Drug Administration and the EU-AIA regulatory frameworks focus on pre-market approval and validation yet largely fail to provide a workable solution regarding the post-deployment need for continuous monitoring, re-training and re-validation of AI models (2). Scholars have warned that as AI models are exposed to new data in clinical settings, their performance may degrade or alter overtime, necessitating ongoing oversight (7). Recognizing this, the FDA asked developers to be accountable for the real-world performance of their AI systems and update the FDA on the changes in terms of performance and input (8). The significant financial and administrative requirements implied by such re-assessment almost always lead to avoiding system retraining, or re-validation, after deployment, compensated by selective use patterns. An illustrative case is ChestLink, an X-ray imaging AI, which in 2022 was the first fully autonomous healthcare AI to receive the EU-CE mark certificate, including post-deployment AI re-training. The catch is that ChestLink is only allowed to autonomously produce reports for healthy patients (9). It does not autonomously analyze or report on chest X-rays with abnormalities. Pooling real-world evidence from an adaptive Chestlink could have shown the trade-off between the risk reduced by suppressing its utility versus letting it learn from diagnosing all cases, under clinicians’ supervision.

The need for in situ customization of healthcare AI tools gives way to experimental strategies, including vendor-appointed “champions” selected from healthcare deployers as well as on-site feedback loops, with physicians taking the role of the “AI’s QA testers,” figuring out in which areas it is good enough and where it is over-sensitive and over-alerting due to “factory settings.” Nevertheless, such customization requires further systematic and standardized monitoring of bias metrics in model outputs and selective use by physicians.

Recent EU regulation also asks for AI transparency and explainability. Art. 13 of the EU AIA states that sufficient transparency shall be provided to those using the system under their authority (10). Art. 86 of the AI Act (Right to explanation) states that patients have the right to obtain clear and meaningful explanations of the role of the AI system in the decision-making procedure from the deployer (10). However, such provisions may be interpreted as a right to a partial explanation only of the role of AI, rather than its functioning. The lack of specific regulation concerning explainability, reflecting the lack of consensus on what it should consist of, may lead to inadvertent consequences, as explored in the next section.

3 How healthcare providers view explainable AI (XAI)

Defining explainability is a debated issue. Defined by one expert group as “the ability to explain both the technical processes of an AI system and the related human decisions” (11), others see explainability more broadly as a combination of both intelligibility and accountability (12). Transparency, closely related to explainability, is seen by some as a more passive characteristic of the system (13), while others claim that transparency (rather than explainability) relates to the initial training dataset and system architecture (14).

Is explainability a necessary condition for trust? Practically, the value of explainability requires a trade-off, e.g., with accuracy (15). At the relatively early stage of deployment that we are in, healthcare professionals arguably have significantly limited trust in AI tools, especially as they are charged with their QA. Explainable AI (XAI) was found to have the same influence on doctors’ decision-making as AI by itself (16). Nevertheless, explainability is arguably a liability issue. If the AI makes a mistake that causes harm to a patient, who is responsible? In addition, the opaque nature of AI/ML may hinder the conveyance of predictions to the patient, creating in the future a potential flood of medical malpractice lawsuits. Thus, clinicians’ preference for AI accuracy rather than explainability may very well change as liability models evolve, increasing the demand for XAI as a tool for creating a defensible record of clinical decision-making.

Expecting every clinical AI system to provide an explanation of its actions imposes a considerable limitation. Contemporary AI technologies markedly outperform older AI technology but are often not able to provide explanations understandable to humans. Regulators do not currently require human-comprehensible explanations for AI in other industries that have potential risks, such as autonomous vehicles. Health providers are often more concerned with the prediction power and efficiency of AI than with its explainability, especially due to time constrains and the need to be efficient with the exponential increase in the workload worldwide.

Post-deployment, a pragmatic shift is arguably emerging from explainability as an inherent characteristic of an algorithm, to explainability as an emergent and approximated property, increasingly tweaked around prediction power. In image detection algorithms, usually Convolutional Neural Networks, their first layers will contain references to shading and edge detection. The human doctor might never have to explicitly define an edge or a shadow, but their clustering may be referred to as “Grade X occlusion” of a blood vessel. The recent suggestion (17) to train medical AI to speak in “higher level” diagnostic concepts, such as “stage 1 cancer” rather than “X density of pixels in location Y,” arguably promoting more explainability and therefore more trust among deployers, reflects certain pre-assumptions regarding the perceived positive value of XAI and the dynamics of clinical teamwork. Is there really a “crisis of explainability” (other than by its legal demand), that leads to distrust by physicians of clinical AI?

Physicians may also be critical of letting AI make clinical recommendations, acknowledging that “an alert” might “make you think of [an] alternative diagnosis” but should not “sway” or “influence” the clinician. From a perspective of professional jurisdiction, it is to be expected that physicians insist on retaining the AI system as a “second pair of eyes” which should not disrupt the clinician’s authority (18).

Regulation of healthcare AI also forbids its full autonomy because of the potential risk assumingly involved in crossing the line between autonomous AI-driven diagnosis and augmented AI-supported human decision-making. Consequently, some medical imaging AI systems may highlight pixels on the scan (“first order” data) but not circle around these pixels (“second order” interpretation); or present predictions (“% of recurring cancer”) which doctors do not share with patients because of their opaqueness.

4 How AI deployment influences, and is influenced by, team routines in clinical settings

A related discussion is rapidly emerging in the fields of Human-Robot Interaction and Human-Autonomy Teaming, exploring how the introduction of AI agents influences intragroup processes in human teams (19). Conflicting findings regarding the correlation between AI adoption and teamwork performance in human-only vs human-AI teams highlight the need to gain deeper understanding of intragroup routines and AI teaming as subject to mutual change.

Teaming with agentic AI often means disrupting team routines. A socio-organizational view of routines shows they are not only fixed and bureaucratic. The implementation of AI changes according to pre-deployment organizational routines and team relations, and these routines and relations also change with the implementation of AI. Routines (like morning routines) embody both administrative/organizational and social/personal patterns of actions. We are only beginning to understand the formalized and negotiated aspects of routine-making in human-AI teaming, including the effects of intervening variables such as team processes (distributed/hierarchical communication and leadership, team diversity in terms of ranks and specialties, routine formalization vs negotiability) as well as the AI adaptability – e.g., automation and augmentation features that are used wholly, partially or not at all, by deployers along different phases in the workflow cycle (2022).

As the new EU AIA takes effect, AI systems that mimic how human teams collaborate might improve explainability in high-risk situations of clinical medicine. This suggestion is an important move forward given that much of the research has so far focused on dyadic relationships between a human user and AI or treated the “team” as a unified unit (23). Our own on-going ethnographic study of the deployment of clinical AI in hospitals that follow FDA regulation problematizes this notion of mimicking how human teams collaborate by showing how AI-human teaming is negotiated vis-a-vis the spectrum of (in)formality of routines, from credentialing, licensing and reporting to putting together a medical performance that withstands trials of accountability. When routines are well-defined and formalized (“protocols”), teams are more likely to integrate the AI tool into their workflow, for example in allowing AI to autonomously detect and alert strokes from ICU scans; in contexts where routines are negotiated and interactions are voluntary, such as intragroup communication, AI’s agentic features, such as the documenting feature, might be restrained or avoided, for example not using the chat option to record consultations between clinicians.

5 The way forward

Medical progress in clinical AI requires parallel evolution of regulatory frameworks and clinical practice workflows. The precautionary nature of the regulation of AI in healthcare creates an experimental space where healthcare managers and physician deployers/users engage in configuring a usable framework for AI during an early adoption phase.

To become a sustainable approach in medical AI implementation, such “use first, trust later” deploying mindset we described cannot be left unmonitored. Major policy implications thus include the following:

(1) Regulation should ensure AI developers disclose their training data and model architecture for full transparency.

(2) Specific mechanisms could include mandating the use of shared data registries, across different demographic groups, for collaborative monitoring and evaluations of system performance by AI vendors, healthcare deployers, patient groups (especially those most likely to be harmed by these technologies) and independent auditing bodies. This would require a process of education and certification among these stakeholders.

(3) Hospital administrators and healthcare deployers should evaluate teaming routines with AI, and inform AI developers.

(4) Explainable AI should be developed within practical medico-legal framings of accountability. A clear disclosure, and even choice, regarding the trade-off between explainability and accuracy in specific healthcare AI systems should be provided to patients.

Striking a balance between regulation and beneficial use, accuracy and explainability, formal and negotiated AI-human teaming, requires empirical and interdisciplinary efforts. The challenge of post-deployment AI oversight in healthcare does not belong to one group such as either AI vendors or healthcare deployers, but involves a diverse range of stakeholders, including AI manufacturers, healthcare providers, patients - especially those from historically underrepresented groups - and regulatory bodies, fostering a system that is constantly learning and consistently adaptive.

Data availability statement

The original contributions presented in this study are included in this article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

AR: Conceptualization, Funding acquisition, Investigation, Project administration, Writing – original draft, Writing – review & editing. YI: Funding acquisition, Supervision, Writing – review & editing. NA: Formal analysis, Investigation, Writing – review & editing. LB: Investigation, Project administration, Writing – review & editing. BH: Investigation, Writing – review & editing. AH: Writing – review & editing. ZP: Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work received funding from ISF grant1325/23.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Correction note

This article has been corrected with minor changes. These changes do not impact the scientific content of the article.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Endsley MR. From here to autonomy: lessons learned from human-automation research. Human Factors. (2017) 59:5–27. doi: 10.1177/0018720816681350

PubMed Abstract | Crossref Full Text | Google Scholar

2. Abulibdeh R, Celi L, Sejdic E. The illusion of safety: a report to the FDA on AI healthcare product approvals. PLoS Digit Health. (2025) 4:e0000866. doi: 10.1371/journal.pdig.0000866

PubMed Abstract | Crossref Full Text | Google Scholar

3. US Food and Drug Administration. Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) Action Plan. Silver Spring, MA: US Food and Drug Administration (2021).

Google Scholar

4. US Food and Drug Administration. Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices. Silver Spring, MA: US Food and Drug Administration (2024).

Google Scholar

5. US Food and Drug Administration. Marketing Submission Recommendations for a Predetermined Change Control Plan for Artificial Intelligence/Machine Learning (AI/ML)-Enabled Device Software Functions. Silver Spring, MA: US Food and Drug Administration (2023/4).

Google Scholar

6. Thomas L, Hyde C, Mullarkey D, Greenhalgh J, Kalsi D, Ko J. Real-world post-deployment performance of a novel machine learning-based digital health technology for skin lesion assessment and suggestions for post-market surveillance. Front Med. (2023) 10:1264846. doi: 10.3389/fmed.2023.1264846

PubMed Abstract | Crossref Full Text | Google Scholar

7. Feng J, Phillips RV, Malenica I, Bishara A, Hubbard AE, Celi LA, et al. Clinical artificial intelligence quality improvement: towards continual monitoring and updating of AI algorithms in healthcare. NPJ Digital Med. (2022) 5:66. doi: 10.1038/s41746-022-00611-y

PubMed Abstract | Crossref Full Text | Google Scholar

8. Harvey HB, Gowda V. How the FDA regulates AI. Acad Radiol. (2020) 27:58–61. doi: 10.1016/j.acra.2019.09.017

PubMed Abstract | Crossref Full Text | Google Scholar

9. Díaz-Rodríguez N, Del Ser J, Coeckelbergh M, López de Prado M, Herrera-Viedma E, Herrera F. Connecting the dots in trustworthy artificial intelligence: from AI Principles, ethics, and key requirements to responsible AI systems and regulation. Information Fusion. (2023) 99:101896. doi: 10.1016/j.inffus.2023.101896

Crossref Full Text | Google Scholar

10. European Union. European Union Regulation 2024/1689, Laying Down Harmonised Rules on Artificial Intelligence. Maastricht: EU (2024).

Google Scholar

11. European Commission. Directorate-General for Communications Networks, Content and Technology, Ethics Guidelines for Trustworthy AI. Brussels: European Commission (2019). doi: 10.2759/346720

Crossref Full Text | Google Scholar

12. Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion. (2020) 58:82–115. doi: 10.1016/j.inffus.2019.12.012

Crossref Full Text | Google Scholar

13. Esposito E. Does explainability require transparency? Sociologica. (2022) 16:17–27. doi: 10.6092/issn.1971-8853/15804

Crossref Full Text | Google Scholar

14. Whicher D, Ahmed M, Israni S. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril. Washington, DC: National Academy of Medicine (2023).

Google Scholar

15. Raz A, Heinrichs B, Avnoon N, Eyal G, Inbar Y. Prediction and explainability in AI: striking a new balance? Big Data Soc. (2024) 11: doi: 10.1177/20539517241235871

Crossref Full Text | Google Scholar

16. Nagendran M, Festor P, Komorowski M, Gordon A, Faisal A. Quantifying the impact of AI recommendations with explanations on prescription decision making. Npj Digit Med. (2023) 6:1–7. doi: 10.1038/s41746-023-00955-z

PubMed Abstract | Crossref Full Text | Google Scholar

17. Banerji C, Chakraborti T, Ismail A, Ostmann F, MacArthur B. Train clinical AI to reason like a team of doctors. Nature. (2025) 639:32–4. doi: 10.1038/d41586-025-00618-x

PubMed Abstract | Crossref Full Text | Google Scholar

18. Henry K, Kornfield R, Sridharan A, Linton R, Groh C, Wang T, et al. Human-machine teaming is key to AI adoption: clinicians’ experiences with a deployed machine learning system. NPJ Digit Med. (2022) 5:97. doi: 10.1038/s41746-022-00597-7

PubMed Abstract | Crossref Full Text | Google Scholar

19. Zercher D, Jussupow E, Heinzl A. When AI Joins the team: a literature review on intragroup processes and their effect on team performance in team-AI collaboration. In Proceedings of the ECIS 2023 Research Papers. Atlanta: (2023).

Google Scholar

20. Hauptman A, Schelble B, McNeese N, Madathil K. Adapt and overcome: Perceptions of adaptive autonomous agents for human-AI teaming. Comput Hum Behav. (2023) 138:107451. doi: 10.1016/j.chb.2022.107451

Crossref Full Text | Google Scholar

21. Jussupow E, Spohrer K, Heinzl A, Gawlitza J. Augmenting medical diagnosis decisions? An investigation into physicians’ decision-making process with artificial intelligence. Information Syst Res. (2021) 32:713–35. doi: 10.1287/isre.2020.0980

Crossref Full Text | Google Scholar

22. O’Neill T, McNeese N, Barron A, Schelble B. Human–autonomy teaming: a review and analysis of the empirical literature. Human Factors. (2020) 64:904–38.

Google Scholar

23. Agarwal R, Dugas M, Gao G. Augmenting physicians with artificial intelligence to transform healthcare: challenges and opportunities. J Econ Manag Strategy. (2024) 2:360–74. doi: 10.1111/jems.12555

Crossref Full Text | Google Scholar

Keywords: AI, explainability, healthcare, regulation, human-AI teaming, early adoption, XAI

Citation: Raz A, Inbar Y, Avnoon N, Bela Lifshitz-Milwidsky L, Hofman B, Honig A and Paz Z (2025) Who moved my scan? Early adopter experiences with pre- and post-market healthcare AI regulation challenges. Front. Med. 12:1651934. doi: 10.3389/fmed.2025.1651934

Received: 22 June 2025; Accepted: 11 August 2025;
Published: 26 August 2025;
Corrected: 05 September 2025.

Edited by:

Antonina Argo, University of Palermo, Italy

Reviewed by:

Dinesh Mendhe, Robert Wood Johnson University Hospital, United States
Joshua Ohde, Mayo Clinic, United States
Hassan Sami Adnan, University of Oxford, United Kingdom

Copyright © 2025 Raz, Inbar, Avnoon, Bela Lifshitz-Milwidsky, Hofman, Honig and Paz. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Aviad Raz, YXZpYWRyYXpAYmd1LmFjLmls

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.