- 1Dr. Gilles Arcand Centre for Health Equity, NOSM University, Thunder Bay, ON, Canada
- 2Faculty of Education, Lakehead University, Thunder Bay, ON, Canada
- 3Health Sciences Library, NOSM University, Thunder Bay, ON, Canada
- 4Human Sciences Division, NOSM University, Thunder Bay, ON, Canada
This rapid review explores how artificial intelligence (AI) is integrated into healthcare and examines the factors influencing trust between users and AI systems. By systematically identifying trust-related determinants, this review provides actionable insights to support effective AI adoption in clinical settings. A comprehensive search of MEDLINE (Ovid), Embase (Ovid), and CINAHL (Ebsco) using keywords related to AI, healthcare, and trust yielded 872 unique citations, of which 40 studies met the inclusion criteria after screening. Three core themes were identified. AI literacy highlights the importance of user understanding of AI inputs, processes, and outputs in fostering trust among patients and clinicians. AI psychology reflects demographic and experiential influences on trust, such as age, gender, and prior AI exposure. AI utility emphasizes perceived usefulness, system efficiency, and integration within clinical workflows. Additional considerations include anthropomorphism, privacy and security concerns, and trust-repair mechanisms following system errors, particularly in high-risk clinical contexts. Overall, this review advances the understanding of trustworthy AI in healthcare and offers guidance for future implementation strategies and policy development.
Introduction
Artificial intelligence (AI) refers to software systems capable of performing tasks that mimic human reasoning through data processing and algorithms (Kaplan et al., 2023; West, 2018). As open-source AI platforms have expanded, adoption across sectors—including healthcare—has accelerated, heightening the need to understand and establish trust in these technologies (Lyons et al., 2024). When appropriately applied, AI can enhance task accuracy, improve safety, and support decision-making efficiency (Baduge et al., 2022; Sardar et al., 2019; Soori et al., 2023; Abdar et al., 2022; Begoli et al., 2019). In healthcare, AI is increasingly used to streamline workflows and reduce human error. Examples include electronic medical record (EMR) automation to support coordinated care (Rojahn et al., 2023), clinical decision support systems (CDSS) that aid diagnosis and treatment planning (Laxar et al., 2023; Osheroff et al., 2007), and robot-assisted surgery designed to minimize procedural risks (Hamet and Tremblay, 2017; Soori et al., 2023). Despite these benefits, challenges remain, such as inaccurate chatbot responses, diagnostic errors, and risks of exacerbating health inequities (Kaplan et al., 2023; D’Elia et al., 2022). Growing concerns around reliability, bias, transparency, and ethical use make trust a critical component of successful AI integration (Kostick-Quenet et al., 2024; Asan et al., 2020).
Trust involves vulnerability and uncertainty and shapes how individuals engage with automated systems as much as with other people (Deng et al., 2024; Hoff and Bashir, 2015). Both insufficient trust—leading to system avoidance—and overtrust—leading to misuse—pose risks (Muir and Moray, 1996). In healthcare, trust underpins patient-provider relationships and clinician reliance on diagnostic and treatment supports (Birkhäuer et al., 2017; Gaube et al., 2021). Trust has long been fundamental to healthcare, shaping both patient–clinician relationships and clinical decision-making processes. Patients must trust that clinicians act in their best interests and possess the expertise to deliver safe, effective care (Birkhäuer et al., 2017). Likewise, clinicians must have confidence in the tools and systems that support their diagnostic and treatment decisions (Gaube et al., 2021). As AI becomes increasingly embedded in clinical workflows, establishing trust from both patients and clinicians is essential for adoption and meaningful use (Asan et al., 2020). Ultimately, trust determines whether AI technologies are accepted, integrated into practice, and relied upon in patient care, underscoring the need to understand the evolving dynamics of human-machine trust in healthcare (Asan et al., 2020; Gaube et al., 2021; Lee and See, 2004). Therefore, fostering trust in AI requires addressing the needs and expectations of both clinicians and patients (Asan et al., 2020).
This rapid review synthesizes current evidence on the factors shaping human-machine trust in healthcare, with the aim of informing safe, effective, and trusted AI implementation.
Methods
The search strategy for this rapid review was developed in consultation with a Health Sciences Librarian. Searches were conducted on March 22, 2024, in MEDLINE (Ovid), Embase (Ovid), and CINAHL (Ebsco). These databases were chosen for this rapid review due to their extensive coverage and relevance to medical and allied health literature, especially in relation to the application of artificial intelligence in healthcare.
To capture the scope of the published research, the final search strategy relies on Medical Subject Headings (MeSH) from MEDLINE (Ovid) and is translated to the equivalent term in the other resources. Target articles were also reviewed for relevant subject headings. Previous iterations of the search strategy were complex, and many citations were found to be too specific for this review. Because of the complex and individualized nature of trust relationships, we decided not to define trust before commencing our study (Table 1).
Table 1. A summary of the comprehensive database search strategy used, including search terms, Boolean operators, and databases (MEDLINE, Embase, and CINAHL) employed to identify studies.
Results
Search results
Figure 1 summarizes the search results and article selection process. A search of three databases identified 1,082 (EMBASE n = 802; CINAHL n = 242; MEDLINE n = 38) articles for screening. All duplicates and editorial articles were removed (n = 210), and a total of 872 articles underwent title, abstract and citation screening. This process excluded 737 articles, leaving 135 articles for full-text screening. 135 full-text articles were assessed for eligibility, with a further 95 articles being excluded because of the wrong article type (n = 21), not related to healthcare (n = 30), trust was not a central aspect of the article/the paper did not relate to trust in AI (n = 43), and a lack of AI or related terms (n = 1). Data was extracted and analyzed from the remaining 40 articles.
Study characteristics
Of the 40 included studies, 20 were quantitative, 12 were qualitative, and 8 were mixed methods (Table 2). A summary of the specific healthcare domains is included in Table 2, although many (n = 11) took a broad approach and did not identify a specific area of healthcare.
Definitions of trust
Of the 40 included articles, 15 articles explicitly defined trust, 10 of the articles defined trust in relation to technology, three of the articles took a more general, psychological approach to defining trust and two of the articles defined trust in the context of the healthcare system (Table 3). Although there were similarities among the definitions used, no two articles included in our study used the same trust definition. The differing definitions of trust are showcased in Table 4.
Results
An inductive approach to thematic analysis was employed, allowing themes to emerge directly from the data without being shaped by pre-existing theoretical frameworks. Thematic analysis of the included articles revealed 3 main themes relating to the factors affecting user trust of AI in healthcare. These themes emerged as (1) AI Literacy, (2) AI Psychology, and (3) AI Utility. Each theme was made up of different subthemes. Table 5 shows a breakdown of the number of articles placed in each theme and subtheme. Many articles discuss more than one theme or subtheme, and are included in multiple sections (Table 6).
Theme 1: AI Literacy for acceptance, understanding, and explainability
A major theme in the literature is centered around the need for user understanding of the input, processes and outputs of the AI system. 32 articles referred to “user literacy” as being a factor affecting either patient or physician trust in AI. The need for literacy was further broken down in the literature in terms of acceptance, understanding, and explainability.
Acceptance–trust enough to use
20 articles discussed the need for a level of trust that was high enough to influence the adoption and acceptance of AI systems into the healthcare system (Alhur et al., 2023; Bergquist et al., 2024; Cheng et al., 2022; Choudhury et al., 2022; Choudhury and Shamszare, 2023; Çitil and Çitil Canbay, 2022; Costantino et al., 2022; Hsieh, 2023; Kim and Kim, 2022; Kosan et al., 2022; Kostick-Quenet et al., 2024; Liu and Tao, 2022; Nash et al., 2023; Rieger et al., 2024; Shin et al., 2023; Sonmez, 2024; Stevens and Stetson, 2023; Szabo et al., 2024; Viberg Johansson et al., 2024). Many articles stressed the importance of system accuracy, reliability, and overall trustworthiness as major factors in building enough trust in a system to adopt it into practice (Bergquist et al., 2024; Fischer et al., 2023; Hsieh, 2023; Kostick-Quenet et al., 2024; Nash et al., 2023; Rieger et al., 2024; Stevens and Stetson, 2023; Shin et al., 2023). The level of risk associated with the task given to the AI system can also influence physician trust in the system. For example, in a study focusing on a clinical decision support system for survival rate in cardiology patients, Kostick-Quenet et al. (2024) state that: “some physicians noted that they demand higher levels of accuracy for estimates intended to inform choices about pursuing an intervention with life-or-death implications versus those intended to predict postoperative adverse events and manage postoperative care”. Physicians want to trust that the system will work as expected, and patients want to trust that their physician trusts the system will work as expected (Cheng et al., 2022; Choudhury et al., 2022). Two studies noted that many physicians who were initially apprehensive toward the implementation of AI into healthcare gained trust in the system after working with the system (Hsieh, 2023; Shin et al., 2023).
Understanding–knowledge of AI systems
Understanding the inner workings of AI system operation was discussed in 13 articles (Asokan et al., 2023; Bergquist et al., 2024; Choudhury and Shamszare, 2023; Hallowell et al., 2022; Hsieh, 2023; Shahzad Khan et al., 2024; Kostick-Quenet et al., 2024; Liu et al., 2022; Louca et al., 2023; Moilanen et al., 2023; Nash et al., 2023; Qassim et al., 2023; Viberg Johansson et al., 2024). Many articles iterated that in using AI systems, physicians do not require a comprehensive engineering understanding of how the system works but would rather have transparency in terms of the datasets used for building the algorithms, and on the development and testing processes (Asokan et al., 2023; Bergquist et al., 2024; Choudhury and Shamszare, 2023; Hallowell et al., 2022; Hsieh, 2023; Shahzad Khan et al., 2024; Kostick-Quenet et al., 2024; Louca et al., 2023; Moilanen et al., 2023; Qassim et al., 2023). This point is illustrated by Kostick-Quenet et al. (2024), who state that, “both patients and physicians from our study emphasized a desire to know more about the nature of data sets—rather than algorithms—used to train algorithmic models, in order to gage relevance of outputs for making inferences about target users or subjects.” According to Qassim et al., physicians view AI systems as a supplement to their clinical judgment; therefore, they do not require a system understanding (Qassim et al., 2023). Three studies found that transparency on the limitations of the system was key to developing physician trust (Moilanen et al., 2023; Liu et al., 2022; Qassim et al., 2023). Asokan et al. (2023) highlight the need for transparency in the development and testing of AI systems, when they state that “respondents consistently desired transparency on how an AI tool was developed and tested.” In contrast, two studies did note that providing information on the algorithmic processes and inner workings of the AI systems did improve physician trust and intent to use (Choudhury and Shamszare, 2023; Hallowell et al., 2022). For example, Hallowell et al. (2022) state that: “trust must be built upon transparency about, and awareness of, how algorithms work, rather than having ‘blind faith’ in algorithmic output”.
Explainability–knowledge of AI outputs
Ten studies discuss how user trust in the AI system is built when the system can explain its final decision (Bergquist et al., 2024; Fischer et al., 2023; Goel et al., 2022; Helenason et al., 2024; Kostick-Quenet et al., 2024; Lancaster Farrell, 2022; Liu et al., 2022; Qassim et al., 2023; Rainey et al., 2022; Van Bulck and Moons, 2024). Many of these studies noted that physician trust in AI systems, most specifically in clinical decision support systems (CDSS), increases with system explainability (Goel et al., 2022; Lancaster Farrell, 2022; Liu et al., 2022; Qassim et al., 2023; Rainey et al., 2022). When analyzing the implementation of a CDSS for psychiatric use, Qassim et al. (2023) found that “doctors felt comfortable using the app in practice because they were, as per one physician, “never surprised” by the CDSS’s recommendations and the “AI explained reasoning behind its choices.” Two studies found that explanations become even more imperative when an AI response may differ from a clinician’s point of view and can aid in exploring different clinical treatments (Bergquist et al., 2024; Kostick-Quenet et al., 2024). The bias many radiologists tend to face is explained by Bergquist et al. (2024), when they state that “more recent cases tended to influence them [radiologists] the most, whereas the AI considered all cases it had been trained on and thus provided them with a more extensive frame of reference.”
Theme: 2: AI psychology and human development
The second theme, identified in 27 articles, explores the psychological factors underlying the development of trust in AI systems. This theme focused more on the population demographic and profile of individuals trusting the AI system. This theme was further broken down in the literature in terms of human development, human motivation, and human relations.
Human development—who is more likely to trust
In 16 articles, the literature pointed to individual traits or lived experiences that may make an individual more or less likely to trust an AI system (Asokan et al., 2023; Choudhury and Asan, 2023; Hsieh, 2023; Huo et al., 2022; Kim and Kim, 2022; Kosan et al., 2022; Liu and Tao, 2022; Moilanen et al., 2023; Nash et al., 2023; Schulz et al., 2023; Sonmez, 2024; Stevens and Stetson, 2023; Szabo et al., 2024; Yakar et al., 2022; Yi-No Kang et al., 2023). Three studies stated that experience using AI systems is a precursor to increased trust in the system (Asokan et al., 2023; Liu and Tao, 2022; Schulz et al., 2023). Schulz et al. (2023) highlight the need for increased AI exposure when they state: “our findings show that trust and beliefs need to go hand in hand with exposure to AI. Therefore, introductions to AI need to be more effectively done through trust enhancers, such as involving trusted professional sources.” Two studies found that positive experiences with AI systems lead to increased trust (Moilanen et al., 2023; Szabo et al., 2024), whereas negative experiences may lead to decreased trust in the systems (Moilanen et al., 2023; Nash et al., 2023). Despite most of the literature agreeing that experience with AI technologies increased physician trust and willingness to implement new technologies, one of the included studies did not find any significant correlation between trust and previous experience (Choudhury and Asan, 2023). This is highlighted in the study by Choudhury et al. (2022), on the implementation of blood utilization calculators (BUC) into clinical practice when they state that “according to our study, clinicians’ experience of using the AI system (BUC) or their familiarity with AI technology, in general, had no significant impact on their trust in–or intent to use–BUC.” Four studies noted that men tend to show increased levels of trust toward AI systems in the healthcare field rather than women (Çitil and Çitil Canbay, 2022; Kim and Kim, 2022; Liu and Tao, 2022; Sonmez, 2024). Conversely, Hsieh (2023) found no significant effect of gender on AI usage. Age was discussed as a factor affecting trust in AI systems in three studies (Kosan et al., 2022; Liu and Tao, 2022; Stevens and Stetson, 2023). Stevens and Stetson (2023) note the effect of age when they state: “as the clinician’s age group increased, the trustworthiness of AI became increasingly important to their willingness to use it.” Another study emphasized that elderly patients tend to be more skeptical of technology being introduced into their healthcare, and that, in general, younger individuals are more trusting of technology (Kosan et al., 2022). Two studies noted that higher levels of education often correlated with higher levels of technology trust (Kosan et al., 2022; Yi-No Kang et al., 2023). Notably, higher education can increase health literacy among patients, and an increased health literacy has a positive effect on trust of AI systems in the healthcare field (Yi-No Kang et al., 2023). One study found that individuals who have a higher cumulative affinity for technology tend to have more trust in AI being integrated into the healthcare field (Rodler et al., 2024).
Human motivation–experience of trust
When analyzing the factors that influence trust in AI, one article in our review highlighted certain psychological aspects of trust that, while not specific to AI, are fundamental to the development of trust in any context (Hallowell et al., 2022). Hallowell et al. (2022) explain that if the “cost” of not trusting something is greater than the “cost” of trusting, then individuals may act as though they have trust in a particular thing, but that “this does not indicate the existence of a fully-fledged trust relationship.” They also stress the idea that “developing trust in any technology is reliant on one’s experience of using it; trust is learnt” (Hallowell et al., 2022).
Human relations–human validation of AI
Sixteen articles included in our review pointed to the idea that trust in the decisions made by AI systems is built when a human or professional in the field validates the output given by an AI system (Asokan et al., 2023; Cheng et al., 2022; Fischer et al., 2023; Helenason et al., 2024; Hallowell et al., 2022; Kosan et al., 2022; Kostick-Quenet et al., 2024; Nash et al., 2023; Qassim et al., 2023; Rodler et al., 2024; Rojahn et al., 2023; Schulz et al., 2023; Shin et al., 2023; Sonmez, 2024; Szabo et al., 2024; Van Bulck and Moons, 2024; Viberg Johansson et al., 2024). From the point of view of the patient, one study found that generally, “the public significantly prefers a human physician over an AI system” (Rojahn et al., 2023). However, three studies found that when AI is endorsed by their clinical team, patients are more likely to trust the AI system (Hallowell et al., 2022; Kostick-Quenet et al., 2024; Qassim et al., 2023). Additionally, it was found that patients have more trust in AI systems controlled by physicians than those not controlled by a physician (Hallowell et al., 2022; Qassim et al., 2023; Rodler et al., 2024). This is highlighted by Hallowell et al. (2022) who found: “as far as our interviewees were concerned if a trusted person—your doctor—uses AI, then you are more likely to trust the algorithmic output.” Eight studies found that physicians are more likely to trust the system when there is proof of validation by professionals within their field (Asokan et al., 2023; Fischer et al., 2023; Hallowell et al., 2022; Helenason et al., 2024; Kostick-Quenet et al., 2024; Nash et al., 2023; Qassim et al., 2023; Schulz et al., 2023). Kostick-Quenet et al. (2024) highlight the need for human validation by saying: “both physicians and patients explained that they would be more likely to trust algorithmic estimates that were endorsed by members of the medical community who are themselves perceived as reputable and trustworthy.” Five studies report physicians have improved trust in the AI system when it matches their clinical judgment (Asokan et al., 2023; Hallowell et al., 2022; Helenason et al., 2024; Qassim et al., 2023; Viberg Johansson et al., 2024), as described in the study by Qassim et al. (2023) when they state that “5/7 doctors reported they trusted the tool because, as one doctor put it, it “was aligned with doctor’s clinical opinions and was good reinforcement from an exogenous source.”
Human attitudes–distrust of AI
An aspect of AI Psychology captured in three of the included articles is the idea of distrust, which captures those who are apprehensive toward the implementation of AI systems in the healthcare field (Hsieh, 2023; Van Bulck and Moons, 2024; Yi-No Kang et al., 2023). One study states that mistrust of the systems is a significant barrier to physician implementation, and states a lack of transparency and accuracy as major factors influencing this mistrust (Hsieh, 2023). One study reports that when asked about patients using ChatGPT to seek medical advice, many experts were worried about the trustworthiness of the responses given, noting that “certain information was missing, too vague, a bit misleading, and not written in a patient-centered way” (Van Bulck and Moons, 2024). Another study found that “individuals with higher levels of digital literacy have poor attitudes toward AI-assisted medical consultations because of their higher perceived distrust of AI” (Yi-No Kang et al., 2023).
Human growth–trust repair
Five articles in our review explore the idea of trust in an AI system, specifically after the AI system has made an error (Huo et al., 2022; Nash et al., 2023; Rieger et al., 2024; Rojahn et al., 2023; Viberg Johansson et al., 2024). One article expressed that in general, patients have much greater trust in human physicians than in AI systems, even when made aware of the greater likelihood of a human making a biased error (Rojahn et al., 2023). Two studies found that patients feel as though a human can learn from past mistakes, but that AI systems cannot grow in the same capacity, making trust repair more difficult in human-machine relationships (Rojahn et al., 2023; Viberg Johansson et al., 2024). This idea is contrasted by research conducted by Rieger et al. (2024), which found no differences in forgiveness and trust restoration after failure between an AI decision support agent and a human. From the point of view of the physician, one study reports that previous negative experiences with technology implementation can deter physicians from trusting new AI technologies (Nash et al., 2023).
Theme: 3: AI utility
The final theme identified in this review was present in 26 articles and focused on the utility of the AI system and how this impacts user trust. We have further broken this main theme into the categories of anthropomorphism, privacy and protection, and perceived value.
Anthropomorphism
The literature defines anthropomorphism as “the pervasive human tendency to attribute human characteristics to non-human entities” (Sonmez, 2024). There were four articles that discussed the idea of anthropomorphic AI systems, and found the literature is in consensus that patient trust in AI systems, such as surgical robots, chatbots or robot doctors, is improved when these machines exemplify human-like traits (Sonmez, 2024; Kim and Kim, 2022; Liu and Tao, 2022; Moilanen et al., 2023). Moilanen et al. (2023) highlight this in their study regarding chatbot use in mental healthcare when they state that “chatbot behavior and human likeness are essential factors informing trust in chatbots.” Two studies note that increased trust in anthropomorphic AI systems stems from the idea of a social presence (Kim and Kim, 2022; Liu and Tao, 2022). Liu and Tao (2022) describe this in their study when they state that:
“We found that while anthropomorphism failed to produce direct effect on behavioral intention, it indeed exerted an indirect effect on behavioral intention through the mediating role of trust. It is likely that, when people are interacting with anthropomorphic smart healthcare services, the feeling of trust would be emerged due to the perception of a social presence”.
They additionally noted that anthropomorphism is increasingly important for developing trust in females and younger adults (Liu and Tao, 2022).
Privacy and protection
Nine articles in our study discuss the interrelationship between trust and the protection of privacy and data (Alhur et al., 2023; Asokan et al., 2023; Bergquist et al., 2024; Costantino et al., 2022; Shahzad Khan et al., 2024; Liu and Tao, 2022; Moilanen et al., 2023; Nash et al., 2023; Viberg Johansson et al., 2024). Four studies emphasize that inadequate protection of patient privacy is a major patient concern regarding the implementation of AI into the healthcare system and can have significant effects on patient trust in these systems being incorporated into their care (Asokan et al., 2023; Moilanen et al., 2023; Nash et al., 2023; Yi-No Kang et al., 2023). These privacy concerns are highlighted in the study conducted by Liu and Tao (2022), looking at public acceptance of smart healthcare services, when they state that “loss of privacy did not directly influence behavioral intention, but it was found to have a negative influence on trust, indicating that consumers concerned with privacy are less likely to trust such services.” One study found these concerns become increasingly relevant in sensitive areas of healthcare where data privacy is much more important, such as mental health (Moilanen et al., 2023). One study found that trust increases when patient privacy concerns are addressed and robust security measures are put in place (Alhur et al., 2023). Furthermore, two studies found that patient trust in AI systems is built when privacy and data security are maintained (Costantino et al., 2022; Shahzad Khan et al., 2024). Bergquist et al. (2024) found that from the physician’s point of view, having control over the data is an important factor in determining trust.
Perceived value
19 articles included in our study found that the perceived value of the AI tool was a factor influencing trust (Asokan et al., 2023; Bergquist et al., 2024; Cheng et al., 2022; Choudhury et al., 2022; Çitil and Çitil Canbay, 2022; Costantino et al., 2022; Helenason et al., 2024; Hsieh, 2023; Liu et al., 2022; Liu and Tao, 2022; Nash et al., 2023; Qassim et al., 2023; Rojahn et al., 2023; Schulz et al., 2023; Shamszare and Choudhury, 2023; Shin et al., 2023; Stevens and Stetson, 2023; Szabo et al., 2024; Yakar et al., 2022; Yi-No Kang et al., 2023). Without added value to the patient-physician encounter or to clinical workflow, the adoption of AI systems would not be effective, as identified by Liu et al., when they state, “if medical AI is of little value to physicians, physicians will resist accepting AI” (Liu et al., 2022). Seven studies report that among both patients and physicians, trust and willingness to adopt a particular AI system increased when the user perceived the system to be useful (Alhur et al., 2023; Cheng et al., 2022; Liu and Tao, 2022; Qassim et al., 2023; Schulz et al., 2023; Stevens and Stetson, 2023; Yi-No Kang et al., 2023). Additionally, nine studies found that for both patients and physicians, ease of use of the system had major implications on trust and willingness to use. Three studies report that patients felt that the AI system needed to be convenient for them to use and add efficiency to their encounters (Alhur et al., 2023; Yakar et al., 2022; Yi-No Kang et al., 2023). Regarding physicians, one study that focused on otolaryngologists’ views on AI noted that “physicians valued efficiency more than accuracy or understanding AI algorithms and design” (Asokan et al., 2023). Six studies noted that to trust a system enough to implement it into clinical practice, it must be capable of speeding up complex tasks, reducing reading times or reducing physician workload (Choudhury et al., 2022; Çitil and Çitil Canbay, 2022; Hsieh, 2023; Qassim et al., 2023; Shamszare and Choudhury, 2023; Shin et al., 2023). The relationship between physician workload and trust in AI is highlighted by Shamszare and Choudhury (2023) when they state that “clinicians who view AI as a workload reducer are more inclined to trust it and are more likely to use it in clinical decision making.” To add to efficiency and ease of use, two studies found that physicians require that AI systems be compatible with other systems and practices within their organization (Bergquist et al., 2024; Viberg Johansson et al., 2024). This idea is highlighted by the study investigating the implementation of AI tools in radiology by Bergquist et al. (2024), as their results state that:
“Radiologists’ trust in AI depends on the experience that AI is compatible with other systems and practices in the organization, increasing their capacity and providing control. Trust in AI emerges when a variegated range of data formats are integrated into existing modalities so that experts across organizational or functional boundaries can share and use data to collaborate efficiently and safely”
Finally, three studies found that physicians must believe that utilizing an AI system will lower their risk of missed diagnoses and will be an added aid to confirm their diagnoses (Hsieh, 2023; Qassim et al., 2023; Shin et al., 2023).
Discussion
AI holds significant potential to enhance decision-making, increase efficiency, and transform existing healthcare practices. Research indicates that the impact of automated systems will depend on whether and how human factors are incorporated into their design (Asan et al., 2020). Thus, as AI continues to be implemented throughout the healthcare system, considering the factors that impact patient and physician trust in AI technologies is of vital importance (Baduge et al., 2022; Glikson and Woolley, 2020). Trust is a complex construct involving many factors, and it plays a critical role in both human relationships and human-machine interactions (Asan et al., 2020). Our review identified that only 35% of the included studies provided an explicit definition of trust. Furthermore, none of the articles utilized the same definition, underscoring the complexity of evaluating this variable using a standardized approach. Definitions of trust were subsequently categorized according to the context under evaluation, including trust in technology, psychological perspectives on trust, and trust within healthcare settings. Across various disciplines, workers’ trust in AI technology is an important aspect of the successful integration of AI into a workplace (Glikson and Woolley, 2020). This rapid review addresses the factors that contribute to developing a trusting relationship in AI technologies within the healthcare industry.
Several factors were identified as essential prerequisites for fostering trust in AI technologies, highlighting the minimum requirements that AI systems must meet to gain user confidence. These factors include AI technologies addressing privacy concerns, maintaining accuracy, providing credible information, high levels of system performance, addressing biases, and overall reliability (Bergquist et al., 2024; Fischer et al., 2023; Kostick-Quenet et al., 2024; Stevens and Stetson, 2023).
When users understood the capabilities and limitations of the AI system they were using, a trusting relationship could be developed (Louca et al., 2023; Moilanen et al., 2023; Viberg Johansson et al., 2024). Both patients and physicians expressed a desire to understand how the overall system was developed and how datasets are utilized to build and refine the algorithms used in healthcare settings (Asokan et al., 2023; Bergquist et al., 2024; Kostick-Quenet et al., 2024; Liu et al., 2022). Gender was linked to attitudes toward AI technology, with females having less trust and more negative attitudes toward technology compared to males in all but one study (Hsieh, 2023) where gender was insignificant in predicting physician trust (Çitil and Çitil Canbay, 2022; Hsieh, 2023; Kim and Kim, 2022; Liu and Tao, 2022; Sonmez, 2024). Sonmez (2024) specifically found that women had a higher risk aversion compared to men, with less positive attitudes, regardless of the human-likeness of AI. Additionally, experience using technology was a contributing factor leading to trust in AI technology. Those with more experience with technology were more likely to trust AI technology and their decisions (Asokan et al., 2023; Moilanen et al., 2023; Schulz et al., 2023). In contrast, previous negative experiences caused distrust and deterred the use of the systems (Nash et al., 2023). Although some studies report the impact of age on trust in technology, one study in this review found that neither prior experience nor age was a factor impacting trust in AI systems or intentions to utilize AI systems (Choudhury and Asan, 2023). This contradicts findings that demonstrated a significant impact of age on trust in AI technology, with older individuals generally exhibiting lower levels of trust in AI compared to younger individuals (Nash et al., 2023; Oksanen et al., 2020). Overall, the more familiar a user is with AI technology, the more likely they are to trust the system. Many studies have found explainable artificial intelligence to be an essential factor impacting user trust (Bergquist et al., 2024; Fischer et al., 2023; Goel et al., 2022; Helenason et al., 2024; Qassim et al., 2023). As outlined in Theme 1, several included studies reported higher adoption and comfort when AI provided interpretable outputs rather than “black box” predictions. This underscores explainability as a critical design feature for trustworthy AI in healthcare (Bergquist et al., 2024; Fischer et al., 2023; Goel et al., 2022; Helenason et al., 2024; Qassim et al., 2023). However, two studies found that explainability was not a factor that predicted whether physicians would or would not utilize AI technology and did not act as a factor to improve trust in healthcare contexts (Lancaster Farrell, 2022; Liu et al., 2022). Understanding the complexity of trust and how trust is built was another important factor influencing trust relationships. For example, trust relationships were impacted when users of AI technologies recognized that trust is a learned behavior and evaluated the overall impact of whether they were to trust or not trust a system (Hallowell et al., 2022).
Our review found that trust is built when the user believes there will be positive benefits from using the AI system. Positive perceptions regarding the usefulness of AI technologies resulted in greater user trust (Alhur et al., 2023; Stevens and Stetson, 2023). Specifically, physicians were likely to use and trust AI technologies if they found them to be efficient, easy to use, and useful in their practice (Asokan et al., 2023; Cheng et al., 2022; Hsieh, 2023; Liu and Tao, 2022; Nash et al., 2023; Qassim et al., 2023; Schulz et al., 2023; Shamszare and Choudhury, 2023; Yakar et al., 2022; Yi-No Kang et al., 2023). As the perceived value of an AI technology increases, physicians are more likely to rely on the system, thereby enhancing their trust in it (Bergquist et al., 2024; Goel et al., 2022; Liu et al., 2022). Furthermore, trust in the AI system was enhanced when the system could be seamlessly integrated into the physicians’ practice and demonstrated compatibility with existing systems (Bergquist et al., 2024).
The human-likeness of AI systems plays a key role in shaping patient trust, fostering a greater openness to incorporating AI technology into their care networks (Choudhury and Shamszare, 2023; Kim and Kim, 2022; Liu and Tao, 2022; Sonmez, 2024). The greater the degree of human-like attributes exhibited by AI technologies, such as surgical robots, the higher the likelihood that individuals will extend their trust in these systems, owing to an enhanced sense of social presence (Kim and Kim, 2022; Liu and Tao, 2022). Trust repair after system failure remains challenging for AI, though findings were mixed, as outlined in Theme 2 (Asokan et al., 2023; Moilanen et al., 2023). This reflects the psychological aspect of trust: human errors are often seen as part of a learning process, while AI mistakes are viewed as indicative of systemic flaws (Moilanen et al., 2023). Consequently, this underscores the necessity for effective trust repair strategies to rebuild confidence in AI systems within healthcare settings (Moilanen et al., 2023). This review found that different psychological and external influences can shape the feelings of the patient toward AI. Therefore, understanding these factors can help build patient acceptance of AI systems and build patient trust in AI even when it makes mistakes (Huo et al., 2022).
As detailed in Theme 2, trust in AI often follows a cascade: patients are more likely to trust systems endorsed by their physicians, and physicians are more comfortable when the technology has been validated by peers, experts, or reputable organizations. Rather than revisiting individual study findings, this reinforces the importance of human validation as a bridge between developers, clinicians, and patients (Asokan et al., 2023; Cheng et al., 2022; Fischer et al., 2023; Helenason et al., 2024; Hallowell et al., 2022; Kostick-Quenet et al., 2024; Qassim et al., 2023; Rodler et al., 2024; Sonmez, 2024). Trust was further enhanced if professionals outside the healthcare field, such as software engineers, program designers, and technology leaders, confirmed the trustworthiness of AI systems (Szabo et al., 2024). Physician trust would fluctuate, though, based on the result of the AI technology. For example, if the system results in decisions that the physician disagrees with, they may grow to distrust it (Hallowell et al., 2022). However, trust was enhanced if the system tended to come to the same decision as the physician (Asokan et al., 2023; Hallowell et al., 2022; Helenason et al., 2024; Qassim et al., 2023; Viberg Johansson et al., 2024). Additionally, physicians who recognize the potential for bias in the development of AI technologies tend to be more cautious and less likely to trust these systems sufficiently to integrate them into their clinical practice (Van Bulck and Moons, 2024). To trust a system, physicians want to be confident that the system is trained on a robust dataset and that these systems are constantly updating. This can lead to a disconnect between patients’ privacy of their data and accurately training the AI systems to be as accurate as possible (Bergquist et al., 2024). In a study conducted by Viberg Johansson et al. (2024) looking at women’s perceptions and attitudes toward the implementation of AI in mammography, patients were willing to share their data for life-saving measures or research purposes but were reluctant to share with private entities, as they considered it “an intrusion into their private lives.” Patients may not understand that data sharing with private entities is essential for technology advancements (Viberg Johansson et al., 2024).
The Organization for Economic Co-Operation and Development (OECD) states that trustworthy AI will be transparent and explainable, where the capabilities and limitations of the system are readily available and easily explained to users (OECD Legal Instruments, 2019). A major area of concern with AI in sensitive fields such as healthcare is the “black box” nature of the system. Output and decisions are often given without explanation behind them, making implementation and trust difficult (Liu et al., 2022). The introduction of explainable AI (XAI) refers to the actions and measures taken to ensure transparency in the AI system that is both explainable and interpretable (Adadi and Berrada, 2018; Liu et al., 2022). Implementing XAI could be a major step toward generating trust in AI systems and furthering their implementation. Trustworthy AI should have security and safety measures in place to avoid harm in the case of misuse or adverse conditions (OECD Legal Instruments, 2019). Finally, trustworthy AI should be accountable and ensure traceability of datasets, processes and decisions made by the system (OECD Legal Instruments, 2019). Because AI can be implemented in vastly different ways in different fields, major factors impacting trust in one field may vary greatly from those in another field. Understanding the key aspects that play a role in determining user trust in AI systems is important for the successful implementation of AI (Abeywickrama et al., 2023; Louca et al., 2023). The ability of the AI system to explain its decision-making process can reveal bias that a diagnosing practitioner may have been unaware of (Bergquist et al., 2024). Therefore, collaboration between technology programmers and healthcare professionals may ensure that the outputs of AI technology are valid and aligned with clinicians, resulting in stronger trust relationships (Nash et al., 2023). Furthermore, when patients believed the physician was making the final decision, not the AI system, they were more open to trusting AI technologies to assist with healthcare decisions (Rodler et al., 2024; Van Bulck and Moons, 2024; Viberg Johansson et al., 2024). However, studies found that patients preferred physicians over AI systems, even if trust was formed (Rojahn et al., 2023; Viberg Johansson et al., 2024). Thus, a cascading relationship exists in user trust for AI systems. For a patient to trust the system, the physician must first trust the system, and the physician’s trust depends on their confidence in the program developer who created it (Hallowell et al., 2022). From the patient’s perspective, a common barrier to trust in AI technologies is the concern over privacy invasion and data collection practices (Liu and Tao, 2022; Moilanen et al., 2023; Nash et al., 2023; Viberg Johansson et al., 2024). In contexts where users exhibit a moderate level of trust in AI systems, the implementation of direct safeguards for data protection and privacy significantly enhanced trust in these systems (Alhur et al., 2023). Healthcare organizations need to adopt standardized procedures for addressing data collection and management issues associated with AI technologies to foster trust among patients (Bergquist et al., 2024; Costantino et al., 2022). This includes physicians ensuring open communication and transparency with patients surrounding privacy and data security (Shahzad Khan et al., 2024).
To trust a system, physicians want to be confident that the system is trained on a robust dataset and that these systems are constantly updating. This can lead to a disconnect between patients privacy of their data and accurately training the AI systems to be as accurate as possible (Bergquist et al., 2024). Physicians may be more willing to share data for research purposes; however, patients may not understand that data sharing with private entities is essential for technology advancements (Viberg Johansson et al., 2024). Overall, the successful integration of AI technologies into healthcare will depend on fostering trust among both physicians and patients. The multifaceted nature of trust is evident, as well as the importance of transparency, system reliability, and privacy protections. As the development and implementation of AI evolves, understanding and addressing the factors that influence trust will be critical to ensure effective adoption and long-term success in improving healthcare outcomes.
Limitations
This study has several limitations. Firstly, as this study is a rapid review, the search was less comprehensive than a scoping or systematic review. The scope of this study was limited to healthcare contexts, so we caution against any interpretations of findings beyond the healthcare setting. Future studies should take a broader approach to consider factors of trust in other contexts. Our search was limited to academic publications, so there could be valuable information in the gray literature that was excluded from this review.
Future research
This review does not address methods for building trust between humans and artificial intelligence in healthcare settings; thus, future research should address this. Further studies should examine the factors of trust in AI outside of healthcare settings. Additionally, longitudinal studies should examine how trust evolves and explore effective recovery strategies related to AI errors. These studies could further investigate how initial impressions of AI evolve with sustained use. This may include investigating effective trust repair mechanisms, including system transparency, error communication strategies, and human oversight models, to restore confidence after AI failure. Future studies may adopt large-language models (LLMs) in order to assist with literature analysis. For the purposes of this review LLMs were not utilized as our goal was to ensure transparency, reproducibility, and methodological rigor by relying on traditional, human-driven review methods such as systematic searching, critical appraisal, and manual synthesis of findings. While LLMs have shown promise for assisting in scientific research and paper drafting, their capabilities, limitations, and best practices for reliable use in scholarly reviews were not yet well defined at the time of this work, and integrating them was considered outside the scope of our study.
An effort should be made to develop standardized metrics and validated scales to measure trust in AI systems consistently across healthcare contexts. The current literature varies and often involves non-comparable definitions of trust, making cross-study synthesis challenging. It also suggested that a co-design and participatory development framework is utilized, engaging clinicians, patients, and technology developers to ensure that AI tools align with user needs, workflows, and ethical expectations. Future studies should also investigate how trust is shaped across diverse cultural contexts and low-resource settings, as well as how trust develops within interdisciplinary teams that include not only clinicians but also information technology developers, administrators, and other stakeholders involved in AI implementation. Although this rapid review did not aim to evaluate or propose specific methodological frameworks, future work would benefit from the development and application of structured approaches—such as standardized trust metrics, longitudinal study designs to track trust over time, and co-design models that engage clinicians, patients, and developers—to guide the creation and assessment of trustworthy AI systems.
Conclusion
Through a rapid review of the literature, our findings suggest that trust is imperative for the successful implementation of AI in healthcare settings and that there are numerous factors that contribute to patient and physician trust in AI systems. By examining broad categories that shape trust, including AI literacy, psychology and utility, this review underscores the complexity of building trust across diverse user groups. Trust in AI is not a static construct that is well defined but involves a dynamic interplay of system transparency, perceived value, and user experience. Furthermore, concepts such as anthropomorphism, privacy, and trust repair strategies must be considered when evaluating trust in AI systems. In order to use AI systems to their full potential in healthcare settings, developers, clinicians, and policymakers must collaborate to address these multifaceted aspects of trust. This involves designing AI systems that are transparent, explainable, and well-aligned with ethical standards to maintain robust privacy safeguards. Future research should continue exploring the nuanced relationships among trust factors and expand beyond healthcare to inform AI applications in other critical domains. Trustworthy AI systems not only have the ability to enhance clinical decision-making and efficiency but could pave the way for a more equitable and inclusive healthcare landscape.
Author contributions
MM: Methodology, Formal analysis, Data curation, Writing – original draft, Conceptualization, Writing – review & editing. KT: Data curation, Writing – review & editing, Writing – original draft, Formal analysis. GS: Writing – review & editing, Formal analysis, Methodology, Conceptualization. GA: Writing – review & editing. JD: Methodology, Writing – review & editing, Writing – original draft. EC: Writing – review & editing, Methodology, Conceptualization, Supervision.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. The study was supported by an AMS Grant as well as NOHFC.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Abdar, M., Khosravi, A., Islam, S. M. S., Acharya, U. R., and Vasilakos, A. V. (2022). The need for quantification of uncertainty in artificial intelligence for clinical data analysis: increasing the level of trust in the decision-making process. IEEE Syst. Man Cybern. Mag. 8, 28–40. doi: 10.1109/MSMC.2022.3150144
Abeywickrama, D. B., Bennaceur, A., Chance, G., Demiris, Y., Kordoni, A., Levine, M., et al. (2023). On specifying for trustworthiness. Commun. ACM 67, 98–109. doi: 10.1145/3624699
Adadi, A., and Berrada, M. (2018). Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138–52160. doi: 10.1109/ACCESS.2018.2870052
Alhur, A. A., Aldhafeeri, M. D., Alghamdi, S. E., Bazuhair, W. S., Al-Thabet, A. A., Alharthi, K. M., et al. (2023). Telemental health and artificial intelligence: knowledge and attitudes of Saudi Arabian individuals towards AI-integrated telemental health. J. Popul. Ther. Clin. Pharmacol. 30, 1993–2009. doi: 10.53555/jptcp.v30i17.2711
Asan, O., Bayrak, A. E., and Choudhury, A. (2020). Artificial intelligence and human Trust in Healthcare: focus on clinicians. J. Med. Internet Res. 22:e15154. doi: 10.2196/15154
Asokan, A., Massey, C. J., Tietbohl, C., Kroenke, K., Morris, M., and Ramakrishnan, V. R. (2023). Physician views of artificial intelligence in otolaryngology and rhinology: a mixed methods study. Laryngoscope Investig. Otolaryngol. 8, 1468–1475. doi: 10.1002/lio2.1177
Baduge, S. K., Thilakarathna, S., Perera, J. S., Arashpour, M., Sharafi, P., Teodosio, B., et al. (2022). Artificial intelligence and smart vision for building and construction 4.0: machine and deep learning methods and applications. Autom. Constr. 141:104440. doi: 10.1016/j.autcon.2022.104440
Begoli, E., Bhattacharya, T., and Kusnezov, D. (2019). The need for uncertainty quantification in machine-assisted medical decision making. Nat. Mach. Intell. 1, 20–23. doi: 10.1038/s42256-018-0004-1
Bergquist, M., Rolandsson, B., Gryska, E., Laesser, M., Hoefling, N., Heckemann, R., et al. (2024). Trust and stakeholder perspectives on the implementation of AI tools in clinical radiology. Eur. Radiol. 34, 338–347. doi: 10.1007/s00330-023-09967-5
Birkhäuer, J., Gaab, J., Kossowsky, J., Hasler, S., Krummenacher, P., Werner, C., et al. (2017). Trust in the health care professional and health outcome: a meta-analysis. PLoS One 12:e0170988. doi: 10.1371/journal.pone.0170988
Cheng, M., Li, X., and Xu, J. (2022). Promoting healthcare workers’ adoption intention of artificial-intelligence-assisted diagnosis and treatment: the chain mediation of social influence and human–computer trust. Int. J. Environ. Res. Public Health 19:13311. doi: 10.3390/ijerph192013311
Choudhury, A., and Asan, O. (2023). Impact of cognitive workload and situation awareness on clinicians’ willingness to use an artificial intelligence system in clinical practice. IISE Trans. Healthc. Syst. Eng. 13, 89–100. doi: 10.1080/24725579.2022.2127035
Choudhury, A., Asan, O., and Medow, J. E. (2022). Effect of risk, expectancy, and trust on clinicians’ intent to use an artificial intelligence system - blood utilization calculator. Appl. Ergon. 101:103708. doi: 10.1016/j.apergo.2022.103708
Choudhury, A., and Shamszare, H. (2023). Investigating the impact of user trust on the adoption and use of ChatGPT: survey analysis. J. Med. Internet Res. 25:e47184. doi: 10.2196/47184
Çitil, E. T., and Çitil Canbay, F. (2022). Artificial intelligence and the future of midwifery: what do midwives think about artificial intelligence? A qualitative study. Health Care Women Int. 43, 1510–1527. doi: 10.1080/07399332.2022.2055760
Costantino, A., Caprioli, F., Elli, L., Roncoroni, L., Stocco, D., Doneda, L., et al. (2022). Determinants of patient trust in gastroenterology televisits: results of machine learning analysis. Inform. Med. Unlocked 29:100867. doi: 10.1016/j.imu.2022.100867
D’Elia, A., Gabbay, M., Rodgers, S., Kierans, C., Jones, E., Durrani, I., et al. (2022). Artificial intelligence and health inequities in primary care: a systematic scoping review and framework. Fam. Med. Community Health 10:e001670. doi: 10.1136/fmch-2022-001670
Deng, M., Chen, J., Wu, Y., Ma, S., Li, H., Yang, Z., et al. (2024). Using voice recognition to measure trust during interactions with automated vehicles. Appl. Ergon. 116:104184. doi: 10.1016/j.apergo.2023.104184
Fischer, A., Rietveld, A., Teunissen, P., Hoogendoorn, M., and Bakker, P. (2023). What is the future of artificial intelligence in obstetrics? A qualitative study among healthcare professionals. BMJ Open 13:e076017. doi: 10.1136/bmjopen-2023-076017
Gaube, S., Suresh, H., Raue, M., Merritt, A., Berkowitz, S. J., Lermer, E., et al. (2021). Do as AI say: susceptibility in deployment of clinical decision-aids. NPJ Digit. Med. 4:31. doi: 10.1038/s41746-021-00385-9
Glikson, E., and Woolley, A. W. (2020). Human trust in artificial intelligence: review of empirical research. Acad. Manage. Ann. 14, 627–660. doi: 10.5465/annals.2018.0057
Goel, K., Sindhgatta, R., Kalra, S., Goel, R., and Mutreja, P. (2022). The effect of machine learning explanations on user trust for automated diagnosis of COVID-19. Comput. Biol. Med. 146:105587. doi: 10.1016/j.compbiomed.2022.105587
Hallowell, N., Badger, S., Sauerbrei, A., Nellåker, C., and Kerasidou, A. (2022). “I don’t think people are ready to trust these algorithms at face value”: trust and the use of machine learning algorithms in the diagnosis of rare disease. BMC Med. Ethics 23:112. doi: 10.1186/s12910-022-00842-4
Hamet, P., and Tremblay, J. (2017). Artificial intelligence in medicine. Metabolism 69, S36–S40. doi: 10.1016/j.metabol.2017.01.011
Helenason, J., Ekström, C., Falk, M., and Papachristou, P. (2024). Exploring the feasibility of an artificial intelligence based clinical decision support system for cutaneous melanoma detection in primary care–a mixed method study. Scand. J. Prim. Health Care 42, 51–60. doi: 10.1080/02813432.2023.2283190
Hoff, K. A., and Bashir, M. (2015). Trust in automation: integrating empirical evidence on factors that influence trust. Hum. Factors 57, 407–434. doi: 10.1177/0018720814547570
Hsieh, P.-J. (2023). Determinants of physicians’ intention to use AI-assisted diagnosis: an integrated readiness perspective. Comput. Hum. Behav. 147:107868. doi: 10.1016/j.chb.2023.107868
Huo, W., Zheng, G., Yan, J., Sun, L., and Han, L. (2022). Interacting with medical artificial intelligence: integrating self-responsibility attribution, human–computer trust, and personality. Comput. Human Behav. 132:107253. doi: 10.1016/j.chb.2022.107253
Kaplan, A. D., Kessler, T. T., Brill, J. C., and Hancock, P. A. (2023). Trust in artificial intelligence: Meta-analytic findings. Hum. Factors 65, 337–359. doi: 10.1177/00187208211013988
Kim, D. K. D., and Kim, S. (2022). What if you have a humanoid AI robot doctor?: an investigation of public trust in South Korea. J. Commun. Healthc. 15, 276–285. doi: 10.1080/17538068.2021.1994825
Kosan, E., Krois, J., Wingenfeld, K., Deuter, C. E., Gaudin, R., and Schwendicke, F. (2022). Patients’ perspectives on artificial intelligence in dentistry: a controlled study. J. Clin. Med. 11:2143. doi: 10.3390/jcm11082143
Kostick-Quenet, K., Lang, B. H., Smith, J., Hurley, M., and Blumenthal-Barby, J. (2024). Trust criteria for artificial intelligence in health: normative and epistemic considerations. J. Med. Ethics 50, 544–551. doi: 10.1136/jme-2023-109338
Lancaster Farrell, C.-J. (2022). Explainability does not improve biochemistry staff trust in artificial intelligence-based decision support. Ann. Clin. Biochem. 59, 447–449. doi: 10.1177/00045632221128687
Laxar, D., Eitenberger, M., Maleczek, M., Kaider, A., Hammerle, F. P., and Kimberger, O. (2023). The influence of explainable vs non-explainable clinical decision support systems on rapid triage decisions: a mixed methods study. BMC Med. 21:359. doi: 10.1186/s12916-023-03068-2
Lee, J. D., and See, K. A. (2004). Trust in Automation: designing for appropriate reliance. Hum. Factors 46, 50–80. doi: 10.1518/hfes.46.1.50_30392
Liu, C. F., Chen, Z. C., Kuo, S. C., and Lin, T. C. (2022). Does AI explainability affect physicians’ intention to use AI? Int. J. Med. Inform. 168:104884. doi: 10.1016/j.ijmedinf.2022.104884
Liu, K., and Tao, D. (2022). The roles of trust, personalization, loss of privacy, and anthropomorphism in public acceptance of smart healthcare services. Comput. Human Behav. 127:107026. doi: 10.1016/j.chb.2021.107026
Louca, J., Vrublevskis, J., Eder, K., and Tzemanaki, A. (2023). Elicitation of trustworthiness requirements for highly dexterous teleoperation systems with signal latency. Front. Neurorobot. 17:1187264. doi: 10.3389/fnbot.2023.1187264
Lyons, J. B., Jessup, S. A., and Vo, T. Q. (2024). The role of decision authority and stated social intent as predictors of trust in autonomous robots. Top. Cogn. Sci. 16, 430–449. doi: 10.1111/tops.12601
Moilanen, J., van Berkel, N., Visuri, A., Gadiraju, U., van der Maden, W., and Hosio, S. (2023). Supporting mental health self-care discovery through a chatbot. Front. Digit. Health 5:1034724. doi: 10.3389/fdgth.2023.1034724
Muir, B. M., and Moray, N. (1996). Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. Ergonomics 39, 429–460. doi: 10.1080/00140139608964474
Nash, D. M., Thorpe, C., Brown, J. B., Kueper, J. K., Rayner, J., Lizotte, D. J., et al. (2023). Perceptions of artificial intelligence use in primary care: a qualitative study with providers and staff of Ontario community health Centres. J. Am. Board Fam. Med. 36, 221–228. doi: 10.3122/jabfm.2022.220177R2
OECD Legal Instruments. (2019). Available online at: https://legalinstruments.oecd.org/en/instruments/oecd-legal-0449#mainText (Accessed May 4, 2024).
Oksanen, A., Savela, N., Latikka, R., and Koivula, A. (2020). Trust toward robots and artificial intelligence: an experimental approach to human–technology interactions online. Front. Psychol. 11:568256. doi: 10.3389/fpsyg.2020.568256
Osheroff, J. A., Teich, J. M., Middleton, B., Steen, E. B., Wright, A., and Detmer, D. E. (2007). A roadmap for national action on clinical decision support. J. Am. Med. Inform. Assoc. 14, 141–145. doi: 10.1197/jamia.M2334
Qassim, S., Golden, G., Slowey, D., Sarfas, M., Whitmore, K., Perez, T., et al. (2023). A mixed-methods feasibility study of a novel AI-enabled, web-based, clinical decision support system for the treatment of major depression in adults. J. Affect. Disord. Rep. 14:100677. doi: 10.1016/j.jadr.2023.100677
Rainey, C., O'Regan, T., Matthew, J., Skelton, E., Woznitza, N., Chu, K. Y., et al. (2022). UK reporting radiographers’ perceptions of AI in radiographic image interpretation–current perspectives and future developments. Radiography 28, 881–888. doi: 10.1016/j.radi.2022.06.006
Rieger, T., Kugler, L., Manzey, D., and Roesler, E. (2024). The (im) perfect automation schema: who is trusted more, automated or human decision support? Hum. Factors 66, 1995–2007. doi: 10.1177/00187208231197347
Rodler, S., Kopliku, R., Ulrich, D., Kaltenhauser, A., Casuscelli, J., Eismann, L., et al. (2024). Patients’ trust in artificial intelligence–based decision-making for localized prostate cancer: results from a prospective trial. Eur. Urol. Focus 10, 654–661. doi: 10.1016/j.euf.2023.10.020
Rojahn, J., Palu, A., Skiena, S., and Jones, J. J. (2023). American public opinion on artificial intelligence in healthcare. PLoS One 18:e0294028. doi: 10.1371/journal.pone.0294028
Sardar, P., Abbott, J. D., Kundu, A., Aronow, H. D., Granada, J. F., and Giri, J. (2019). Impact of artificial intelligence on interventional cardiology: from decision-making aid to advanced interventional procedure assistance. JACC Cardiovasc. Interv. 12, 1293–1303. doi: 10.1016/j.jcin.2019.04.048
Schulz, P. J., Lwin, M. O., Kee, K. M., Goh, W. W., Lam, T. Y. T., and Sung, J. J. (2023). Modeling the influence of attitudes, trust, and beliefs on endoscopists’ acceptance of artificial intelligence applications in medical practice. Front. Public Health 11:1301563. doi: 10.3389/fpubh.2023.1301563
Shahzad Khan, K., Imran, A., and Nadir, R. (2024). AI-driven transformations in healthcare marketing: a qualitative inquiry into the evolution and impact of artificial intelligence on online strategies. J. Popul. Ther. Clin. Pharmacol. 1, 179–190. doi: 10.53555/jptcp.v31i1.3954
Shamszare, H., and Choudhury, A. (2023). Clinicians’ perceptions of artificial intelligence: focus on workload, risk, trust, clinical decision making, and clinical integration. Healthcare 11:2308. doi: 10.3390/healthcare11162308
Shin, H. J., Lee, S., Kim, S., Son, N. H., and Kim, E. K. (2023). Hospital-wide survey of clinical experience with artificial intelligence applied to daily chest radiographs. PLoS One 18:e0282123. doi: 10.1371/journal.pone.0282123
Sonmez, F. (2024). Going under Dr. robot’s knife: the effects of robot anthropomorphism and mortality salience on attitudes toward autonomous robot surgeons. Psychol. Health 39, 1112–1129. doi: 10.1080/08870446.2022.2130311
Soori, M., Arezoo, B., and Dastres, R. (2023). Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cogn. Robot. 3, 54–70. doi: 10.1016/j.cogr.2023.04.001
Stevens, A. F., and Stetson, P. (2023). Theory of trust and acceptance of artificial intelligence technology (TrAAIT): an instrument to assess clinician trust and acceptance of artificial intelligence. J. Biomed. Inform. 148:104550. doi: 10.1016/j.jbi.2023.104550
Szabo, B., Orsi, B., and Csukonyi, C. (2024). Robots for surgeons? Surgeons for robots? Exploring the acceptance of robotic surgery in the light of attitudes and trust in robots. BMC Psychol. 12:45. doi: 10.1186/s40359-024-01529-8
Van Bulck, L., and Moons, P. (2024). What if your patient switches from Dr. Google to Dr. ChatGPT? A vignette-based survey of the trustworthiness, value, and danger of ChatGPT-generated responses to health questions. Eur. J. Cardiovasc. Nurs. 23, 95–98. doi: 10.1093/eurjcn/zvad038
Viberg Johansson, J. V., Dembrower, K., Strand, F., and Grauman, Å. (2024). Women’s perceptions and attitudes towards the use of AI in mammography in Sweden: a qualitative interview study. BMJ Open 14:e084014. doi: 10.1136/bmjopen-2024-084014
West, D. (2018). What is artificial intelligence? Brookings. Available online at: https://www.brookings.edu/articles/what-is-artificial-intelligence/ (Accessed May 4, 2024).
Yakar, D., Ongena, Y. P., Kwee, T. C., and Haan, M. (2022). Do people favor artificial intelligence over physicians? A survey among the general population and their view on artificial intelligence in medicine. Value Health 25, 374–381. doi: 10.1016/j.jval.2021.09.004
Yi-No Kang, E., Chen, D.-R., and Chen, Y.-Y. (2023). Associations between literacy and attitudes toward artificial intelligence–assisted medical consultations: the mediating role of perceived distrust and efficiency of artificial intelligence. Comput. Human Behav. 139:107529. doi: 10.1016/j.chb.2022.107529
Keywords: artificial intelligence, trust, trust factors, healthcare, AI, factors impacting trust, AI integration
Citation: Mertz M, Toskovich K, Shields G, Attema G, Dumond J and Cameron E (2025) Exploring trust factors in AI-healthcare integration: a rapid review. Front. Artif. Intell. 8:1658510. doi: 10.3389/frai.2025.1658510
Edited by:
Deepanjali Vishwakarma, University of Limerick, IrelandReviewed by:
Valentina Ilija Janev, Institut Mihajlo Pupin, SerbiaSajjad Karimian, University College Dublin, Ireland
Copyright © 2025 Mertz, Toskovich, Shields, Attema, Dumond and Cameron. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Megan Mertz, bW1lcnR6QG5vc20uY2E=
Kelvi Toskovich1