Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med., 01 August 2025

Sec. Family Medicine and Primary Care

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1631565

This article is part of the Research TopicPatient-Centered Care: Strengthening Trust and Communication in Healthcare RelationshipsView all 18 articles

Leveraging LLM to identify missed information in patient-physician communication: improving healthcare service quality

  • 1Department of Mechanical & Industrial Engineering, University of Toronto, Toronto, ON, Canada
  • 2Institute for Global Health, University College London, London, United Kingdom

Background and objective: Electronic medical records (EMRs) have significantly changed the dynamics of physician-patient interactions, leading to a shift in communication patterns. Although various studies have developed guidelines for these new dynamics, different EMRs result in different modes of interaction, which can contribute to missed information during clinical encounters. Therefore, this study aims to develop a method that can automate the identification process of missed information to increase patient safety and satisfaction.

Methods: A total of 98 transcripts of clinical consultations from two primary care clinics in the United States were used for identifying missed information and patient unsatisfactory factors. We first examine those factors through ordinal logistic regression. Then we leveraged large language model (Phi-3.5) to develop the automation model for identifying missed information of physicians.

Results: We show that showing care and empathy to patients ( β =1.283, OR = 3.609 [95% CI: 1.836, 7.091], p <0.001) and explaining things clearly to patients ( β =1.620, OR = 5.051 [95% CI: 2.138, 11.938], p <0.001) can significantly increase the level of patient satisfaction. And our model has an average accuracy of 90.09% with F1-score of 93.75% on identifying missed information during clinical practices in primary care.

Conclusion: This study demonstrates the potential of automated analysis using Phi-3.5 to evaluate the identification of communication gaps in physician-patient interactions, ultimately enhancing patient safety and satisfaction. Further research is needed to refine this approach and explore its application across diverse healthcare settings.

1 Introduction

New technologies in healthcare are designed to assist physicians to provide healthcare services with higher efficiency which have changed the visiting patterns permanently for both physicians and patients (1, 2). For example, the use of Electronic Medical Records (EMR) saves time for physicians spending time on communication with patients so that they can take care of more patients (3, 4). However, the increasing use of EMRs and other computer-based systems in primary care has added complexity to physician-patient interactions. While these technologies have numerous benefits, including improved record-keeping and data accessibility, they can also be a source of distraction for physicians. The workload of the physicians has also increased due to the increasing number of visits (5). Studies have shown that physicians spend a significant portion of their time interacting with computers during patient visits, which can distract from communication with the patient (6, 7).

Diagnostic errors in primary care are a significant concern, with studies indicating that they occur in approximately 5 to 15% of encounters, depending on the conditions examined and the study parameters (8, 9). Essential information during the clinical practices can be missed by communication breakdowns between physicians and patients which leads to the result of incorrect diagnosis (10). Physicians may miss information from patients, including verifying personal information in health records with them and gathering details of their medical history, medication use, and findings from physical and medical examinations (11, 12). Increased computer usage has raises concerns about its impact on the quality of care and the potential for missing crucial patient information (13).

As the new technology of Artificial Intelligence (AI) is deploying in the healthcare system, the physicians are able to manage with much more information as before and take care of more patients (1416). Since physicians have less communications with patients, the new type of communication pattern is emphasized since the period of using of EMR that the patients are encouraged to join the diagnosis process during the visiting with more active interactions with physicians instead of being passive (1719). The quality of the communication between the physicians and patients is extremely important during this process, as any pieces of missed information will lead to the incorrect diagnosis and decision-makings of physicians.

Patient satisfaction and trust are also significant factors for studying communication patterns in healthcare. Although new healthcare technologies have increased efficiency of physicians. Reduced direct communication with patients often leads to perceptions of decreased attentiveness and care, resulting in lower patient satisfaction and weakened trust in the healthcare experience. Previous studies shows that the lack of interactions between physicians and patients will lead to the feeling of insufficient care for patients (19, 20). In particular, it is essential to balance the use of new technology with strategies to maintain patient-centered communication to ensure the quality of healthcare service provided and preserve the trust in the physician-patient relationship.

Therefore, it is necessary to study communication patterns and effectiveness to uncover factors that contribute to suboptimal healthcare services and instances of incorrect diagnosis (21). To identify the causes of incorrect diagnoses, previous studies have used an event-based reporting system to identify the information physicians may have missed from patients. This system is structured to monitor specific events that could potentially lead to the incidents (22, 23). However, the unique nature of primary care presents challenges for this system. In primary care, the scope of interaction is much more complex and broader than in specialized departments, such as emergency department, where primary care emphasizes a long-term, continuous relationship between physicians and patients, with comprehensive, ongoing conversations (10, 24). Hence, instead of focusing solely on critical events in emergency or surgical settings, routine patient–physician interactions are recognized as key factors that physicians need to prioritize in primary care (25, 26). And physicians in primary care are tasked to understand and address the concerns and needs from patients (27, 28). That is, patient-centered care is valued more by patients in primary care than in emergency department settings, identifying key drivers that influence the quality of interactions during clinical practice is significant (29).

Traditional methods such as manual coding are widely used to study behavior and communication patterns during clinical encounters (26, 30, 31). However, these methods present significant challenges in terms of cost and efficiency, making it difficult to analyze large datasets. As the advancements in artificial intelligence, machine learning and natural language processing approaches are increasingly being used for automating event analysis (32, 33). With the development of deep learning techniques, particularly through transformer-based Large Language Models (LLMs), conversation dialogs can be effectively modeled and understood with the technique of attention mechanism in transformers (3436). Hence, we are able to analyze complex communication patterns and capture events from a conversation by considering full context of the text, including the sequence of messages and underlying meanings across the turnovers (3740). This capability presents an opportunity to analyze information may missed during conversations in primary care while interactions and turnovers during primary care clinical practices are centralized with immediate solutions and reactions from physicians to patients (41, 42).

Therefore, this study first aims to identify the missed information from physicians in primary care and reveal factors that affect the quality of patient-centered care during clinical encounters, with a focus on their impact on patient satisfaction. Next, we develop a framework with large language model (LLM) to automate the identification of missed information from physicians, with a goal of increase patient safety in primary care clinics. Lastly, we discussed the unique nature of the missed associated with incorrect diagnosis and the needs of patients in order to provide a solution for increasing their satisfaction with physicians.

2 Materials and methods

2.1 Data collection and processing

The data used for this study was collected from previous studies which included 110 medical encounters recordings with high enough quality and the survey (31). The protocols of this study and previous studies were received and approved by Research Ethics Boards (Protocol #: 00045360) and clinicians consented to participation. Two clinical centers in the US participated in this study, eligible patients must have infected with common cold. There are 110 patients, and 5 physicians participated in this study, including 41 males and 69 females. And the mean age of patients was 34.2 (min 12.2, max 71.8). Eligible patients who agreed or authorized by guardian to participate were taken to a private consultation room with video cameras and a survey was filled out by each patient after the consultation. In the survey questionnaire, each patient was asked to rate their satisfaction with their physicians, and the level of showing care (i.e., needs, compassion) and explaining things clearly to patients. There are 5 levels of the rating, from 1 to 5 for each category.

The audio data were extracted from each encounter. Following Research Ethics Board (REB) guidelines, we used offline versions of the python programs. Transcriptions were obtained using whisper (43), and speakers were identified using pyannote.audio (44). Each transcription of an encounter was saved into a CSV file for analysis. A flowchart of the data processing process is shown in Figure 1.

Figure 1
Flowchart illustrating a three-stage process. Stage 1: Data Collection and Processing involves converting raw audio data (n=100) to text transcription (n=100). Stage 2: Data Coding and Automation transitions from manual to automatic coding. Stage 3: Data Analysis and Interpretation includes identifying missed information and conducting qualitative and quantitative analysis. Arrows indicate progression through the stages.

Figure 1. Flowchart for this study.

2.2 Clinical encounters and patient satisfaction

In the diagnostic steps for the common cold in primary care, physicians often perform physical examinations to identify symptoms and make a diagnosis (45). Studies have shown that most cases of the common cold are caused by rhinoviruses, and symptoms such as rhinorrhea, nasal congestion, and sore throat are typically identifiable through a physician’s examination, around 20% of cold cases are caused by unknown viruses (46). These cases may require laboratory tests to guide treatment effectively. However, physicians often avoid lab testing during initial visits due to the defensive healthcare practices, where simpler diagnoses are prioritized to reduce patient expenses and avoid unnecessary interventions (47, 48). In primary care, medical examinations are not required for the common cold unless symptoms have lasted more than typical duration or atypical symptoms are present (45). Therefore, based on guidance from previous studies, we developed a framework designed to identify missed information of diagnostic process toward patients performed by physicians, with the aim of increasing patient safety and satisfaction (45, 46, 49). A framework of our model on identifying missed information and procedures in clinical encounters is shown in Figure 2, with the explanation of each stage provided in Table 1. In addition, all participating physicians in this study made accurate diagnoses, and the annotation of missed information in each diagnostic step was performed by a single annotator (XZ) through a manual review of the dialog text of the encounters. And we selected showing care, explaining things clearly to patients and their satisfaction levels in the survey data (each factor is rated by 1 to 5) collected by Osan and Montague as factors in examine patient centered care to patient satisfaction (26, 50, 51). The distribution of the diagnostic process and patient centered care factors is provided in Table 2. Figure 3 is the distribution of patient satisfaction levels.

Figure 2
Flowchart showing a diagnostic process. It begins with

Figure 2. Framework of identifying missed information and diagnosis procedures.

Table 1
www.frontiersin.org

Table 1. Explanation of framework.

Table 2
www.frontiersin.org

Table 2. Distribution of diagnostic process and patient centered care factors.

Figure 3
Bar chart titled

Figure 3. Distribution of patient satisfaction.

2.3 Model specification

In this study, we first implemented an ordinal logistic model to analyze patient satisfaction. The ordinal logistic model accounts for the ordered nature of outcome without assuming equal intervals between different outcome levels (52). That is, the ordinal logistic model allows the analysis of patient satisfaction levels to inherent the order of responses without treating difference between adjacent satisfaction levels as equivalent. Given a response variable Y i with 5 satisfaction levels, the probability of a patient satisfaction for a clinical encounter can be defined as:

P ( Y i > j ) = exp ( k = 1 K β k X ik α j ) 1 + exp ( k = 1 K β k X ik α j ) , j = 1 , , 4

where j represents the level of patient satisfaction, α j is the cut-off point (threshold) for the j th satisfaction level, β k is a vector of model parameters, and X ik is a vector of all independent variables. In particular, the details of all independent variables are provided in Table 3. DurationDays, Age and Education are controlled.

Table 3
www.frontiersin.org

Table 3. Explanation of independent variables.

Then we used LLM to automize the manual qualitative coding of missed information during the clinical encounters. The Large Language Model (LLM) is a type of deep learning model designed to understand, generate, and process human language. It is built on transformer architectures, which allows the model to capture complex patterns in text through self-attention mechanisms (34, 53). It can do several tasks effectively, such as text generation, summarization and natural language understanding (37, 38).

In this study, we aim to identify the missed information from the interactions between physicians and patients. The LLM we used for this study is Phi-3.5-mini-instruct (Phi3.5) (54), a relatively small language model designed for faster inference, lower computational resource requirements, and reduced energy consumption. While many large language models, such as Llama3-8B come in different sizes (55), Phi-3.5-mini-instruct is specifically optimized its efficiency, which is more suitable for deployment in local environments.

To improve the performance of the model and increase its understanding of conversations of primary care encounters, we implemented zero-shot in-context learning on the model with the guideline of BMJ Best Practice and a dictionary of epidemiology (45, 56, 57). This allows the model to better recognize and interpret medical terms, improving its accuracy in understanding health-related texts. We first randomly selected approximately 30% of the data as a held-out validation dataset to tune the prompts for the Phi-3.5 model. The rest of data were then used for testing set to evaluate the model’s performance, including accuracy, precision, recall, and F1 score (58). Since our model identifies missed events based on a framework based on a “checklist,” we evaluate its performance by comparing the accuracy between predicted label and actual label on information collected by physicians, and compute precision, recall, F1-score are computed for the clinical encounters in testing dataset.

3 Results

3.1 Patient centered care

To examine the potential factors influencing patient satisfaction, an ordinal logistic model was fitted to our data. The estimates of the fitted model are shown in Table 4. According to previous studies (51, 59), we selected showing care, explaining things clearly, and the number of diagnostic steps performed by physicians as explanatory variables, and included duration of common cold (in days), age, and education level of patients as control variables. The result shows that showing care to patients ( β =1.283, OR = 3.609 [95% CI: 1.836, 7.091], p <0.001) and explaining things clearly to patients ( β =1.620, OR = 5.051 [95% CI: 2.138, 11.938], p <0.001) can significantly increase the level of patient satisfaction. And following proper diagnostic steps does not have impact toward patient satisfaction. However, this may be attributed to the distribution of our data (Table 2), as most of the information missed during the clinical encounter was checking patients’ daily habits, whereas physicians performed well in other diagnostic steps. That is, we can only show that missed information in daily habits does not affect patient satisfaction.

Table 4
www.frontiersin.org

Table 4. Ordinal logistics model regression results.

3.2 Automation of identifying missed information during diagnostic process

We then automated the identification of missing information in clinical encounters by utilizing the LLM (Phi3.5) for improving patient safety during the diagnostic process. Rather than monitoring for missed events, we counted the events that occurred (22, 60). In this context, a positive label indicates that the information was not missing. Additionally, a total of 98 recordings were used for this study, 12 recording were excluded due to low audio quality.

Since the data set was positively skewed (i.e., missed information was rare in clinical encounters) we first generated 50 synthetic clinical encounters. In particular, interactions within each synthetic encounter were randomly selected from original data and combined to form a coherent and realistic conversation with missed information during random diagnostic process. We then applied zero-shot learning using Phi-3.5 (57), with 40 encounters (25 actual and 15 synthetic) held out as validation set for prompt tuning on Phi-3.5, and 108 encounters (73 actual and 35 synthetic) as the test set. The prompt for zero-shot learning is provided in Appendix. The automated identification model performance is presented in Table 5. Although our model has lower accuracy (0.77), recall (0.78) and F1-score (0.85) in DailyHabits than other categories (>0.95). We can identify missed events very well since low recall in DailyHabits means that there are many times the doctor actually did collect information about daily habits, but the model failed to detect it. In overall, our model shows strong performance in capturing the events occurs (58), which means that our model can identify missed information during clinical encounters with a high performance.

Table 5
www.frontiersin.org

Table 5. Performance table of automated identification model.

4 Discussion

This study explored the impact of missed information during clinical encounters on patient safety and satisfaction in primary care using samples from university primary care clinics in the US. To identify missed information in clinical encounters, we categorized the essential diagnostic information for assessing the common cold during primary care encounters into three domains: daily habits, medical history, and physical examination. Our analysis of missed information reveals that physicians failed to collect information related to daily habits of patients in 19% of the clinical encounters. Although missed information from physicians does not show a significant effect on diagnostic outcomes in this study, this may be attributed to the imbalanced distribution on missed information, previous studies have indicated that missed information is one of the significant factors contributes to errors in primary care settings, which can lead to incorrect diagnoses (8, 9). For instance, missed information in patients’ daily habits can have various negative consequences that missing smoking habits of patients with respiratory symptoms may result in poorer disease control, increased healthcare utilization, and higher healthcare costs (61), and missing information of patients’ alcohol consumption habits may result in adverse health outcomes, particularly when medications that interact negatively with alcohol are prescribed (62).

Notably, although this study did not directly examine the impact of EMR use on missed information, previous research has shown that physicians who are proficient with electronic medical records are less likely to miss critical information during clinical encounters, as they need to input and retrieve data through the EMR system (3, 26). Furthermore, the increasing use of EMRs and emerging AI tools in clinical practice can help reduce likelihood of missed information by providing more intuitive and supportive interfaces for physicians through EMR system (4, 18). These technologies can also save time, particularly under conditions of high workload, by minimizing the need to repeatedly collect the same patient information across clinical visits (45, 63). However, workload of physicians may paradoxically increase with the use of more efficient EMR systems, particularly in the context of physician shortage (3, 64). One consequence of increased workload is burnout of physicians, which has been associated with a higher likelihood of missed information where essential information may be ignored during clinical decision-making and diagnosis (7).

With respect to patient satisfaction, our data show a high proportion of responses indicates high satisfaction levels. One significant factor contributes to these ratings appears to be the effective documentations by physicians (26, 30, 31). The ability of physicians to provide attentive care while efficiently doing documentation may enhance patients’ trust and overall satisfaction. The results of the ordinal logistic regression analysis on attentive care reveal similar findings that when physicians demonstrate compassionate care and provide clear explanations to patients, it significantly increases patient satisfaction. Our ordinal logistic regression analysis does not show a statistically significant correlation with missed information and patient satisfaction. One of the reasons is that physicians are the primary source of information for patients (19). That is, patients are likely to trust the diagnostic process even in cases where errors occur or information is missed and satisfy with physicians when they show compassionate and effective communication skills (1921). However, this may vary across different clinical settings, particularly the encounters with longer waiting times. Waiting time has been identified as a significant factor contributing to lower patient satisfaction (64, 65). Key determinants of patient satisfaction include the amount of time spent with physicians (65) and patients’ perceptions of being treated attentively and respectfully during clinical visits (64). In particular, when patients experiencing extended waiting times, they expect more comprehensive diagnostic procedures than their previous visits.

As our goal is to examine the impact of missed information during clinical encounters on patient safety and satisfaction using a larger sample in future studies, analyzing the interactions between physicians and patients through manual annotation presents substantial challenges. To address this challenge, we developed a method for automatically identifying missed information using a large language model (LLM), specifically Phi-3.5. Our model demonstrated strong performance in identifying missed information across all diagnostic steps (F1-score > 0.8) by using the designed prompt to LLM (57, 58). In addition, we found that identifying missed information in daily habit (F1-score = 0.8468) has slightly lower F1-score than other categories (F1-score > 0.97), the challenge may arise by the difficulty of collecting all relevant habits that are involved. For instance, we have only included most common habits related to the diagnosis of the common cold in this study (45). However, some physicians are interested with specific habits, such as coffee consumption, which were not considered in our model. This discrepancy can lead to incorrect coding during the identification of missed information, where physicians may have documented that certain habits were assessed, but our model lacks the context to recognize the relevance of these specific habits.

This study has several limitations that should be considered. Although previous research has demonstrated that missed information have negative impact on patient safety and satisfaction, our study did not replicate these findings, primarily due to a limited sample size. In particular, the dataset lacked sufficient cases involving incorrect diagnoses, which restricted our ability to examine the specific effects of missed information on diagnostic accuracy and patient satisfaction. However, this study is still valuable for future works. Our Phi3.5 model can classify the information during clinical practice with high performance. That is, we provided an effective method to analyze with larger sample sizes on missed information.

Furthermore, future studies should consider incorporating additional data sources, such as using EMR usage logs during clinical practices to identify the information retrieved from EMR system by physicians when patients have previously visited (4, 26). Video recordings of clinical encounters could provide valuable insights into physician behavior and cognitive load, particularly in their burnout. Physicians experiencing burnout during work may fail to properly process or apply the information available to them, even when it is accessible (7). Also, body language can be analyzed through video data, as nonverbal communication has been shown to significantly influence patient satisfaction (26, 31). In addition, a standardized scaling system for assessing ratings of physician empathy and explanatory communication should be implemented. This would enable more consistent evaluation of patient-centered care factors, potentially allowing for sentiment analysis using language models (26, 30, 54).

5 Conclusion

Our study provides evidence that improving interactions during clinical practices receive higher satisfaction from patients. The method regarding LLM we provided with, have contribute to the automated analysis on missed information in clinical interactions, ultimately enhancing patient safety by preventing incorrect diagnosis. It demonstrates potential capabilities in information identification through a multimodal approach. Future studies should examine diverse healthcare settings to ensure that the proposed framework can better align with the essential information needs of clinical practice. Additionally, technologies such as EMRs in primary care should be designed to strengthen the connection between physicians and patients, not only by improving information transparency for both sides, but also by better accommodating user preferences.

Data availability statement

The datasets presented in this article are not readily available because original data include recording of clinical practices, which can identify the participants. Requests to access the datasets should be directed to ZW5pZC5tb250YWd1ZUB1dG9yb250by5jYQ==.

Ethics statement

The studies involving humans were approved by Health Sciences REB, University of Toronto. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin.

Author contributions

XyZ: Conceptualization, Data curation, Formal analysis, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing. EC: Conceptualization, Formal analysis, Methodology, Supervision, Visualization, Writing – review & editing. XzZ: Formal analysis, Writing – review & editing. EM: Conceptualization, Data curation, Formal analysis, Funding acquisition, Methodology, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) RGPIN-2024-05057.

Acknowledgments

The authors thank all the physicians, nurses, staff and patients who participated in this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Makoul, G, Curry, RH, and Tang, PC. The use of electronic medical records: communication patterns in outpatient encounters. J Am Med Inform Assoc. (2001) 8:610–5. doi: 10.1136/jamia.2001.0080610

PubMed Abstract | Crossref Full Text | Google Scholar

2. Montague, E, and Perchonok, J. Health and wellness technology use by historically underserved health consumers: systematic review. J Med Internet Res. (2012) 14:e78. doi: 10.2196/jmir.2095

PubMed Abstract | Crossref Full Text | Google Scholar

3. Park, SY, Lee, SY, and Chen, Y. The effects of EMR deployment on doctors’ work practices: a qualitative study in the emergency department of a teaching hospital. Int J Med Inform. (2012) 81:204–17. doi: 10.1016/j.ijmedinf.2011.12.001

PubMed Abstract | Crossref Full Text | Google Scholar

4. Asan, OD, Smith, P, and Montague, E. More screen time, less face time – implications for EHR design. Evaluation Clin Prac. (2014) 20:896–901. doi: 10.1111/jep.12182

PubMed Abstract | Crossref Full Text | Google Scholar

5. Doerr, E, Galpin, K, Jones-Taylor, C, Anander, S, Demosthenes, C, Platt, S, et al. Between-visit workload in primary care. J Gen Intern Med. (2010) 25:1289–92. doi: 10.1007/s11606-010-1470-2

PubMed Abstract | Crossref Full Text | Google Scholar

6. Tai-Seale, M, Olson, CW, Li, J, Chan, AS, Morikawa, C, Durbin, M, et al. Electronic health record logs indicate that physicians Split time evenly between seeing patients and desktop medicine. Health Aff. (2017) 36:655–62. doi: 10.1377/hlthaff.2016.0811

PubMed Abstract | Crossref Full Text | Google Scholar

7. Downing, NL, Bates, DW, and Longhurst, CA. Physician burnout in the electronic health record era: are we ignoring the real cause? Ann Intern Med. (2018) 169:50–1. doi: 10.7326/m18-0139

PubMed Abstract | Crossref Full Text | Google Scholar

8. Singh, H, Meyer, AND, and Thomas, EJ. The frequency of diagnostic errors in outpatient care: estimations from three large observational studies involving US adult populations. BMJ Qual Saf. (2014) 23:727–31. doi: 10.1136/bmjqs-2013-002627

PubMed Abstract | Crossref Full Text | Google Scholar

9. Taitz, JM, Lee, TH, and Sequist, TD. A framework for engaging physicians in quality and safety. BMJ Qual Saf. (2012) 21:722–8. doi: 10.1136/bmjqs-2011-000167

PubMed Abstract | Crossref Full Text | Google Scholar

10. Murphy, DR, Singh, H, and Berlin, L. Communication breakdowns and diagnostic errors: a radiology perspective. Diagnosi. (2014) 1:253–61. doi: 10.1515/dx-2014-0035

PubMed Abstract | Crossref Full Text | Google Scholar

11. Smith, PC. Missing clinical information during primary care visits. JAMA. (2005) 293:565. doi: 10.1001/jama.293.5.565

Crossref Full Text | Google Scholar

12. Singh, H, and Carayon, P. A roadmap to advance patient safety in ambulatory care. JAMA. (2020) 324:2481–2. doi: 10.1001/jama.2020.18551

PubMed Abstract | Crossref Full Text | Google Scholar

13. Clarke, A, Adamson, J, Watt, I, Sheard, L, Cairns, P, and Wright, J. The impact of electronic records on patient safety: a qualitative study. BMC Med Inform Decis Mak. (2016) 16:62. doi: 10.1186/s12911-016-0299-y

PubMed Abstract | Crossref Full Text | Google Scholar

14. Aminololama-Shakeri, S, and López, JE. The doctor-patient relationship with artificial intelligence. Am J Roentgenol. (2019) 212:308–10. doi: 10.2214/ajr.18.20509

PubMed Abstract | Crossref Full Text | Google Scholar

15. Palanica, A, Docktor, MJ, Lee, A, and Fossat, Y. Using mobile virtual reality to enhance medical comprehension and satisfaction in patients and their families. Perspect Med Educ. (2019) 8:123–7. doi: 10.1007/s40037-019-0504-7

PubMed Abstract | Crossref Full Text | Google Scholar

16. Rossettini, G, Cook, C, Palese, A, Pillastrini, P, and Turolla, A. Pros and cons of using artificial intelligence Chatbots for musculoskeletal rehabilitation management. J Orthop Sports Phys Ther. (2023) 53:728–34. doi: 10.2519/jospt.2023.12000

PubMed Abstract | Crossref Full Text | Google Scholar

17. Truog, RD. Patients and doctors — the evolution of a relationship. N Engl J Med. (2012) 366:581–5. doi: 10.1056/nejmp1110848

PubMed Abstract | Crossref Full Text | Google Scholar

18. Lorenzini, G, Arbelaez Ossa, L, Shaw, DM, and Elger, BS. Artificial intelligence and the doctor–patient relationship expanding the paradigm of shared decision making. Bioethics. (2023) 37:424–9. doi: 10.1111/bioe.13158

PubMed Abstract | Crossref Full Text | Google Scholar

19. Asan, O, Yu, Z, and Crotty, BH. How clinician-patient communication affects trust in health information sources: temporal trends from a national cross-sectional survey. PLoS One. (2021) 16:e0247583. doi: 10.1371/journal.pone.0247583

PubMed Abstract | Crossref Full Text | Google Scholar

20. Pearson, SD, and Raeke, LH. Patients’ trust in physicians: many theories, few measures, and little data. J Gen Intern Med. (2000) 15:509–13. doi: 10.1046/j.1525-1497.2000.11002.x

PubMed Abstract | Crossref Full Text | Google Scholar

21. Ratna, H. The importance of effective communication in healthcare practice. HPHR. (2019) 1–6. doi: 10.54111/0001/w4

Crossref Full Text | Google Scholar

22. Bhasale, A. The wrong diagnosis: identifying causes of potentially adverse events in general practice using incident monitoring. Fam Pract. (1998) 15:308–18.

Google Scholar

23. Hewitt, TA, and Chreim, S. Fix and forget or fix and report: a qualitative study of tensions at the front line of incident reporting. BMJ Qual Saf. (2015) 24:303–10. doi: 10.1136/bmjqs-2014-003279

PubMed Abstract | Crossref Full Text | Google Scholar

24. Deledda, G, Moretti, F, Rimondini, M, and Zimmermann, C. How patients want their doctor to communicate. A literature review on primary care patients’ perspective. Patient Educ Couns. (2013) 90:297–306. doi: 10.1016/j.pec.2012.05.005

PubMed Abstract | Crossref Full Text | Google Scholar

25. Gabriel, SE. Primary care: specialists or generalists. Mayo Clin Proc. (1996) 71:415–9.

Google Scholar

26. Asan, O, and Montague, E. Physician interactions with electronic health records in primary care. Health Systems. (2012) 1:96–103. doi: 10.1057/hs.2012.11

PubMed Abstract | Crossref Full Text | Google Scholar

27. Porter, ME, Pabo, EA, and Lee, TH. Redesigning primary care: a strategic vision to improve value by organizing around patients’ needs. Health Aff. (2013) 32:516–25. doi: 10.1377/hlthaff.2012.0961

Crossref Full Text | Google Scholar

28. Tong, ST, Liaw, WR, Kashiri, PL, Pecsok, J, Rozman, J, Bazemore, AW, et al. Clinician experiences with screening for social needs in primary care. J Am Board Fam Med. (2018) 31:351–63. doi: 10.3122/jabfm.2018.03.170419

PubMed Abstract | Crossref Full Text | Google Scholar

29. Wu, N, and Woloski, JR. Emergency department versus primary care use: a patient perspective. PRiMER. (2024) 8:44. doi: 10.22454/primer.2024.526921

PubMed Abstract | Crossref Full Text | Google Scholar

30. Asan, O, Xu, J, and Montague, E. Dynamic comparison of physicians’ interaction style with electronic health Records in Primary Care Settings. Aust J Gen Pract. (2013) 2:e1000137. doi: 10.4172/2329-9126.1000137

PubMed Abstract | Crossref Full Text | Google Scholar

31. Montague, E, Chen, P-Y, Xu, J, Chewning, B, and Barrett, B. Nonverbal interpersonal interactions in clinical encounters and patient perceptions of empathy. J Participat Med. (2013) 5:1–17.

Google Scholar

32. Fong, A. Realizing the power of text mining and natural language processing for analyzing patient safety event narratives: the challenges and path forward. J Patient Saf. (2021) 17:e834–6. doi: 10.1097/pts.0000000000000837

PubMed Abstract | Crossref Full Text | Google Scholar

33. Fong, A, Behzad, S, Pruitt, Z, and Ratwani, RM. A machine learning approach to reclassifying miscellaneous patient safety event reports. J Patient Saf. (2021) 17:e829–33. doi: 10.1097/pts.0000000000000731

PubMed Abstract | Crossref Full Text | Google Scholar

34. Vaswani, A, Shazeer, N, Parmar, N, Uszkoreit, J, Jones, L, Gomez, AN, et al. Attention Is All You Need. Advances in Neural Information Processing Systems (2017) 30:6000–6010. doi: 10.48550/ARXIV.1706.03762

Crossref Full Text | Google Scholar

35. Butow, P, and Hoque, E. Using artificial intelligence to analyse and teach communication in healthcare. Breast. (2020) 50:49–55. doi: 10.1016/j.breast.2020.01.008

PubMed Abstract | Crossref Full Text | Google Scholar

36. Gianola, S, Bargeri, S, Castellini, G, Cook, C, Palese, A, Pillastrini, P, et al. Performance of ChatGPT compared to clinical practice guidelines in making informed decisions for lumbosacral radicular pain: a cross-sectional study. J Orthop Sports Phys Ther. (2024) 54:222–8. doi: 10.2519/jospt.2024.12151

PubMed Abstract | Crossref Full Text | Google Scholar

37. Blank, IA. What are large language models supposed to model? Trends Cogn Sci. (2023) 27:987–9. doi: 10.1016/j.tics.2023.08.006

PubMed Abstract | Crossref Full Text | Google Scholar

38. Zhou, L, Suominen, H, and Gedeon, T. Adapting state-of-the-art deep language models to clinical information extraction systems: potentials, challenges, and solutions. JMIR Med Inform. (2019) 7:e11499. doi: 10.2196/11499

PubMed Abstract | Crossref Full Text | Google Scholar

39. Mitchell, JR, Szepietowski, P, Howard, R, Reisman, P, Jones, JD, Lewis, P, et al. A question-and-answer system to extract data from free-text oncological pathology reports (CancerBERT network): development study. J Med Internet Res. (2022) 24:e27210. doi: 10.2196/27210

PubMed Abstract | Crossref Full Text | Google Scholar

40. Wornow, M, Xu, Y, Thapa, R, Patel, B, Steinberg, E, Fleming, S, et al. The shaky foundations of large language models and foundation models for electronic health records. NPJ Digit Med. (2023) 6:135. doi: 10.1038/s41746-023-00879-8

PubMed Abstract | Crossref Full Text | Google Scholar

41. MacKichan, F, Brangan, E, Wye, L, Checkland, K, Lasserson, D, Huntley, A, et al. Why do patients seek primary medical care in emergency departments? An ethnographic exploration of access to general practice. BMJ Open. (2017) 7:e013816. doi: 10.1136/bmjopen-2016-013816

PubMed Abstract | Crossref Full Text | Google Scholar

42. Kuzel, AJ. Patient reports of preventable problems and harms in primary health care. Annals Family Med. (2004) 2:333–40. doi: 10.1370/afm.220

PubMed Abstract | Crossref Full Text | Google Scholar

43. Radford, A, Kim, JW, Xu, T, Brockman, G, McLeavey, C, and Sutskever, I. Robust Speech Recognition via Large-Scale Weak Supervision. International conference on machine learning. PMLR (2022) p. 28492–28518.

Google Scholar

44. Plaquet, A, and Bredin, H. Powerset multi-class cross entropy loss for neural speaker diarization. arXiv preprint arXiv. (2023):231013025.

Google Scholar

45. Common Cold. BMJ best practice. (2024). Available online at: https://bestpractice.bmj.com/topics/en-gb/252 [Accessed January 9, 2025]

Google Scholar

46. The common cold. Principles and practice of pediatric infectious diseases. Netherland: Elsevier (2018). p. 199–202.e1

Google Scholar

47. Lorenc, T, Khouja, C, Harden, M, Fulbright, H, and Thomas, J. Defensive healthcare practice: systematic review of qualitative evidence. BMJ Open. (2024) 14:e085673. doi: 10.1136/bmjopen-2024-085673

PubMed Abstract | Crossref Full Text | Google Scholar

48. Naugler, C. Laboratory test use and primary care physician supply. Can Fam Physician. (2013) 59:e240–5.

PubMed Abstract | Google Scholar

49. Wetterneck, TB, Lapin, JA, Krueger, DJ, Holman, GT, Beasley, JW, and Karsh, B-T. Development of a primary care physician task list to evaluate clinic visit workflow: table 1. BMJ Qual Saf. (2012) 21:47–53. doi: 10.1136/bmjqs-2011-000067

PubMed Abstract | Crossref Full Text | Google Scholar

50. Street, RL. The many “disguises” of patient-centered communication: problems of conceptualization and measurement. Patient Educ Couns. (2017) 100:2131–4. doi: 10.1016/j.pec.2017.05.008

PubMed Abstract | Crossref Full Text | Google Scholar

51. Świątoniowska-Lonc, N, Polański, J, Tański, W, and Jankowska-Polańska, B. Impact of satisfaction with physician–patient communication on self-care and adherence in patients with hypertension: cross-sectional study. BMC Health Serv Res. (2020) 20:1046. doi: 10.1186/s12913-020-05912-0

PubMed Abstract | Crossref Full Text | Google Scholar

52. Ordinal Logistic Regression. Springer Series in Statistics. Cham: Springer International Publishing (2015). p. 311–325

Google Scholar

53. Tai, RH, Bentley, LR, Xia, X, Sitt, JM, Fankhauser, SC, Chicas-Mosier, AM, et al. An examination of the use of large language models to aid analysis of textual data. Int J Qual Methods. (2024) 23:14. doi: 10.1177/16094069241231168

Crossref Full Text | Google Scholar

54. Haider, E, Perez-Becker, D, Portet, T, Madan, P, Garg, A, Ashfaq, A, et al. Phi-3 safety post-training: aligning language models with a “break-fix” cycle. arXiv preprint (2024) arXiv:2407.13833. doi: 10.48550/ARXIV.2407.13833

Crossref Full Text | Google Scholar

55. Touvron, H, Lavril, T, Izacard, G, Martinet, X, Lachaux, M-A, Lacroix, T, et al. LLaMA: open and efficient foundation language models. arXiv preprint (2023) arXiv:2302.13971. doi: 10.48550/ARXIV.2302.13971

Crossref Full Text | Google Scholar

56. Porta, M. A dictionary of epidemiology (DRAFT). Oxford: Oxford University Press (2014).

Google Scholar

57. Pan, J, Gao, T, Chen, H, and Chen, D. What in-context learning “learns” in-context: disentangling task recognition and task learning. arXiv preprint (2023). doi: 10.48550/ARXIV.2305.09731,

Crossref Full Text | Google Scholar

58. Rainio, O, Teuho, J, and Klén, R. Evaluation metrics and statistical tests for machine learning. Sci Rep. (2024) 14:6086. doi: 10.1038/s41598-024-56706-x

PubMed Abstract | Crossref Full Text | Google Scholar

59. Rakel, DP, Hoeft, TJ, Barrett, BP, Chewning, BA, Craig, BM, and Niu, M. Practitioner empathy and the duration of the common cold. Fam Med. (2009) 41:494–501.

PubMed Abstract | Google Scholar

60. Hales, BM, and Pronovost, PJ. The checklist—a tool for error management and performance improvement. J Crit Care. (2006) 21:231–5. doi: 10.1016/j.jcrc.2006.06.002

PubMed Abstract | Crossref Full Text | Google Scholar

61. Mäkelä, MJ, Backer, V, Hedegaard, M, and Larsson, K. Adherence to inhaled therapies, health outcomes and costs in patients with asthma and COPD. Respir Med. (2013) 107:1481–90. doi: 10.1016/j.rmed.2013.04.005

PubMed Abstract | Crossref Full Text | Google Scholar

62. Traccis, F, Presciuttini, R, Pani, PP, Sinclair, JMA, Leggio, L, and Agabio, R. Alcohol-medication interactions: a systematic review and meta-analysis of placebo-controlled trials. Neurosci Biobehav Rev. (2022) 132:519–41. doi: 10.1016/j.neubiorev.2021.11.019

PubMed Abstract | Crossref Full Text | Google Scholar

63. Cheung, A, Stukel, TA, Alter, DA, Glazier, RH, Ling, V, Wang, X, et al. Primary care physician volume and quality of diabetes care: a population-based cohort study. Ann Intern Med. (2017) 166:240–7. doi: 10.7326/m16-1056

PubMed Abstract | Crossref Full Text | Google Scholar

64. Spechbach, H, Rochat, J, Gaspoz, J-M, Lovis, C, and Ehrler, F. Patients’ time perception in the waiting room of an ambulatory emergency unit: a cross-sectional study. BMC Emerg Med. (2019) 19:41. doi: 10.1186/s12873-019-0254-1

PubMed Abstract | Crossref Full Text | Google Scholar

65. Anderson, RT, Camacho, FT, and Balkrishnan, R. Willing to wait?: the influence of patient wait time on satisfaction with primary care. BMC Health Serv Res. (2007) 7:31. doi: 10.1186/1472-6963-7-31

PubMed Abstract | Crossref Full Text | Google Scholar

Appendix

In this study, we applied prompt engineering with Phi3.5, with the following prompts:

Here are examples of diagnostic steps:

1. A history eliciting a constellation of symptoms compatible with the diagnosis

2. Identification of risk factors suggestive of the condition (for example, seasonal occurrence, smoking, exposure to affected individuals)

3. A brief physical examination, including temperature, pulse and blood pressure, and examination of oropharynx, nares, neck, and chest. If the patient is unwell or observations are outside of normal limits, consider other causes or complications such as influenza serious bacterial infections such as pneumonia or meningitis, or sepsis, and tailor physical examination accordingly

4. Excluding alternative diagnoses by screening for distinguishing features of conditions with overlapping symptoms, such as allergic rhinitis.

The following is a conversation between a physician and a patient in a clinical encounter that patient infect with common cold (Speaker 1: Physician, Speaker 0: Patient, Speaker U: Nurse/Staff):

{Conversation dialog}.

Please summarize the following into ‘Yes’ or ‘No’ answers (If patient shares the information before physician asks, it also counts as yes):

1. Did Physician receive information about the daily habits of patients?

a. Examples: alcohol, smoke/smoking/tobacco, jogging, etc.

1. Did Physician receive information about the medical history of patients?

a. Examples: symptoms (A symptom is any subjective evidence of disease perceived by the patient), allergies, genetic diseases (such as Heart Disease, Asthma or other genetic disease affecting breath) compatible with the diagnosis.

1. Did Physician receive physical exam information from patients?

a. Patient provide information about the information or feelings of their temperature, oropharynx, nares, neck, chest

b. Physician perform physical exam to patients with instructions such as Now I will check your XXX. (XXX: temperature, oropharynx, nares, neck, chest of patients).

Keywords: large language model (LLM), automation, patient satisfaction, patient safety, missed information

Citation: Zhou X, Cohen E, Zhou X and Montague E (2025) Leveraging LLM to identify missed information in patient-physician communication: improving healthcare service quality. Front. Med. 12:1631565. doi: 10.3389/fmed.2025.1631565

Received: 19 May 2025; Accepted: 15 July 2025;
Published: 01 August 2025.

Edited by:

Waseem Jerjes, Imperial College London, United Kingdom

Reviewed by:

Giacomo Rossettini, University of Verona, Italy
Vishal Shetty, University of Massachusetts Amherst, United States

Copyright © 2025 Zhou, Cohen, Zhou and Montague. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Enid Montague, ZW5pZC5tb250YWd1ZUB1dG9yb250by5jYQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.