SYSTEMATIC REVIEW article

Front. Digit. Health, 03 February 2025

Sec. Connected Health

Volume 7 - 2025 | https://doi.org/10.3389/fdgth.2025.1482712

Comparative analysis of ChatGPT and Gemini (Bard) in medical inquiry: a scoping review

  • FH

    Fattah H. Fattah 1,2

  • AM

    Abdulwahid M. Salih 1,2

  • AM

    Ameer M. Salih 1,3

  • SK

    Saywan K. Asaad 1,2

  • AK

    Abdullah K. Ghafour 1

  • RB

    Rawa Bapir 1,4,5

  • BA

    Berun A. Abdalla 1,5

  • SO

    Snur Othman 5

  • SM

    Sasan M. Ahmed 1,5

  • SJ

    Sabah Jalal Hasan 1

  • YM

    Yousif M. Mahmood 1

  • FH

    Fahmi H. Kakamad 1,2,5*

  • 1. Scientific Affairs Department, Smart Health Tower, Sulaymaniyah, Iraq

  • 2. College of Medicine, University of Sulaimani, Sulaymaniyah, Iraq

  • 3. Civil Engineering Department, College of Engineering, University of Sulaimani, Sulaymaniyah, Iraq

  • 4. Department of Urology, Sulaimani Surgical Teaching Hospital, Sulaymaniyah, Iraq

  • 5. Kscien Organization for Scientific Research (Middle East Office), Sulaymaniyah, Iraq

Article metrics

View details

30

Citations

8,1k

Views

1,5k

Downloads

Abstract

Introduction:

Artificial intelligence and machine learning are popular interconnected technologies. AI chatbots like ChatGPT and Gemini show considerable promise in medical inquiries. This scoping review aims to assess the accuracy and response length (in characters) of ChatGPT and Gemini in medical applications.

Methods:

The eligible databases were searched to find studies published in English from January 1 to October 20, 2023. The inclusion criteria consisted of studies that focused on using AI in medicine and assessed outcomes based on the accuracy and character count (length) of ChatGPT and Gemini. Data collected from the studies included the first author's name, the country where the study was conducted, the type of study design, publication year, sample size, medical speciality, and the accuracy and response length.

Results:

The initial search identified 64 papers, with 11 meeting the inclusion criteria, involving 1,177 samples. ChatGPT showed higher accuracy in radiology (87.43% vs. Gemini's 71%) and shorter responses (907 vs. 1,428 characters). Similar trends were noted in other specialties. However, Gemini outperformed ChatGPT in emergency scenarios (87% vs. 77%) and in renal diets with low potassium and high phosphorus (79% vs. 60% and 100% vs. 77%). Statistical analysis confirms that ChatGPT has greater accuracy and shorter responses than Gemini in medical studies, with a p-value of <.001 for both metrics.

Conclusion:

This Scoping review suggests that ChatGPT may demonstrate higher accuracy and provide shorter responses than Gemini in medical studies.

Introduction

Artificial Intelligence (AI) and Machine Learning (ML) are interconnected technologies recently gaining significant popularity. AI involves creating intelligent machines capable of performing tasks that typically require human intelligence, such as visual perception, speech recognition, decision-making, and language translation. ML, a subset of AI, focuses on developing algorithms and statistical models that enable machines to learn from data and improve their performance over time without explicit programming (1, 2). AI is at the forefront of transforming various aspects of our lives by altering how we analyze information and enhancing decision-making through problem-solving, reasoning, and learning (3).

In the dynamic domain of AI chatbots, the comparative analysis of ChatGPT and Gemini (formerly known as Google's Bard) has emerged as a focal point, particularly in medical inquiries (1, 2). Recent investigations have explored the precision and effectiveness of these AI models in fielding medical questions across various specialties (1, 47). These studies demonstrate ChatGPT's capabilities in diagnostic imaging and clinical decision support, underscoring its potential value in healthcare settings (46).

In recent years, AI models like ChatGPT and Gemini have significantly impacted natural language processing, particularly in healthcare. ChatGPT, developed by OpenAI, provides relevant and accurate text-based responses using a large dataset (1, 5). While Gemini, from Google DeepMind, integrates multimodal capabilities, handling text, audio, and video, which is especially useful in medical imaging (5). However, both models face challenges. ChatGPT, for example, shows variability in psychiatric assessments and struggles with complex cases (2). Additionally, AI models still struggle to interpret nuanced human emotions and contexts (4).

While AI chatbots like ChatGPT and Gemini show promise in medicine, extensive research is still required to understand their capabilities properly. It is essential to address the variation in their performance across different medical scenarios and enhance their accuracy for various medical applications (8). The use of AI in healthcare faces several challenges, including data privacy, algorithm accuracy, adherence to ethical standards, societal acceptance, and clinical integration (9, 10). These challenges make it difficult to develop precise and reliable AI systems. Privacy concerns restrict access to relevant data, and potential biases can result in inaccurate outcomes (8, 10).

This scoping review aims to evaluate and compare the accuracy and length of ChatGPT and Gemini (Google's Bard) in addressing medical inquiries across diverse fields, focusing on their strengths, limitations, and practical implications for healthcare. As AI models become increasingly integrated into clinical and educational settings, understanding their performance variability is essential. Both models face challenges, including inconsistencies in complex cases, privacy concerns, and ethical issues. This review offers insights to help researchers, practitioners, and developers optimize these tools for more effective decision-making and patient care.

Methods

Study protocols

We applied a systematic approach to assess the methodological quality of our scoping review, including comprehensive literature searches, double screening, bias assessment, and evaluation of publication bias.

Data sources and search strategy

A systematic search was conducted in databases and search engines, including Google Scholar, PubMed/MEDLINE, Cochrane Library, Web of Science, CINAHL, and EMBASE, using keywords such as (“ChatGPT” OR “GPT-3” OR “GPT-4” OR “Bard” OR “Gemini”) AND (“Medical” OR “Healthcare” OR “Clinical” OR “Health Inquiry” OR “Medical Inquiry”) AND (“comparison” OR “comparative” OR “analysis” OR “review”) to identify studies published from January 1 to October 20, 2023. The search was restricted to studies published in English and related to human health subjects.

Eligibility criteria

To be included in this study, studies needed to meet the following criteria: focus on the application of ChatGPT and Gemini across different branch specialties, evaluate outcomes based on the accuracy and character count of ChatGPT and Gemini, and be verified against the most recent predatory journal list (11). Additionally, review articles and case reports were excluded.

Study selection process

The initial screening involved two researchers reviewing all titles and abstracts to check if they met the eligibility criteria. In case of disagreements, a third author was consulted to reach a final decision and resolve conflicts between the initial researchers.

Data items

The data collected from the studies included the first author's name, country of study, type of study design, publishing year, sample size, type of medical specialty, accuracy, and length (character) of ChatGPT and Gemini. Accuracy refers to the ability of ChatGPT and Gemini to provide contextually appropriate and correct responses to medical questions based on the standard guidelines specific to each medical specialty.

Data analysis and synthesis

Microsoft Excel (2019) was utilized to collect and organize the extracted data, while descriptive analysis was conducted using the Statistical Package for Social Sciences (SPSS) software (version 26). The data is displayed as frequencies, percentages, means, and standard deviations.

Results

Study selection

During the initial database search, a total of 64 articles were identified. Pre-screening procedures removed one duplicate, two articles in non-English languages, and eight with unretrievable data. Following a comprehensive review of titles and abstracts, 53 studies were assessed, excluding 22 for lack of relevance. The remaining 31 studies underwent full-text evaluation, excluding 19 for failing to meet the inclusion criteria. Among the 12 studies that proceeded to the eligibility assessment phase, one was excluded due to its publication in a predatory journal. Ultimately, 11 studies met the criteria for inclusion in the review (Figure 1).

Figure 1

Characteristics of the included studies

The summarized raw data from the included studies are all observational in Tables 1, 2. India and the United States were the primary contributors, providing two studies. Additionally, Canada, Singapore, Turkey, Australia, Ecuador, and Iraq each contributed one study (Table 1).

Table 1

NoAuthorType of studyPublishing yearsCountrySample sizeSpecialty
1Patil (12)Cross-sectional2023Canada29 (9.12%)Neuroradiology
19 (5.97%)Mammography
89 (27.99%)General & physics
30 (9.43%)Nuclear medicine
16 (5.03%)Pediatric Radiology
26 (8.18%)Interventional radiology
29 (9.12%)Gastrointestinal radiology
11 (3.46%)Genitourinary radiology
16 (5.03%)Cardiac radiology
6 (1.89%)Chest radiology
25 (7.86%)Musculoskeletal radiology
22 (6.92%)Ultrasound
2Kumari (13)Cross-sectional2023India50Hematology
3Dhanvijay (14)Cross-sectional2023India77Physiology
4Muhialdeen (15)Cross-sectional2023Iraq20Clinical Diagnosis
5Koga (16)Cross-sectional2023USA25neurodegenerative disorders
6Zhi Wei Lim (17)Cross-sectional2023Singapore31myopia care
7Ilgaz (18)Cross-sectional2023Turkey131Anatomy
8Seth (19)Cross-sectional2023Australia6Rhinoplasty
9Salazar (20)Cross-sectional2023Ecuador75Emergency
101Non-emergency
10Qarajeh (21)Cross-sectional2023USA81 (33.75%)Renal Diet High potassium
68 (28.33%)Renal Diet Low potassium
91 (37.91%)Renal Diet High phosphorus
11Toyama (22)Cross-sectional2023Japan103Radiology

Baseline characteristics of the included studies.

Table 2

NoAuthorSpecialtyChatGPT accurateGemini accurateChatGPT length (character)Gemini length (character
1Patil (12)aNeuroradiology100.00%86.21%840.90 (±426.35)1,443.52 (±415.88)
Mammography84.21%68.42%787.63 (±447.38)1,454.95 (±442.34)
General & physics85.39%68.54%1,022.38 (±453.50)1,490.69 (±406.58)
Nuclear medicine80.00%56.67%947.30 (±486.57)1,321.57 (±374.86)
Pediatric Radiology93.75%68.75%764.63 (±330.04)1,368.88 (±547.91)
Interventional radiology88.46%80.77%952.31 (±510.00)1,538.31 (±446.90)
Gastrointestinal radiology89.66%79.31%901.93 (±423.01)1,427.66 (±322.21)
Genitourinary radiology72.73%63.64%1,048.82 (±338.28)1,373.09 (±342.2)
Cardiac radiology75.00%68.75%915.50 (±3.48)1,537.94 (±692.44)
Chest radiology100.00%83.33%816.33 (±303.77)1,492.33 (±295.28)
Musculoskeletal radiology80.00%64.00%945.48 (±394.93)1,326.40 (±316.33)
Ultrasound100.00%63.64%944.91 (±518.11)1,371.95 (±352.66)
2Kumari (13)bHematology3.15/5 (63%)A2.23/5 (44%)A--
3Dhanvijy (14)bPhysiology3.19/4 (79%)A2.15/4 (53%)A--
4Muhialdeen (15)bClinical Diagnosis90%80%--
5Koga (16)aneurodegenerative disorders84%76%--
6Zhi Wei Lim (17)amyopia care80.6%54.8%1,221.13 (±323.32)1,275.87 (±393.25)
7Ilgaz (18)bAnatomy44.27%41.98%--
8Seth (19)bRhinoplasty4/5 (80%)A4/5 (80%)A--
9Salazar (20)bEmergency77%87%--
Non-emergency36%33%--
10Qarajeh (21)aRenal Diet High potassium99%79%--
Renal Diet Low potassium60%79%--
Renal Diet high phosphorus77%100%--
11Toyama (22)aRadiology65%39%--

Comparison between ChatGPT and Bard.

a

ChatGPT-4.

b

ChatGPT-3.5.

Main findings

The research included 1,177 samples from 11 different medical specialties (1222). In radiology, there were 421 samples, encompassing various fields such as neuroradiology (9.12%), mammography (5.97%), general and physics (27.99%), nuclear medicine (9.43%), pediatric radiology (5.03%), interventional radiology (8.18%), gastrointestinal radiology (9.12%), genitourinary radiology (3.46%), cardiac radiology (5.03%), chest radiology (1.89%), musculoskeletal radiology (7.86%), and ultrasound (6.92%). The renal sample size was 240, divided into the renal diet with high potassium (33.75%), the renal diet with low potassium (28.33%), and the renal diet with high phosphorus (37.91%). Emergency and non-emergency cases had sample sizes of 176, while the smallest samples were in clinical diagnosis and neurodegenerative disorders, with sizes of 20 and 25, respectively (Table 1).

The comparison between ChatGPT and Gemini across various specialties reveals accuracy and response length differences. ChatGPT generally may demonstrate higher accuracy than Gemini, especially in radiology specialties. The average accuracy of ChatGPT was 87.43%, higher than Gemini 71%. Additionally, the average response length of ChatGPT was 907 characters, shorter than Gemini's 1,428 characters. This indicates that ChatGPT's accuracy relative to response length may be more reliable and accurate than Gemini. Accuracy in the hematology specialty, ChatGPT, was 63%, compared to Gemini's 44%. Similar trends are observed in physiology, clinical diagnosis, neurodegenerative disorders, anatomy, renal diet, high potassium, and radiology. Conversely, in myopia care, the response lengths of ChatGPT and Gemini were nearly the same (1,221.13 vs. 1,275.87), with ChatGPT achieving a higher accuracy of 80.6% compared to Gemini's 54.8%. In rhinoplasty, both ChatGPT and Gemini demonstrate the same accuracy. In contrast, Gemini may be more accurate than ChatGPT in emergency scenarios, a renal diet with low potassium and a renal diet with high phosphorus (87% vs. 77%, 79% vs. 60%, and 100% vs. 77%, respectively) (Table 2).

The statistical analysis compared the accuracy and response length of ChatGPT and Gemini. The results indicate that ChatGPT has a higher accuracy (72.06%) than Gemini (63.38%), with a mean difference of 8.68, a confidence interval of 7.77–9.58, and a statistically significant p-value of <.001. In terms of response length, ChatGPT produces shorter responses (960.84 words) compared to Gemini (1,423.15 words), with a mean difference of 462.31 and a similarly significant p-value of <.001. This statistical comparison emphasizes that ChatGPT may be more accurate and generates shorter responses than Gemini (Table 3).

Table 3

MeanMean differenceStd. deviationConfidence intervalp-value
LowerUpper
ChatGPT Accuracy72.068.6815.827.779.58<.001
Gemini Accuracy63.38
ChatGPT Length960.84462.31158.10445.64478.98<.001
Gemini Length1,423.15

Statistical analysis of ChatGPT and Gemini.

Discussion

Implementing large language models (LLMs) in medical education shows significant potential for transforming traditional teaching methods. Models like ChatGPT and Gemini process extensive medical literature, providing valuable, contextually relevant information for educators and students (23, 24). LLMs create interactive, dynamic learning by giving students access to current medical data, clarifying complex concepts, and enhancing problem-solving. They also improve knowledge retrieval and support evidence-based decision-making. Incorporating LLMs encourages self-directed learning, critical thinking, and ongoing professional growth. However, recognizing their limitations and biases is essential for responsible, ethical use, complemented by practical training and clinical mentorship (25, 26).

AI models have demonstrated significant potential in assisting medical professionals by enhancing efficiency in problem-solving, diagnosis, and data interpretation. For instance, ChatGPT has consistently outperformed models like Bard and Bing in accuracy when addressing medical vignettes (13, 14). This underscores AI's pivotal role in supporting clinical decision-making, particularly in complex fields such as hematology (13). However, despite these encouraging results, AI models still face limitations, including inconsistencies in performance across various medical specialties, which necessitate further refinement before full integration into clinical practice (21).

The comparative analysis of ChatGPT and Gemini across various medical specialties reveals distinct patterns in their accuracy and response length performance. ChatGPT consistently demonstrates higher accuracy rates compared to Gemini in most specialties. This trend is evident in specialties such as neuroradiology (100% vs. 86.21%), hematology (63% vs. 44%), physiology (79% vs. 53%), clinical diagnosis (90% vs. 80%) and neurodegenerative disorders (84% vs. 76%). ChatGPT's superior accuracy indicates its potential as a reliable tool for medical inquiries, providing precise and dependable information across various medical fields (1216).

Despite Gemini's lower accuracy rates, it consistently delivers longer responses than ChatGPT. In neuroradiology, Gemini's responses averaged 1,443.52 characters compared to ChatGPT's 840.90 characters. This pattern is repeated across other specialties, such as mammography (1,454.95 vs. 787.63) and general & physics (1,490.69 vs. 1,022.38) (12).

The longer response length of Gemini suggests that it may offer more detailed and comprehensive information, which could be beneficial in scenarios where a more exhaustive explanation is needed. While comparing ChatGPT and Gemini for accuracy and response length in chest radiology and ultrasound, ChatGPT consistently outperforms Gemini in accuracy. ChatGPT achieves 100% accuracy for chest radiology compared to Gemini's 83.33%, with a shorter average response length of 816.33 vs. Gemini's 1,492.33 characters. ChatGPT also has a perfect accuracy rate of 100% in ultrasound, while Gemini's accuracy drops to 63.64%. Similarly, ChatGPT's responses are more concise, averaging 944.91 characters compared to Gemini's 1,371.95 characters (12).

ChatGPT and Gemini in myopia care respond to similar lengths (1,221.13 and 1,275.87 characters, respectively). However, there is a difference in accuracy: ChatGPT achieves an accuracy of 80.6%, whereas Gemini achieves 54.8%. This disparity suggests that while both models may offer comparable responses in terms of content, ChatGPT tends to provide more reliable and accurate information in this specialized medical context (17).

ChatGPT and Gemini exhibit nearly identical accuracy in anatomy and rhinoplasty. ChatGPT achieves 44.27% accuracy in anatomy, slightly higher than Gemini's 41.98%. For rhinoplasty, both models perform equally well, each with an accuracy rate of 80%. This comparison demonstrates that ChatGPT and Gemini perform similarly in these medical specialties (18, 19).

Exceptions to this trend were observed in emergency scenarios, where Gemini achieved higher accuracy (87%) compared to ChatGPT (77%) (20). This highlights that Gemini may have strengths in specific contexts, such as emergencies where detailed information could be critical. However, both models showed lower accuracy rates in non-emergency scenarios, with ChatGPT slightly outperforming Gemini (36% vs. 33%) (20).

The performance of ChatGPT and Gemini in providing dietary advice for renal conditions also varied. ChatGPT excelled in high potassium contexts (99% vs. 79%) but was less accurate in low potassium and high phosphorus scenarios compared to Gemini (77% vs. 100%) (21). This variability suggests that each model may have specialized strengths in specific medical contexts, and their combined use could potentially enhance the quality of medical inquiry responses.

The comparative analysis of ChatGPT and Gemini (Bard) in medical inquiry highlights several limitations. ChatGPT may provide inaccurate medical information due to its limited understanding of complex contexts, and biases in training data can affect accuracy. Ethical concerns include the risk of outdated information and issues related to patient data privacy. Additionally, the evolving nature of large language models means that ChatGPT and Gemini are frequently updated, potentially rendering some findings obsolete as newer versions are released. The study's focus on specific models and predefined case vignettes may restrict its findings, as the scope of medical inquiries is limited to particular scenarios, which may not fully capture the broad range of medical topics these models could encounter. Moreover, potential biases in the responses of these language models were not fully explored, affecting the generalizability of the results. There may be limitations and potential bias in measuring accuracy, as each specialty uses different standard answers to compare with the responses of ChatGPT and Gemini across various studies. This variability makes it challenging to determine how accurately the models perform in each specialty. The findings indicate that ChatGPT generally offers more accurate and concise responses across various medical specialties, while Gemini provides more detailed but less accurate answers. The choice between these AI models should be guided by the specific needs of the medical inquiry—whether precision or detail is prioritized. Future improvements should aim to integrate the strengths of both models, enhancing accuracy while maintaining the comprehensiveness of responses to support better clinical decision-making and patient care.

Conclusion

This scoping review indicates that ChatGPT has shown promise in the included medical studies. It may demonstrate higher accuracy and a shorter response than Gemini. Therefore, further research is needed to maximize ChatGPT's accuracy compared to Gemini in the medical field.

Statements

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

FF: Investigation, Methodology, Validation, Visualization, Writing – review & editing. AbS: Conceptualization, Investigation, Methodology, Resources, Writing – review & editing. AmS: Conceptualization, Formal Analysis, Methodology, Resources, Validation, Writing – review & editing. SKA: Formal Analysis, Resources, Software, Supervision, Writing – review & editing. AG: Conceptualization, Formal Analysis, Methodology, Resources, Validation, Writing – review & editing. RB: Data curation, Formal Analysis, Investigation, Methodology, Validation, Visualization, Writing – review & editing. BA: Conceptualization, Data curation, Visualization, Writing – original draft, Writing – review & editing. SO: Conceptualization, Methodology, Project administration, Visualization, Writing – review & editing. SMA: Conceptualization, Formal Analysis, Methodology, Visualization, Writing – review & editing. SH: Formal Analysis, Methodology, Resources, Validation, Writing – review & editing. YM: Data curation, Formal Analysis, Software, Supervision, Visualization, Writing – review & editing. FK: Conceptualization, Data curation, Validation, Visualization, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1.

    HiwaDSAbdallaSSMuhialdeenASHamasalihHMKarimSO. Assessment of nursing skill and knowledge of ChatGPT, Gemini, Microsoft Copilot, and Llama: a comparative study. Barw Med J. (2024) 2(2):36. 10.58742/bmj.v2i2.87

  • 2.

    AbbasYNMahmoodYMHassanHAHamadDQHasanSJOmerDAet alRole of ChatGPT and google bard in the diagnosis of psychiatric disorders: a cross sectional study. Barw Med J. (2023) 1(4):149. 10.58742/4vd6h741

  • 3.

    XuLSandersLLiKChowJC. Chatbot for health care and oncology applications using artificial intelligence and machine learning: systematic review. JMIR Cancer. (2021) 7(4):e27850. 10.2196/27850

  • 4.

    FuchsATrachselTWeigerREggmannF. ChatGPT’s performance in dentistry and allergy immunology assessments: a comparative study. Swiss Dent J. (2024) 134(2):17. 10.61872/sdj-2024-06-01

  • 5.

    MasalkhiMOngJWaisbergELeeAG. Google DeepMind’s Gemini AI versus ChatGPT: a comparative analysis in ophthalmology. Eye. (2024) 14:16. 10.1038/s41433-024-02958-w

  • 6.

    SalihAMMohammedNAMahmoodYMHasanSJNamiqHSGhafourAKet alChatGPT insight and opinion regarding the controversies in neurogenic thoracic outlet syndrome; a case based-study. Barw Med J. (2023) 1(3):24. 10.58742/bmj.v1i2.48

  • 7.

    WeiQYaoZCuiYWeiBJinZXuX. Evaluation of ChatGPT-generated medical responses: a systematic review and meta-analysis. J Biomed Inform. (2024) 8:104620. 10.1016/j.jbi.2024.104620

  • 8.

    FarhudDDZokaeiS. Ethical issues of artificial intelligence in medicine and healthcare. Iran J Public Health. (2021) 50(11):iv. 10.18502/ijph.v50i11.7600

  • 9.

    RayPPMajumderP. The potential of ChatGPT to transform healthcare and address ethical challenges in artificial intelligence-driven medicine. J Clin Neurol. (2023) 19(5):509. 10.3988/jcn.2023.0158

  • 10.

    SharmaDKaushalSKumarHGainderS. Chatbots in healthcare: challenges, technologies and applications. 2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST). IEEE (2022). p. 16

  • 11.

    AbdullahHOAbdallaBAKakamadFHAhmedJOBabaHOHassanMNet alPredatory publishing lists: a review on the ongoing battle against fraudulent actions. Barw Med J. (2024) 2(2):2630. 10.58742/bmj.v2i2.91

  • 12.

    PatilNSHuangRSvan der PolCBLarocqueN. Comparative performance of ChatGPT and bard in a text-based radiology knowledge assessment. Can Assoc Radiol J. (2024) 75(2):34450. 10.1177/08465371231193716

  • 13.

    KumariAKumariASinghASinghSKJuhiAKumarAet alLarge language models in hematology case solving: a comparative study of ChatGPT-3.5, Google Bard, and Microsoft Bing. Cureus. (2023) 21:e43861. 10.7759/cureus.43861

  • 14.

    DhanvijayAKPinjarMJDhokaneNSorteSRKumariAMondalH. Performance of large language models (ChatGPT, Bing Search, and Google Bard) in solving case vignettes in physiology. Cureus. (2023) 15(8):e42972. 10.7759/cureus.42972

  • 15.

    MuhialdeenASMohammedSAAhmedNHAhmedSFHassanWNAsaadHRet alArtificial intelligence in medicine: a comparative study of ChatGPT and google bard in clinical diagnostics. Barw Med J. (2023) 6:713. 10.58742/pry94q89

  • 16.

    KogaSMartinNBDicksonDW. Evaluating the performance of large language models: ChatGPT and Google Bard in generating differential diagnoses in clinicopathological conferences of neurodegenerative disorders. Brain Pathol. (2024) 34(3):e13207. 10.1111/bpa.13207

  • 17.

    LimZWPushpanathanKYewSMLaiYSunCHLamJSet alBenchmarking large language models’ performances for myopia care: a comparative analysis of ChatGPT-3.5, ChatGPT-4.0, and google bard. EBioMedicine. (2023) 95:104770. 10.1016/j.ebiom.2023.104770

  • 18.

    IlgazHBÇelikZ. The significance of artificial intelligence platforms in anatomy education: an experience with ChatGPT and Google Bard. Cureus. (2023) 15(9):e45301. 10.7759/cureus.45301

  • 19.

    SethILimBXieYCevikJRozenWMRossRJet alComparing the efficacy of large language models ChatGPT, BARD, and Bing AI in providing information on rhinoplasty: an observational study. Aesthet Surg J Open Forum. (2023) 5:ojad084. 10.1093/asjof/ojad084; US: Oxford University Press.

  • 20.

    SalazarGZZúñigaDVindelCLYoongAMHincapieSZúñigaABet alEfficacy of AI chats to determine an emergency: a comparison between OpenAI’s ChatGPT, Google Bard, and Microsoft Bing AI chat. Cureus. (2023) 15(9):e45473. 10.7759/cureus.45473

  • 21.

    QarajehATangpanithandeeSThongprayoonCSuppadungsukSKrisanapanPAiumtrakulNet alAI-powered renal diet support: performance of ChatGPT, Bard AI, and Bing chat. Clin Pract. (2023) 13(5):116072. 10.3390/clinpract13050104

  • 22.

    ToyamaYHarigaiAAbeMNaganoMKawabataMSekiYet alPerformance evaluation of ChatGPT, GPT-4, and bard on the official board examination of the Japan radiology society. Jpn J Radiol. (2024) 42(2):2017. 10.1007/s11604-023-01491-2

  • 23.

    SinhaRKRoyADKumarNMondalH. Applicability of ChatGPT in assisting to solve higher order problems in pathology. Cureus. (2023) 15(2):e35237. 10.7759/cureus.35237

  • 24.

    DasDKumarNLongjamLASinhaRRoyADMondalHet alAssessing the capability of ChatGPT in answering first-and second-order knowledge questions on microbiology as per competency-based medical education curriculum. Cureus. (2023) 15(3):e36034. 10.7759/cureus.36034

  • 25.

    GhoshABirA. Evaluating ChatGPT’s ability to solve higher-order questions on the competency-based medical education curriculum in medical biochemistry. Cureus. (2023) 15(4):e37023. 10.7759/cureus.37023

  • 26.

    GudisDAMcCoulEDMarinoMJPatelZM. Avoiding bias in artificial intelligence. Int Forum Allergy Rhinol. (2023) 13(3):1935. 10.1002/alr.23129

Summary

Keywords

ChatGPT, Google Bard, medical inquiries, comparison, madical AI

Citation

Fattah FH, Salih AM, Salih AM, Asaad SK, Ghafour AK, Bapir R, Abdalla BA, Othman S, Ahmed SM, Hasan SJ, Mahmood YM and Kakamad FH (2025) Comparative analysis of ChatGPT and Gemini (Bard) in medical inquiry: a scoping review. Front. Digit. Health 7:1482712. doi: 10.3389/fdgth.2025.1482712

Received

21 August 2024

Accepted

21 January 2025

Published

03 February 2025

Volume

7 - 2025

Edited by

Hosna Salmani, Iran University of Medical Sciences, Iran

Reviewed by

Larry R. Price, Texas State University, United States

Thomas F. Heston, University of Washington, United States

Updates

Copyright

*Correspondence: Fahmi H. Kakamad

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics