Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med., 18 March 2025

Sec. Healthcare Professions Education

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1534294

This article is part of the Research TopicArtificial Intelligence Applications in Chronic Ocular Diseases, Volume IIView all 38 articles

Enhancing ophthalmology students’ awareness of retinitis pigmentosa: assessing the efficacy of ChatGPT in AI-assisted teaching of rare diseases—a quasi-experimental study

  • 1Department of Ophthalmology, Chongqing Key Laboratory for the Prevention and Treatment of Major Blinding Eye Diseases, The First Affiliated Hospital of Chongqing Medical University, Chongqing, China
  • 2The First College of Clinical Medicine, Chongqing Medical University, Chongqing, China

Background: Retinitis pigmentosa (RP) is a rare retinal dystrophy often underrepresented in ophthalmology education. Despite advancements in diagnostics and treatments like gene therapy, RP knowledge gaps persist. This study assesses the efficacy of AI-assisted teaching using ChatGPT compared to traditional methods in educating students about RP.

Methods: A quasi-experimental study was conducted with 142 medical students randomly assigned to control (traditional review materials) and ChatGPT groups. Both groups attended a lecture on RP and completed pre- and post-tests. Statistical analyses compared learning outcomes, review times, and response accuracy.

Results: Both groups significantly improved in post-test scores (p < 0.001), but the ChatGPT group required less review time (24.29 ± 12.62 vs. 42.54 ± 20.43 min, p < 0.0001). The ChatGPT group also performed better on complex questions regarding advanced RP treatments, demonstrating AI’s potential to deliver accurate and current information efficiently.

Conclusion: ChatGPT enhances learning efficiency and comprehension of rare diseases like RP. A hybrid educational model combining AI with traditional methods can address knowledge gaps, offering a promising approach for modern medical education.

Background

Retinitis pigmentosa (RP) is a group of inherited retinal dystrophies that leads to the progressive degeneration of photoreceptor cells, resulting in vision loss and eventual blindness. The prevalence of RP is estimated to affect approximately 1 in 4,000 individuals globally, with variations depending on geographical and ethnic populations (1). In China, in 2018, the National Health Commission included RP in the “First Batch of Rare Diseases Catalogue.” However, RP is rarely covered in the training of undergraduate and graduate ophthalmology students, and even in the training of resident physicians, which leads to a general lack of awareness among ophthalmologists in China. Over the years, advancements in molecular diagnostics have allowed for a better understanding of the genetic mutations driving RP. Techniques such as gene sequencing, genome-wide association studies (GWAS), and next-generation sequencing have made it possible to identify causative genes, leading to more precise diagnostic capabilities (24). Emerging treatments include gene therapy, optogenetics, and stem cell therapy, all of which are showing promise in clinical trials (5). However, these advancements are not reflected in traditional textbooks, resulting in a significant portion of RP patients being misdiagnosed or informed that there are no treatment options.

To change this situation, we have been committed to educating ophthalmology students about RP, primarily through traditional lectures using PowerPoint presentations. In recent years, the integration of artificial intelligence (AI), particularly large language models (LLMs), has rapidly evolved, bringing transformative changes to medicine (610). One of the most promising applications of AI is its ability to assist in medical diagnostics, education, and clinical decision-making (1113). The development of these models represents a shift toward more accessible, real-time, and personalized medical solutions (1416). In the ophthalmology domain, LLMs are becoming increasingly essential, primarily because the ability to process large volumes of medical literature and clinical data has exceeded human capacity. Studies have demonstrated that ChatGPT, along with other AI systems, performs well in responding to ophthalmology-related questions (17, 18). For instance, in a study comparing ChatGPT with ophthalmology board exam trainees, the AI model’s responses were rated highly in terms of thematic accuracy, especially for more difficult questions (19).

These advancements underscore the potential role of AI in bridging gaps in knowledge delivery, particularly for medical students and professionals undergoing rigorous training. Research on nurses highlights the potential of artificial intelligence to provide diverse perspectives. However, educators are advised to integrate AI tools with strategies that promote critical thinking, careful data evaluation, and source verification. This suggests a hybrid educational approach, blending AI with traditional methods, to strengthen nursing students’ clinical reasoning and decision-making abilities (20).

This study explores the combined teaching of traditional methods and ChatGPT in teaching rare diseases like retinitis pigmentosa, to determine if combined teaching is superior to traditional methods alone, and to explore the current role of LLMs in ophthalmology, focusing on their accuracy, usability, and future potential in medical education and clinical practice.

Methods

Design and setting

This study employed a quasi-experimental design, including experimental and control groups, to measure learning outcomes through pre- and post-tests. All students first attended a single traditional lecture focused on retinitis pigmentosa. Following the lecture, students were randomly assigned to two groups: control group used ophthamology test books and internet for review, while the ChatGPT group used only ChatGPT for review. To ensure the independence of the trial and the anonymity of students, other than the lecturer delivering the large class, no other instructors were involved in the grouping or subsequent review processes.

Participants

We assigned students with odd ID numbers to the control group and those with even ID numbers to the ChatGPT group. Both groups consisted of third-year medical students who had not yet attended specific courses on retinitis pigmentosa (RP) or engaged in related practical experiences. Consequently, the participants in both groups had comparable levels of prior knowledge and academic performance at the start of the study.

The recruitment process was conducted by a research assistant not involved in teaching the course. Before the course began, the assistant introduced the study and obtained informed consent from the students who agreed to participate. Inclusion criteria included: (1) medical students who had participated in ophthalmology internships; (2) students who had not attended a lecture on retinitis pigmentosa or studied it before; (3) students with no prior case study experience using AI tools. There were no exclusion criteria.

Interventions

All students first attended a standardized lecture on retinitis pigmentosa. Following this, both groups received identical review frameworks specifying four core competencies: (1) pathophysiology, (2) clinical manifestations, (3) diagnostic evaluations (including auxiliary examinations), and (4) therapeutic strategies. The ChatGPT group was directed to conduct topic-driven inquiries aligned with these domains, whereas the control group utilized prescribed resources: Authoritative textbooks (Yang P, Fan X, eds. *Ophthalmology*. 9th ed. Beijing: People’s Medical Publishing House; 2023. ISBN: 978-7-117-26667-3; Wang N, ed. *Ophthalmology*. Beijing: Higher Education Press; 2021. ISBN: 9787040561548) and search engines (Baidu, Bing). No prior training on AI utilization or information retrieval techniques was implemented, replicating authentic self-directed learning scenarios. Participants individually documented their review time. To eliminate confounding factors, collaborative learning was prohibited during the review phase, and outcome assessment relied exclusively on the second examination performance.

Exams

Students took two exams: a pre-test immediately after the lecture to assess their baseline knowledge levels and a post-test after the review period to evaluate the impact of the review methods on learning outcomes. The second examination was administered 24 h after the course concluded. The exam content was designed by a team of experts in retinal pigment degeneration to ensure accuracy and relevance.

Ethics

The institutional review board of the First Affiliated Hospital of Chongqing Medical University approved this study (IRB No. 2024-162-01). Participants were thoroughly informed about the study’s objectives, voluntary nature, confidentiality measures, identity protection, and procedural details during the recruitment process.

Statistical data

Statistical analyses were conducted using SPSS (version 25.0; IBM Inc., Armonk, NY, USA). Independent t-tests compared group differences in scores, times, study times, and correctness, while paired t-tests assessed within-group improvements between attempts. Item-level correctness rates were evaluated both within groups across attempts and between groups for the second attempt. The expected effect size was moderate (Cohen’s d = 0.5), with a significance level of 0.05 and a test power of 0.8, requiring at least approximately 64 participants per group, thus a total of 128 students. This sample size ensures sufficient statistical power to detect differences in learning outcomes between the experimental and control groups.

Results

The study included 142 medical students, with 70 participants in the control group (47.1% male) and 72 in the ChatGPT-assisted group (54.2% male). A 27-item assessment (total score: 135 points; see Supplementary material 1) demonstrated comparable baseline knowledge between groups, with pre-test scores of 87.54 ± 28.27 (control) versus 87.71 ± 24.05 (ChatGPT) (p = 0.97). Post-intervention evaluations revealed significant improvement in both groups (control: 110.93 ± 29.01; ChatGPT: 115.42 ± 22.70; both p < 0.0001), though the intergroup difference remained statistically nonsignificant (p = 0.28).

Notably, temporal efficiency metrics diverged between cohorts. The control group reduced their test completion time by 50% (pre-test: 670.20 ± 448.82 s vs. post-test: 334.19 ± 243.10 s; p < 0.0001), while the ChatGPT group achieved a 37% reduction (490.57 ± 319.93 s vs. 310.67 ± 233.42 s; p < 0.0001). Despite the control group’s initially prolonged pre-test duration (Δ = 179.63 s; p = 0.0067), post-test completion times converged between groups (Δ = 23.52 s; p = 0.56).

A critical distinction emerged in learning time investment. Students utilizing ChatGPT required 43% less review time to achieve comparable outcomes (24.29 ± 12.62 min vs. 42.54 ± 20.43 min; p < 0.0001). Domain-specific analysis (Table 1) identified differential competency development: both groups showed improved accuracy in foundational knowledge (genetic patterns: Q2-3; diagnostic criteria: Q19), but the ChatGPT cohort demonstrated superior performance in clinical reasoning tasks, particularly therapeutic decision-making (Q11: 73% vs. 61%; p = 0.032) and diagnostic integration (Q24: 79% vs. 64%; p = 0.013).

Table 1
www.frontiersin.org

Table 1. Comparison of the accuracy rates for different questions between the control group and the ChatGPT group in two tests.

Persistent challenges were observed in electroretinogram interpretation (Q18 accuracy <65%) and advanced pathophysiology synthesis (Q16/Q23). Crucially, while factual recall performance was equivalent (Q1-5/Q8-10; p > 0.05), the ChatGPT group exhibited enhanced capacity for multidisciplinary knowledge integration (Q27: 84% vs. 72%; p = 0.022).

In synthesizing these outcomes, three salient patterns emerge: (1) AI-enhanced learning produced equivalent knowledge retention with substantially reduced time commitment, (2) the intervention cohort demonstrated unique advantages in synthesizing complex clinical concepts despite equivalent baseline capabilities, and (3) conventional assessment tools failed to fully capture the pedagogical benefits of AI integration, particularly in dynamic clinical reasoning domains.

Discussion

This quasi-experimental study demonstrates that AI-assisted learning platforms can alleviate three persistent challenges in rare disease education: (1) missing specialized knowledge in standard textbooks, (2) slow development of clinical decision-making skills, and (3) time-consuming traditional training methods. Our data reveal that ChatGPT-enhanced instruction not only compensates for the absence of ultra-specialized RP knowledge in standard curricula (e.g., optogenetic therapy timing, OCT diagnostic integration) but also accelerates the development of higher-order clinical competencies. Notably, the ChatGPT group achieved equivalent knowledge retention with 43% reduced preparation time, while outperforming controls in therapeutic decision-making and multidisciplinary knowledge synthesis. These advantages were most pronounced in scenarios requiring dynamic integration of genetic mechanisms, diagnostic findings, and therapeutic strategies—precisely the cognitive domains where traditional RP pedagogy underperforms. The findings suggest AI’s unique capacity to bypass the “clinical exposure bottleneck” inherent to rare disease training, providing learners with simulated expertise-building experiences that conventional resources cannot replicate. This pedagogical shift may prove particularly transformative for diseases like RP, where delayed diagnosis remains prevalent due to clinicians’ limited pattern recognition experience.

Retinitis pigmentosa presents specific challenges in education due to its complexity and rarity. Most ophthalmologists receive limited exposure to RP during their formal education and clinical practice. Traditional educational materials often fail to include the latest advancements in diagnosis and treatment, such as gene therapy and optogenetics, which are crucial for managing such diseases. Compared to the limited content in textbooks, the complexity of RP, along with its genetic and clinical heterogeneity, requires broader educational attention. Our study found significant improvements in all students’ scores from the first to the second exam, demonstrating that both traditional and AI-assisted methods effectively enhance student learning. This indicates that both traditional methods and ChatGPT provided a solid foundation of established knowledge (20).

The comparable post-test performance between groups may stem from two synergistic factors. First, standardized baseline knowledge was ensured through pre-allocation lectures covering RP fundamentals (genetic mechanisms, classic phenotypes), as evidenced by equivalent pre-test scores (Control: 87.54 ± 28.27 vs. ChatGPT: 87.71 ± 24.05, p = 0.97). Second, the assessment’s focus on factual recall (Q1-Q15 testing definitions/pathophysiology) rather than applied clinical reasoning likely obscured ChatGPT’s strengths, given textbooks’ persistent reliability for memorizing established facts despite therapeutic inaccuracies. Nevertheless, the ChatGPT group demonstrated superiority in clinical judgment tasks, particularly on Q11 regarding optogenetic therapy timing (79% vs. 67% accuracy). This discrepancy reflects two critical issues: (1) control students’ dependence on outdated textbooks stating “no available treatment” despite lecture updates, and (2) exposure to unverified medical advertisements during independent online searches (e.g., Baidu), which propagated misconceptions. In contrast, ChatGPT’s capacity to synthesize real-time therapeutic evidence enabled accurate, context-aware responses while filtering commercial misinformation (6, 11, 14, 17, 21, 22). These findings corroborate emerging evidence that LLMs enhance ophthalmology education through three key mechanisms: dynamic knowledge updating in fast-evolving fields (e.g., genetics), multi-perspective clinical scenario analysis, and automated credibility screening of health information (14, 17, 19). Furthermore, the persistent challenges in ERG interpretation (Q18) underscore the limitations of text-based learning modalities for visual diagnostic competencies. Future assessments incorporating clinical vignettes could better elucidate AI’s potential in cultivating higher-order competencies like genetic report interpretation.

Our study also found that although the review materials for the control group included both Chinese and English textbooks and were not extensive, students still spent a considerable amount of time reviewing them. Compared to the experimental group using ChatGPT, the review time was significantly longer. The shorter review time for the ChatGPT group did not compromise the quality of learning, as evidenced by their comparable exam scores to the Control group. This highlights ChatGPT’s effectiveness in improving learning efficiency (2325).

Our findings contribute to the ongoing debate about lecture-based pedagogy in medical education. While traditional lectures effectively establish baseline knowledge (evidenced by equivalent pre-test scores), they show limited capacity to cultivate clinical reasoning for rare diseases. This supports a blended model where AI tools augment – rather than replace – didactic teaching, particularly for maintaining content currency in rapidly evolving fields like gene therapy.

Past literatures reported that that LLMs like ChatGPT are primarily trained on English data, which may limit their application in non-English-speaking regions (26, 27). In response to this challenge, efforts have been made to develop specific language models, such as the Chinese Ophthalmology Model (MOPH) customized for Chinese medical professionals (26). However, in our study, students posed questions to ChatGPT in Chinese and received answers in Chinese. Our evaluation found these responses to be highly accurate, showcasing ChatGPT’s robust capability in handling non-English queries. This finding is significant for demonstrating the viability and effectiveness of AI technology in multilingual environments, especially for its potential application in global medical education and practice. This adaptability of AI models allows them to better serve users in different languages, providing more accurate information and thereby enhancing the quality of global medical education and clinical practice.

Despite the potential of AI to enhance educational outcomes, concerns remain about the accuracy of AI-generated responses. AI models can rapidly respond to medical queries but are sometimes prone to errors or “hallucinations,” where incorrect or misleading information is presented as fact (6, 11). This is particularly concerning in medical settings, where accuracy is crucial. However, in our study, the experimental group using ChatGPT did not encounter issues with “hallucinations” or misleading information. Through expert review, ChatGPT demonstrated high accuracy in answering questions about Retinitis Pigmentosa. This indicates that when AI tools are properly supervised and validated, they can serve as a reliable resource for medical education (2830), especially in complex and evolving medical fields such as genetic eye diseases.

These findings support the adoption of a hybrid educational model that combines AI tools like ChatGPT with traditional teaching methods. This model not only ensures that students access the most current information but also maximizes educational effectiveness by leveraging the strengths of both AI and traditional resources. Future educational strategies should focus on developing AI tools that support these hybrid models, ensuring they are technologically advanced yet educationally sound. Educators should also emphasize the importance of critical thinking and ethical considerations, especially in the use of AI, to prepare students for the complexities of modern medical practice.

Our study has several limitations that warrant consideration. First, the assessment’s emphasis on recall-based questions (Q1-Q15) may have underestimated ChatGPT’s capacity to enhance clinical decision-making skills, as its advantages emerged primarily in complex clinical questions (Q11/Q24: 12–11% accuracy gap, p < 0.05). Second, the homogeneous sample of third-year medical students from a single institution limits generalizability to clinicians or diverse training stages, though baseline equivalence supports internal validity. Third, the single-session intervention (24.29-min ChatGPT review) precludes conclusions about long-term retention of RP-specific competencies like ERG interpretation (Q18). Fourth, observed benefits may be disease-specific, given greater AI performance gains in RP-focused questions (Q6/Q14/Q27) compared to general ophthalmology items. The unanticipated pre-test time imbalance between groups suggests future studies should implement stratified randomization based on digital literacy and include brief surveys to understand students’ exam preparation methods.

Moving forward, we plan to extend this research in three key areas. First, we’ll test the approach with different medical professionals, including doctors in community hospitals and eye specialists-in-training, to see if the AI benefits hold across experience levels (current sample: 142 students). Second, we are redesigning tests to include more real-world tasks – like analyzing eye scan images (Q24) and treatment planning – which better match how AI assists in clinical environments or scenarios. Third, we’ll track how well the knowledge long term retention using mobile tools, especially for complex clinical decision-making scenarios like deciding when to use new therapies. For rare diseases beyond RP (where our early data shows bigger gaps – Q14 baseline errors up to 32%), we are exploring AI tools that help doctors quickly find reliable treatment options. At the same time, we are studying how to best combine AI with expert guidance to prevent critical errors. These steps aim to create practical tools that save time (43% faster learning in our study) while keeping critical information up-to-date, especially for clinics without easy access to specialists.

Conclusion

In conclusion, integrating AI technology, particularly ChatGPT, into the education of complex and rare diseases such as Retinitis Pigmentosa, offers a promising pathway. Our study underscores the need for educational systems to adapt to technological advancements, providing students with a comprehensive, current, and clinically relevant learning experience. As AI technology continues to evolve, its potential to enhance learning outcomes and medical practice appears both promising and extensive. Future research should continue to explore and refine the use of AI in education, ensuring it aligns with educational goals and the realities of medical practice.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the institutional review board of the First Affiliated Hospital of Chongqing Medical University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

JZ: Conceptualization, Methodology, Writing – original draft, Writing – review & editing. KS: Formal analysis, Investigation, Writing – original draft, Writing – review & editing. PQ: Data curation, Writing – original draft, Writing – review & editing. SL: Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Acknowledgments

The authors thank staff members from the First Affiliated Hospital of Chongqing Medical University, for their assistance with information collection.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Gen AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2025.1534294/full#supplementary-material

References

1. Hartong, D, Berson, E, and Dryja, T. Retinitis pigmentosa. Lancet. (2006) 368:1795–809. doi: 10.1016/S0140-6736(06)69740-7

PubMed Abstract | Crossref Full Text | Google Scholar

2. Confalonieri, F, La Rosa, A, Ottonelli, G, Barone, G, Ferraro, V, Di Maria, A, et al. Retinitis pigmentosa and therapeutic approaches: a systematic review. J Clin Med. (2024) 13:4680. doi: 10.3390/jcm13164680

PubMed Abstract | Crossref Full Text | Google Scholar

3. Georgiou, M, Robson, AG, Fujinami, K, de Guimarães, TAC, Fujinami-Yokokawa, Y, Daich Varela, M, et al. Phenotyping and genotyping inherited retinal diseases: molecular genetics, clinical and imaging features, and therapeutics of macular dystrophies, cone and cone-rod dystrophies, rod-cone dystrophies, Leber congenital amaurosis, and cone dysfunction syndromes. Prog Retin Eye Res. (2024) 100:101244. doi: 10.1016/j.preteyeres.2024.101244

Crossref Full Text | Google Scholar

4. Liu, W, Liu, S, Li, P, and Yao, K. Retinitis pigmentosa: Progress in molecular pathology and biotherapeutical strategies. Int J Mol Sci. (2022) 23:4883. doi: 10.3390/ijms23094883

PubMed Abstract | Crossref Full Text | Google Scholar

5. Becherucci, V, Bacci, GM, Marziali, E, Sodi, A, Bambi, F, and Caputo, R. The new era of therapeutic strategies for the treatment of retinitis pigmentosa: a narrative review of pathomolecular mechanisms for the development of cell-based therapies. Biomedicines. (2023) 11:2656. doi: 10.3390/biomedicines11102656

PubMed Abstract | Crossref Full Text | Google Scholar

6. Jowsey, T, Stokes-Parish, J, Singleton, R, and Todorovic, M. Medical education empowered by generative artificial intelligence large language models. Trends Mol Med. (2023) 29:971–3. doi: 10.1016/j.molmed.2023.08.012

PubMed Abstract | Crossref Full Text | Google Scholar

7. Wang, RY, Zhu, SY, Hu, XY, Sun, L, Zhang, SC, and Yang, WH. Artificial intelligence applications in ophthalmic optical coherence tomography: a 12-year bibliometric analysis. Int J Ophthalmol. (2024) 17:2295–307. doi: 10.18240/ijo.2024.12.19

PubMed Abstract | Crossref Full Text | Google Scholar

8. Jagannath, P, Perera, N, Minton, L, Bass, R, Milner, D, and Perchik, J. Increasing accessibility: effectiveness of a remote artificial intelligence education curriculum for international medical graduates. Clin Teach. (2025) 22:e70047. doi: 10.1111/tct.70047

PubMed Abstract | Crossref Full Text | Google Scholar

9. Zong, H, Wu, R, Cha, J, Wang, J, Wu, E, Li, J, et al. Large language models in worldwide medical exams: platform development and comprehensive analysis. J Med Internet Res. (2024) 26:e66114. doi: 10.2196/66114

PubMed Abstract | Crossref Full Text | Google Scholar

10. Liu, Z, Zhang, L, Wu, Z, Yu, X, Cao, C, Dai, H, et al. Surviving ChatGPT in healthcare. Front Radiol. (2023) 3:1224682. doi: 10.3389/fradi.2023.1224682

PubMed Abstract | Crossref Full Text | Google Scholar

11. Xu, T, Weng, H, Liu, F, Yang, L, Luo, Y, Ding, Z, et al. Current status of ChatGPT use in medical education: potentials, challenges, and strategies. J Med Internet Res. (2024) 26:e57896. doi: 10.2196/57896

PubMed Abstract | Crossref Full Text | Google Scholar

12. Hanycz, SA, and Antiperovitch, P. A practical review of generative AI in cardiac electrophysiology medical education. J Electrocardiol. (2025) 90:153903. doi: 10.1016/j.jelectrocard.2025.153903

PubMed Abstract | Crossref Full Text | Google Scholar

13. Xu, Y, and Yang, W. Editorial: artificial intelligence applications in chronic ocular diseases. Front Cell Dev Biol. (2023) 11:1295850. doi: 10.3389/fcell.2023.1295850

PubMed Abstract | Crossref Full Text | Google Scholar

14. Mihalache, A, Grad, J, Patil, NS, Huang, RS, Popovic, MM, Mallipatna, A, et al. Google Gemini and bard artificial intelligence chatbot performance in ophthalmology knowledge assessment. Eye. (2024) 38:2530–5. doi: 10.1038/s41433-024-03067-4

PubMed Abstract | Crossref Full Text | Google Scholar

15. Ji, YK, Hua, RR, Liu, S, Xie, CJ, Zhang, SC, and Yang, WH. Intelligent diagnosis of retinal vein occlusion based on color fundus photographs. Int J Ophthalmol. (2024) 17:1–6. doi: 10.18240/ijo.2024.01.01

PubMed Abstract | Crossref Full Text | Google Scholar

16. Zapata-Rivera, D, Torre, I, Lee, CS, Sarasa-Cabezuelo, A, Ghergulescu, I, and Libbrecht, P. Editorial: generative AI in education. Front Artif Intell. (2024) 7:1532896. doi: 10.3389/frai.2024.1532896

PubMed Abstract | Crossref Full Text | Google Scholar

17. Bahir, D, Zur, O, Attal, L, Nujeidat, Z, Knaanie, A, Pikkel, J, et al. Gemini AI vs. ChatGPT: a comprehensive examination alongside ophthalmology residents in medical knowledge. Graefes Arch Clin Exp Ophthalmol. (2024) 263:527–36. doi: 10.1007/s00417-024-06625-4

PubMed Abstract | Crossref Full Text | Google Scholar

18. Yang, WH, Shao, Y, and Xu, YW. Guidelines on clinical research evaluation of artificial intelligence in ophthalmology (2023). Int J Ophthalmol. (2023) 16:1361–72. doi: 10.18240/ijo.2023.09.02

PubMed Abstract | Crossref Full Text | Google Scholar

19. Chang, LC, Sun, CC, Chen, TH, Tsai, DC, Lin, HL, and Liao, LL. Evaluation of the quality and readability of ChatGPT responses to frequently asked questions about myopia in traditional Chinese language. Digit Health. (2024) 10:20552076241277021. doi: 10.1177/20552076241277021

PubMed Abstract | Crossref Full Text | Google Scholar

20. Shin, H, De Gagne, JC, Kim, SS, and Hong, M. The impact of artificial intelligence-assisted learning on nursing Students' ethical decision-making and clinical reasoning in pediatric care: a quasi-experimental study. Comput Inform Nurs. (2024) 42:704–11. doi: 10.1097/CIN.0000000000001177

PubMed Abstract | Crossref Full Text | Google Scholar

21. Strzalkowski, P, Strzalkowska, A, Chhablani, J, Pfau, K, Errera, MH, Roth, M, et al. Evaluation of the accuracy and readability of ChatGPT-4 and Google Gemini in providing information on retinal detachment: a multicenter expert comparative study. Int J Retina Vitreous. (2024) 10:61. doi: 10.1186/s40942-024-00579-9

PubMed Abstract | Crossref Full Text | Google Scholar

22. Mahajan, A, Tran, A, Tseng, ES, Como, JJ, El-Hayek, KM, Ladha, P, et al. Performance of trauma-trained large language models on surgical assessment questions: a new approach in resource identification. Surgery. (2025) 179:108793. doi: 10.1016/j.surg.2024.08.026

PubMed Abstract | Crossref Full Text | Google Scholar

23. Scaioli, G, Lo Moro, G, Conrado, F, Rosset, L, Bert, F, and Siliquini, R. Exploring the potential of ChatGPT for clinical reasoning and decision-making: a cross-sectional study on the Italian medical residency exam. Ann Ist Super Sanita. (2023) 59:267–70. doi: 10.4415/ANN_23_04_05

PubMed Abstract | Crossref Full Text | Google Scholar

24. Tian, D, Jiang, S, Zhang, L, Lu, X, and Xu, Y. The role of large language models in medical image processing: a narrative review. Quant Imaging Med Surg. (2024) 14:1108–21. doi: 10.21037/qims-23-892

PubMed Abstract | Crossref Full Text | Google Scholar

25. Gong, D, Li, WT, Li, XM, Wan, C, Zhou, YJ, Wang, SJ, et al. Development and research status of intelligent ophthalmology in China. Int J Ophthalmol. (2024) 17:2308–15. doi: 10.18240/ijo.2024.12.20

PubMed Abstract | Crossref Full Text | Google Scholar

26. Zheng, C, Ye, H, Guo, J, Yang, J, Fei, P, Yuan, Y, et al. Development and evaluation of a large language model of ophthalmology in Chinese. Br J Ophthalmol. (2024) 108:1390–7. doi: 10.1136/bjo-2023-324526

PubMed Abstract | Crossref Full Text | Google Scholar

27. Kai, I, Naoya, A, and Kiyotaka, F. Analysis of responses of GPT-4 V to the Japanese National Clinical Engineer Licensing Examination. J Med Syst. (2024) 48:83. doi: 10.1007/s10916-024-02103-w

PubMed Abstract | Crossref Full Text | Google Scholar

28. Wang, C, Xiao, C, Zhang, X, Zhu, Y, Chen, X, Li, Y, et al. Exploring medical students' intention to use of ChatGPT from a programming course: a grounded theory study in China. BMC Med Educ. (2025) 25:209. doi: 10.1186/s12909-025-06807-6

PubMed Abstract | Crossref Full Text | Google Scholar

29. Wu, Y, Zheng, Y, Feng, B, Yang, Y, Kang, K, and Zhao, A. Embracing ChatGPT for medical education: exploring its impact on doctors and medical students. JMIR Med Educ. (2024) 10:e52483. doi: 10.2196/52483

PubMed Abstract | Crossref Full Text | Google Scholar

30. Öncü, S, Torun, F, and Ülkü, HH. AI-powered standardised patients: evaluating ChatGPT-4o's impact on clinical case management in intern physicians. BMC Med Educ. (2025) 25:278. doi: 10.1186/s12909-025-06877-6

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: retinitis pigmentosa, ChatGPT, ophthalmology, education, AI

Citation: Zeng J, Sun K, Qin P and Liu S (2025) Enhancing ophthalmology students’ awareness of retinitis pigmentosa: assessing the efficacy of ChatGPT in AI-assisted teaching of rare diseases—a quasi-experimental study. Front. Med. 12:1534294. doi: 10.3389/fmed.2025.1534294

Received: 26 November 2024; Accepted: 03 March 2025;
Published: 18 March 2025.

Edited by:

Weihua Yang, Southern Medical University, China

Reviewed by:

Krisztina Valter, Australian National University, Australia
Gangjin Kang, The Affiliated Hospital of Southwest Medical University, China
Ziao Hu, Central South University, China

Copyright © 2025 Zeng, Sun, Qin and Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shulin Liu, c2h1bGlubGl1X2NtdUAxNjMuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.