- 1Department of Meridian and Acupoint, College of Korean Medicine, Kyung Hee University, Seoul, Republic of Korea
- 2KM Fundamental Research Division, Korea Institute of Oriental Medicine, Daejeon, Republic of Korea
- 3Jaseng Spine and Joint Research Institute, Jaseng Medical Foundation, Seoul, Republic of Korea
Background: The medical knowledge of GPT-4 has been evaluated on patient data, providing diagnostic and treatment suggestions. However, few studies have directly compared the clinical suggestions of GPT-4 with those of groups of practitioners.
Methods: This study assessed the ability of GPT-4 to make medical decisions regarding acupuncture treatment by comparing its selection of acupoints with those made by human clinicians. Ten case reports published in Korean medical journals were selected and put in a standardized format. The standardized patient information was given to 80 Korean Medicine doctors and GPT-4 to diagnose and prescribe three to five acupoints per case. To evaluate the performance of GPT-4, the similarities in acupoint selection between the doctors and GPT-4 were quantified based on the percentage overlap and correlations of the selection probabilities of acupoints in each case.
Results: The average percentage overlap for acupoints among cases at the 10% cutoff was 51.3%, i.e., more than half of the GPT-4 acupoint suggestions overlapped the acupoints selected by the doctors. In half of the cases, significant correlations were observed in the acupoint selection probabilities, implying that GPT-4 acupoint suggestions are similar to those of doctors.
Conclusions: GPT-4 made reasonable acupoint suggestions, with notable overlap observed with the prescriptions of doctors. This shows its promise for supporting medical decisions, education, and personalized medicine for patients undergoing acupuncture treatment. Future studies and validation are necessary to ensure the reliability and efficacy of applying GPT-4 in real-world settings.
Background
The potential of large language models (LLMs) in medicine has been demonstrated; they can contribute to medical education, research, and clinical decision-making (1). The reliability of ChatGPT, one of the first generative artificial intelligence (AI) chatbots, in medical domains has been tested since its release in 2022 by OpenAI. Its performance in medical licensing examinations in many countries and languages demonstrates its medical knowledge in both conventional and traditional East Asian medicine (TEAM) (2–5). Numerous studies have revealed its utility in medical education and as support for clinical decisions. Healthcare professionals can use ChatGPT to support clinical decisions, such as diagnosis and treatment planning, with reliable accuracy (6, 7). The benefits can be extended to education, strengthening the personalized learning experiences of students or trainees (8).
Acupuncture treatment is a clinical intervention in which the practitioner stimulates specific points on the patient's body using needles. Combinations of these points are selected according to the symptoms, diseases, and state of the patient. Selecting acupoints is a complex activity, and the selection rationale is often intrinsic; many factors affect this process, including traditional and conventional theories regarding the locations of targets as well as the experience of doctors (9, 10). With the advent of AI, data-based algorithms and models were developed to discover the rules governing acupoint selection (11, 12). Since each point has unique properties and actions, choosing the correct points is important for achieving better clinical responses (13, 14). The use of ChatGPT for suggesting acupoints has shown promise, indicating its utility in acupuncture (15).
The medical knowledge of generative AI has been evaluated using real patient cases, and it made accurate diagnoses in challenging medical cases (16). Chatbot has also suggested treatment according to patient information and provided valuable insights for tailored treatment plans (15, 17). However, few studies have directly compared the clinical suggestions of ChatGPT with those of actual practitioners. Therefore, this study assessed the practical utility of GPT-4, the latest version of ChatGPT, in making medical decisions regarding acupuncture treatment by comparing its selection of acupoints with those made by human clinicians.
Methods
Materials preparation
Ten cases were selected from case reports including acupuncture treatment published in Korean medical journals between 2014 and 2016. We selected ten cases to include a diverse range of diseases, covering musculoskeletal symptoms, psychiatric symptoms, and various symptoms treated in internal medicine. All patient information that could influence the diagnosis and treatment plan was extracted from each case. To standardize the information for virtual diagnosis, we organized the extracted information in the order of demographics, chief complaint, medical history, other symptoms and signs, and laboratory test results. The study was approved by the Institutional Review Board of Kyung Hee University, Seoul, Republic of Korea (KHSIRB-17-046).
Data collection and preprocessing
Through an online virtual diagnosis process, 80 Korean medicine doctors (KMDs) were provided with standardized patient information for each case. They were asked to identify TEAM patterns and select three to five acupoints based on the information (18). The patterns of acupoint selection by the doctors were previously published elsewhere (19). GPT-4 was given the same patient information, with the prompt “Based on the patient's information, please identify patterns in Traditional East Asian medicine practice, and suggest 3–5 acupoints for the patient”. The procedure was performed for all 10 cases in May 2024, and every prompt was run in independent chats to prevent any learning effects. The experimental process is described in Figure 1.

Figure 1. Study procedure. Ten cases were selected from case reports published in Korean medical journals. Patient information, including symptoms and signs, was standardized into a consistent format. This information was presented to 80 Korean medicine doctors, who were instructed to select three to five acupoints for each patient, and to GPT-4, which was prompted to perform the same task. KMD, average of the 80 Korean medicine doctors; GPT, average of 10 GPT-4 runs.
Data from 80 doctors and 10 GPT-4 runs were collected for each case. Acupoint data from doctors were initially recorded as free text, a pre-processing step was applied to correct typographical errors and standardize the data by excluding any acupoints not included in the 361 acupoints defined by World Health Organization (WHO) terminology. The GPT-4 data contained Korean, Chinese, and WHO terminology and often included minor incorrect terms. Therefore, we selectively included an acupoint only if at least two of the three correctly recorded terminologies referred to the same point. Each acupoint was converted to standard WHO terminology.
Data analysis
The selection probability for each acupoint was calculated for both the doctors and GPT-4 by dividing the total number of answers in each case by 80 for the doctors and by 10 for GPT-4. We identified the top 30 most frequently used acupoints and visualized their selection probabilities using a heatmap. The similarity in acupoint selection between doctors and GPT-4 was evaluated using the percentage overlap of acupoints and the correlation of selection probabilities in each case. Overlapping acupoints were identified as acupoints selected at a rate above the cutoff level by both the doctors and GPT-4. The percentage overlap at the 10% cutoff was calculated as follows:
Since the percentage overlap varies depending on the cutoff level, we calculated it at different cutoff levels ranging from 10% to 50% and then averaged the values. The resulting measure is referred to as the percentage overlap (avg.) for each case.
Spearman's correlation analysis was used to compare the selection probabilities of acupoints between doctors and GPT-4 for each case. Acupoints in each case were excluded from the correlation analyses when they were not selected by either doctors or GPT-4.
To analyze selection patterns by disease type, we categorized the ten cases into three groups: musculoskeletal symptoms (derangement of meniscus [case 4], chronic prostatitis [case 6], and intervertebral disc disorders [case 8]), psychiatric symptoms (menopausal climacteric states cases [case 3], panic disorder [case 7], and fibromyalgia [case 9]), and various symptoms in internal medicine (benign paroxysmal positional vertigo [case 1], gastroesophageal reflux disease [case 2], diabetic neuropathy [case 5], and puerperal disorder [case 10]). This categorization aligns with the clusters identified in our previous study using a hierarchical approach based on selection probabilities (19). Percentage overlap was averaged for each group.
The data analyses were performed using R software (version 4.0.2, https://cran.r-project.org), and the heatmap was generated with Orange software (version 3.36.1, https://orangedatamining.com).
Analysis of response variability
Given the substantial sample-size imbalance (10 trials of GPT-4 vs. 80 doctors responses), we compared response variability. To quantify the response variability of GPT-4 and doctors, we represented each response as a set of selected acupoints (binary, multi-label) and computed the mean of all pairwise Jaccard dissimilarities, using the following equation:
This approach directly measures pattern heterogeneity between two datasets with identical item availability. To characterize uncertainty under unequal sample sizes, we reported 95% bootstrap percentile confidence intervals (2,000 resamples). To assess the dispersion of heterogeneity, we applied PERMDISP permutation test, which embeds the Jaccard distance matrix via principal coordinates analysis and compared mean distances to the group centroid.
Results
Percentage overlap of selected acupoints between doctors and GPT-4
Table 1 shows the percentage overlap of acupoints chosen by doctors and GPT-4. The average percentage overlap between doctors and GPT-4 at the 10% cutoff among cases was 51.3%, indicating that more than half of the acupoints suggested by GPT-4 matched those selected by the doctors. The average percentage overlap across different cutoff levels was 48.9%, and the results for each cutoff level are presented in Supplementary material 1.
The percentage overlap differed by the type of disease. Generally, acupoint suggestions for musculoskeletal symptoms showed the lowest overlap, with an average of 42.3% at the 10% cutoff and 36% across various cutoff levels. In comparison, psychiatric symptoms had highest overlap, averaging 63.3% at the 10% cutoff and 51% across various cutoffs, while internal medicine symptoms showed 49% and 56%, respectively.
Selection patterns of acupoints of doctors and GPT-4
The selection probabilities of the top 30 acupoints were visualized using a heatmap (Figure 2). Acupoint selection patterns were similar between doctors and GPT-4. Major acupoints such as SP6, LR3, LI4, and ST36 were consistently recommended by both doctors and GPT-4 across the 10 different diseases. Furthermore, similar acupoint selection patterns were observed beyond these major points. For example, GB20 was particularly recommended in case 1, SP10 in case 4, and BL23 in case 8.

Figure 2. Selection probabilities of the top 30 acupoints. The selection probabilities of the top 30 most frequently used acupoints are illustrated in each of the 10 cases using a heatmap. The top row indicates the selection probabilities of the 80 doctors and the bottom row the selection probabilities of GPT-4. The color intensity in the heatmap reflects the selection probability, with darker shades indicating a higher probability of selection for a given acupoint.
Significant correlations were observed for 5 of the 10 cases, including case 3 (ρ = 0.355, p = 0.003), case 5 (ρ = 0.532, p < 0.001), case 6 (ρ = 0.263, p = 0.017), case 7 (ρ = 0.408, p < 0.001), and case 9 (ρ = 0.262, p = 0.019). These results suggest that GPT-4's acupoint selection is similar to that of doctors, as acupoints frequently chosen by doctors were also frequently selected by GPT-4, while less commonly chosen acupoints were similarly selected less often. Table 1 summarizes the acupoint selection similarities between doctors and GPT-4 for the 10 different diseases.
Comparison of response variability between doctors and GPT-4
Across all 10 cases, doctors exhibited higher response variability than GPT-4. Mean pairwise Jaccard dissimilarity was 0.570–0.711 for GPT-4 vs 0.840–0.935 for doctors (mean difference = 0.264). Size-matched bootstrap CIs for doctors –GPT-4 excluded zero in every case, and PERMDISP likewise showed greater dispersion for doctors (all p < 0.05; 0.0001–0.0029). Taken together, both metrices indicate that doctors' responses are consistently more heterogeneous than GPT-4's, independent of sample-size imbalance.
Discussion
In this study, GPT-4 showed its potential for assisting in clinical decision-making by helping to select appropriate acupoints. More than half of the acupoints suggested by GPT-4 overlapped those selected by the 80 doctors across 10 cases involving different symptoms. Moreover, significant correlations were observed in the acupoint selection patterns between doctors and GPT-4 in half of the cases. Our findings suggest that ChatGPT has the ability to suggest appropriate acupoints based on patient information and might be useful for supporting medical decisions and education regarding acupuncture practice.
We compared the acupoints suggested by dozens of doctors, and the results demonstrate how well GPT-4 can reflect real-world clinical decisions. Unlike most medical diagnoses, the selection of appropriate acupuncture points does not involve clear-cut “true” or “false” answers in most clinical settings. Therefore, we identified acupoints that were frequently suggested (those with a frequency exceeding 10%, by both GPT-4 and the doctors) and found significant overlap in these frequently suggested acupoints. Over half of the acupoints suggested by GPT-4 matched those of the doctors, with 51.3% overlap across 10 different cases. The average overlap was also similar (48.9%) when we applied different cutoff levels, indicating that GPT-4 can choose appropriate acupoints based on the clinical information of diverse patients similar to Korean Medicine doctors. Furthermore, the significant correlation of selection probabilities reveals the similarity of acupoint prescription between human doctors and GPT-4. These findings add to the growing evidence of the clinical decision-making ability of ChatGPT using patient information, such as for diagnostic reasoning and treatment planning (15, 16).
Both doctors and GPT-4 suggested major acupoints, such as SP6, LR3, LI4, and ST36, regardless of the disease. These acupoints are widely used in diverse conditions and have general neurological effects (20). One might have concerns about GPT-4 repeatedly suggesting major acupoints, as this could lead to overestimation of its performance. However, similar acupoint selection patterns were observed beyond these major points, such as GB20, SP10, and BL23, each with specific implications. For example, in the first case, GB20 was chosen to target specific symptoms and disease sites. The combination of major points for general effects and non-major points for targeted effects is commonly observed in acupuncture treatment (19). Moreover, the significant correlations in acupoint selection between doctors and GPT-4 in half of the cases further support the similarity in acupoint selection patterns exhibited by GPT-4.
The similarity of acupoint selection patterns varied by disease type. Overlap was higher, and significant correlations were more frequently observed in psychiatric disorders compared to musculoskeletal disorders. These differences may be attributed to the differing rationales behind acupoint selection depending on disease type. In musculoskeletal diseases, doctors more frequently selected local acupoints near the symptom site. While GPT-4 relies on theoretical and literature-based knowledge, doctors make decisions based on both experience and theoretical-based knowledge. Doctors might place greater reliance on their clinical experience when selecting acupoints for musculoskeletal diseases. Differences were also evident at the case level. For example, in case 1, a high percentage of doctors selected ST36 and PC6, but these were never chosen by GPT-4. Conversely, GPT-4 frequently suggested points such as BL18 and HT7, which were rarely or never recommended by doctors. Such discrepancies suggest that higher acupoint overlap does not necessarily guarantee alignment with the underlying clinical rationale of real-world clinical decisions. Caution is therefore warranted when applying language models in clinical settings.
Our study reflects the reasoning process in acupuncture treatment—where doctors integrate signs and consider multiple symptoms when selecting appropriate acupoints (9), rather than relying solely on “extrinsic” diagnoses—leads to identifiable patterns in acupoint selection. Notably, similar patterns were observed in GPT-4, suggesting that it may reflect aspects of this clinical reasoning process. Incorporating information relating to various aspects, such as subjective symptoms and individual lifestyles, could enhance personalized medicine (21). In future studies, this capability can be further leveraged by applying advanced technologies. A modified language processing model capable of handling specialized knowledge and terminology in the field of acupuncture can be utilized (22). Additionally, supervised fine-tuning can be applied to improve reliability and performance by training the model on high-quality datasets. Significant effectiveness of refined generative LLMs has been shown in the field of Traditional Chinese Medicine (23).
An enhanced version of GPT-4 could be applied to acupoint recommendation in various scenarios. A conversational AI system could be developed to allow users to input symptoms and either receive appropriate acupoint recommendations or be prompted with further diagnostic questions. However, it is important to note that, as an LLM, it cannot capture information obtained through direct in-person interaction. The process of translating such observations into text remains a task that only trained doctors can perform. Furthermore, the knowledge that LLMs learn from training data does not inherently include the experiential insights of doctors. Therefore, when applying GPT-4 to decision-making in clinical settings, careful consideration is needed to ensure that medical professionals appropriately integrate their clinical expertise. Additionally, such system could be beneficial in medical education by allowing students to input symptoms, engage in interactive discussions, and explore various approaches to acupoint selection. This could help students develop a deeper understanding of acupuncture treatment strategies while fostering a more dynamic learning experience.
Despite our findings, several limitations and challenges should be addressed regarding the application of GPT-4 in acupuncture treatment. First, this study did not apply advanced techniques for better performance, such as fine-tuning methods or prompt engineering. Although our study showed that GPT-4 performed remarkably, substantial evidence suggest that these techniques can further improve its performance in medical domain. In particular, supervised fine-tuning trains the model on selectively curated datasets, it can help address a key limitation—GPT-4 might have learned detailed clinical interventions and case-specific information from the published case reports used in this study. If GPT-4 has learned from these case reports, its responses may be biased toward the reported treatment interventions, potentially limiting its generalizability in clinical applications and hindering its broader applicability in clinical decision-making. Supervised fine-tuning methods, together with prompt engineering could help address these concerns and lead to more accurate responses in future studies. Second, the lack of representativeness and generalizability should be noted. The selected 10 cases and 80 doctors from Korea do not fully reflect acupuncture treatment practices across different cultural and educational backgrounds. Recent findings have shown that China-based models performed better on Traditional East Asian Medicine tasks (24). Moreover, relying solely on Korean clinicians and case reports may introduce inherent cultural biases, which could limit the representativeness of our findings. These factors highlight the need for future studies to expand the range of sources, clinicians, and models to improve representativeness and mitigate cultural bias. Third, GPT-4's knowledge is limited by a temporal cutoff and may not incorporate the most recent evidence or clinical guidelines. In addition, this study could not include physical examination findings from direct observation, which are critical in real-world clinical decision-making. Nevertheless, our study design ensured that both GPT-4 and doctors were provided with the same patient information and the same task request for acupoint suggestion, allowing comparison under equivalent conditions. Fourth, there was a notable sample-size imbalance (10 GPT-4 runs vs. 80 doctors' responses). Nonetheless, variability analyses showed that doctors' responses were consistently more heterogeneous than GPT-4, suggesting that 10 GPT-4 runs, while still a limitation, can reasonably represent GPT-4's outputs under fixed prompt conditions. Finally, ethical issues, the generation of inaccurate content, and issues of reproducibility and uncertainty have been raised consistently regarding the use of chatbots in healthcare (1). These risks should be considered carefully when applying and interpreting the output of chatbots in clinical practice.
Conclusions
In conclusion, our findings provide preliminary evidence that GPT-4 may support acupoint selection, in some cases making decisions similar to those of human clinicians. This suggests that generative AI could potentially be applied to support medical decision-making as well as acupuncture education and personalized medicine. By offering recommendations comparable to those made by experienced practitioners, such models may serve as a valuable tool for enhancing clinical practice, education, and patient care in acupuncture medicine. However, larger validation studies are needed to develop a more reliable and effective tool for real-world use.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.
Ethics statement
The study was approved by the Institutional Review Board of Kyung Hee University, Seoul, Republic of Korea (KHSIRB-17-046). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
D-EY: Conceptualization, Formal analysis, Writing – original draft. C-HK: Data curation, Methodology, Writing – review & editing. YR: Funding acquisition, Project administration, Writing – review & editing. Y-SL: Conceptualization, Writing – review & editing. YC: Conceptualization, Supervision, Writing – original draft.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2024-00449485), and Korea Institute of Oriental Medicine (KSN2511011) and an Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korea government (MSIT) [RS-2022-00155911, Artificial Intelligence Convergence Innovation Human Resources Development (Kyung Hee University)]. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2025.1632303/full#supplementary-material
Abbreviations
AI, artificial intelligence; GPT-4, generative pre-trained transformer 4; KMD, Korean Medicine doctor; LLM, large language model; TEAM, traditional East Asian medicine; WHO, world health organization.
References
1. Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). (2023) 11:887. doi: 10.3390/healthcare11060887
2. Fang C, Wu Y, Fu W, Ling J, Wang Y, Liu X, et al. How does ChatGPT-4 preform on non-English national medical licensing examination? An evaluation in Chinese language. PLOS Digit Health. (2023) 2:e0000397. doi: 10.1371/journal.pdig.0000397
3. Jang D, Yun TR, Lee CY, Kwon YK, Kim CE. GPT-4 can pass the Korean National Licensing Examination for Korean Medicine Doctors. PLOS Digit Health. (2023) 2:e0000416. doi: 10.1371/journal.pdig.0000416
4. Kung TH, Cheatham M, Medenilla A, Sillos C, De Leon L, Elepano J, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. (2023) 2:e0000198. doi: 10.1371/journal.pdig.0000198
5. Takagi S, Watari T, Erabi A, Sakaguchi K. Performance of GPT-35 and GPT-4 on the Japanese Medical Licensing examination: comparison study. JMIR Med Educ. (2023) 9:e48002. doi: 10.2196/48002
6. Lahat A, Sharif K, Zoabi N, Shneor Patt Y, Sharif Y, Fisher L, et al. Assessing Generative Pretrained Transformers (GPT) in clinical decision-making: comparative analysis of GPT-35 and GPT-4. J Med Internet Res. (2024) 26:e54571. doi: 10.2196/54571
7. Sandmann S, Riepenhausen S, Plagwitz L, Varghese J. Systematic analysis of ChatGPT, Google search and Llama 2 for clinical decision support tasks. Nat Commun. (2024) 15:2050. doi: 10.1038/s41467-024-46411-8
8. Abd-Alrazaq A, Alsaad R, Alhuwail D, Ahmed A, Healy PM, Latifi S, et al. Large language models in medical education: opportunities, challenges, and future directions. JMIR Med Educ. (2023) 9:e48291. doi: 10.2196/48291
9. Birch S, Alraek T. Traditional East Asian medicine: how to understand and approach diagnostic findings and patterns in a modern scientific framework? Chin J Integr Med. (2014) 20:336–40. doi: 10.1007/s11655-014-1809-3
10. Jung WM, Lee SH, Lee YS, Chae Y. Exploring spatial patterns of acupoint indications from clinical data: a STROBE-compliant article. Medicine. (2017) 96:e6768. doi: 10.1097/MD.0000000000006768
11. Jung WM, Park IS, Lee YS, Kim CE, Lee H, Hahm DH, et al. Characterization of hidden rules linking symptoms and selection of acupoint using an artificial neural network model. Front Med. (2019) 13:112–20. doi: 10.1007/s11684-017-0582-z
12. Chen J, Xie Y, Lin Q, Qian Z, Feng J, Zhang J, et al. Investigating acupoint selection and combinations of acupuncture for Tic disorders: an association rule mining and network analysis study. Front Neurol. (2022) 13:894951. doi: 10.3389/fneur.2022.894951
13. Choi EM, Jiang F, Longhurst JC. Point specificity in acupuncture. Chin Med. (2012) 7:4. doi: 10.1186/1749-8546-7-4
14. Hwang YC, Lee IS, Ryu Y, Lee MS, Chae Y. Exploring traditional acupuncture point selection patterns for pain control: data mining of randomised controlled clinical trials. Acupunct Med. (2020) 39:964528420926173. doi: 10.1177/0964528420926173
15. Lee H. Using ChatGPT as a learning tool in acupuncture education: comparative study. JMIR Med Educ. (2023) 9:e47427. doi: 10.2196/47427
16. Kanjee Z, Crowe B, Rodman A. Accuracy of a generative artificial intelligence model in a complex diagnostic challenge. JAMA. (2023) 330:78–80. doi: 10.1001/jama.2023.8288
17. Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. (2023) 183:589–96. doi: 10.1001/jamainternmed.2023.1838
18. Kim CH, Yoon DE, Lee YS, Jung WM, Kim JH, Chae Y, et al. Revealing associations between diagnosis patterns and acupoint prescriptions using medical data extracted from case reports. J Clin Med. (2019) 8:1663. doi: 10.3390/jcm8101663
19. Lee YS, Ryu Y, Yoon DE, Kim CH, Hong G, Hwang YC, et al. Commonality and specificity of acupuncture point selections. Evid Based Complement Alternat Med. (2020) 2020:2948292. doi: 10.1155/2020/2948292
20. Lee IS, Chae Y. Identification of major traditional acupuncture points for pain control using network analysis. Acupunct Med. (2021) 39:553–4. doi: 10.1177/0964528420971309
21. Johnson KB, Wei WQ, Weeraratne D, Frisse ME, Misulis K, Rhee K, et al. Precision medicine, AI, and the future of personalized health care. Clin Transl Sci. (2021) 14:86–93. doi: 10.1111/cts.12884
22. Xu T, Wen J, Wang L, Huang Y, Zhu Z, Zhu Q, et al. Acupuncture indication knowledge bases: meridian entity recognition and classification based on ACUBERT. Database. (2024) 2024:baae083. doi: 10.1093/database/baae083
23. Tan Y, Zhang Z, Li M, Pan F, Duan H, Huang Z, et al. MedChatZH: A tuning LLM for traditional Chinese medicine consultations. Comput Biol Med. (2024) 172:108290. doi: 10.1016/j.compbiomed.2024.108290
Keywords: large language model, artificial intelligence, GPT-4, medical decision-making, acupoint selection
Citation: Yoon D-E, Kim C-H, Ryu Y, Lee Y-S and Chae Y (2025) Performance of GPT-4 for planning acupuncture treatment: comparison with human clinician performance. Front. Med. 12:1632303. doi: 10.3389/fmed.2025.1632303
Received: 22 May 2025; Accepted: 01 September 2025;
Published: 24 September 2025.
Edited by:
Mingliang Suo, Beihang University, ChinaReviewed by:
Hubert Van Griensven, St George's, University of London, United KingdomWeiqi Jiao, Boston Strategic Partners Inc., United States
Copyright © 2025 Yoon, Kim, Ryu, Lee and Chae. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Younbyoung Chae, eWJjaGFlQGtodS5hYy5rcg==