Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 24 May 2024
Sec. Head and Neck Cancer

Assessing the role of advanced artificial intelligence as a tool in multidisciplinary tumor board decision-making for primary head and neck cancer cases

Benedikt SchmidlBenedikt Schmidl1Tobias HüttenTobias Hütten1Steffi PigorschSteffi Pigorsch2Fabian StgbauerFabian Stögbauer3Cosima C. HochCosima C. Hoch1Timon HussainTimon Hussain1Barbara WollenbergBarbara Wollenberg1Markus Wirth*Markus Wirth1*
  • 1Department of Otolaryngology Head and Neck Surgery, Technical University Munich, Munich, Germany
  • 2Department of RadioOncology, Technical University Munich, Munich, Germany
  • 3Institute of Pathology, Technical University Munich, Munich, Germany

Background: Head and neck squamous cell carcinoma (HNSCC) is a complex malignancy that requires a multidisciplinary approach in clinical practice, especially in tumor board discussions. In recent years, artificial intelligence has emerged as a tool to assist healthcare professionals in making informed decisions. This study investigates the application of ChatGPT 3.5 and ChatGPT 4.0, natural language processing models, in tumor board decision-making.

Methods: We conducted a pilot study in October 2023 on 20 consecutive head and neck cancer patients discussed in our multidisciplinary tumor board (MDT). Patients with a primary diagnosis of head and neck cancer were included. The MDT and ChatGPT 3.5 and ChatGPT 4.0 recommendations for each patient were compared by two independent reviewers and the number of therapy options, the clinical recommendation, the explanation and the summarization were graded.

Results: In this study, ChatGPT 3.5 provided mostly general answers for surgery, chemotherapy, and radiation therapy. For clinical recommendation, explanation and summarization ChatGPT 3.5 and 4.0 scored well, but demonstrated to be mostly an assisting tool, suggesting significantly more therapy options than our MDT, while some of the recommended treatment modalities like primary immunotherapy are not part of the current treatment guidelines.

Conclusions: This research demonstrates that advanced AI models at the moment can merely assist in the MDT setting, since the current versions list common therapy options, but sometimes recommend incorrect treatment options and in the case of ChatGPT 3.5 lack information on the source material.

1 Introduction

The incidence of cancer worldwide is more than 24.5 million cases, resulting in 9.6 million deaths per year (1). Head and neck squamous cell carcinoma (HNSCC), ranking as the seventh most common type of cancer, poses unique challenges due to the complexity of the anatomical region and heterogeneous nature of disease. Consequently, the most common approach is the discussion of oncological cases in a multidisciplinary tumor board (MDT). MDTs are designed to improve patient outcomes through an individualized and collaborative strategy (2). They bring together a diverse team of medical professionals, including medical and surgical oncologists, radiation oncologists, pathologists, and radiologists to discuss the treatment of cancer patients (1, 3). MDTs provide a comprehensive perspective on each case, tailoring treatment plans to individual needs, especially through molecular tumor boards for difficult-to-treat cases (3, 4). Despite their benefits, MDTs face obstacles such as costs, responsibilities, geographic barriers, and treatment delays (24).

In parallel, the field of artificial intelligence (AI), especially deep learning (DL) and natural language processing (NLP), has proven significant advancement (5). Models like Generative Pre-trained Transformer (GPT), developed by OpenAI (San Francisco, California) in 2018, have surpassed earlier AI tools in resources and capabilities (5). These AI models have already shown potential in oncology, as exemplified by IBM Watson, and are now being explored for their utility in assisting clinical decision-making (6).

While AI at the moment cannot replace the expertise of healthcare professionals, it may aid in decision-making by efficiently accessing and presenting relevant information. Our group’s previous work demonstrated that ChatGPT accurately answered 65% of ENT tumor related practice questions for the German otolaryngology board certification (7). However, the comparative effectiveness of ChatGPT 3.5 and 4.0 in assisting MDTs for primary cases of head and neck cancer remains to be elucidated.

This study aims to assess the potential of ChatGPT 3.5 and 4.0 in providing suitable recommendations for patients diagnosed with primary head and neck cancer, comparing its performance with traditional MDTs.

2 Materials and methods

2.1 Patient cohort

The inclusion criteria comprised a confirmed diagnosis of head and neck cancer and the absence of distant metastasis. Excluded from consideration were cases involving recurrent situations. Tumor characteristics and the age of the ten consecutive patients before treatment were obtained from the MDT. In this study, a total of 20 patients were included. These cases were discussed in the MDT of October 2023 in the Department of Otorhinolaryngology/Head and Neck Surgery, Klinikum rechts der Isar, Technical University of Munich. The age of patients in this study ranged from 52 to 74 years (Median age 61.5, Average age 63.8). The data were anonymized before being provided to the investigators, ensuring that patient identification was not possible. Subsequently, the extracted information (Table 1) was entered into the ChatGPT 3.5 or 4.0 chat bot. The resulting answers were then copied and categorized.

Table 1
www.frontiersin.org

Table 1 Overview of the patient cohort and the information entered into ChatGPT.

Ethical approval was waived by the ethics committee of the Technical University of Munich due to the retrospective nature of this study.

2.2 Artificial intelligence/ChatGPT

ChatGPT is a publicly accessible artificial intelligence chatbot, that is based on a transformer-based language model. It can produce text that is similar to human language. A special characteristic is that it is only trained on data up to 2021. The user can input questions via a website, and ChatGPT analyzes the contextual relationships among words in the query. The training data for ChatGPT stem from various open-access internet sources, including websites, articles, and books, up until the year 2021 (48) GPT model 3.5 and ChatGPT 4.0. were used, available publicly in October 2023.

2.3 ChatGPT prompt format and data evaluation

A standardized prompt format was used for inputting patient information into ChatGPT, mimicking the setting of an individual patient presentation in MDTs. The prompt used for this study was the standardized way of presenting a case at the MDT to allow a direct comparison of the results to the recommendations of the MDT. Before this prompt was used, four different prompts were tested to validate the results, since a lack of information leads to insufficient and vague answers. The other three prompts are depicted in Supplementary Figure 1. The prompt was as follows: “A (XX) year old patient with a cT (XX) cN (XX) squamous cell carcinoma of the (XX), the patient (XX) smokes, (XX) and has the following secondary diseases (XX) with a Karnofsky Index of (XX). The patient is presented in an interdisciplinary tumor board. What treatment options are available and which option do you think leads to the best prognosis?”. This setup replicates an oncologist/head and neck surgeon interaction with ChatGPT. An example would be: “A 74-year-old patient with a cT4a cN0 squamous cell carcinoma of the left transglottic larynx, the patient no longer smokes, cum. 25 pack/years and has the following secondary diseases: arterial hypertension with a Karnofsky Index of 100%. The patient is presented in an interdisciplinary tumor board pre-therapeutically. What treatment options are available and which option do you think leads to the best prognosis?” Following each response, no further dialogue was initiated; the ChatGPT history was cleared, and the subsequent question was posed. All the prompts can be found in the Supplementary Data. A new ChatGPT session was started for each prompt to avoid any sort of training or influence by prior answers. Each answer by ChatGPT was collected by B.S. and reviewed by B.S. and T.H. independently. For each treatment modality that was mentioned one point was given to ChatGPT. Additionally, the grading scales for Summarization, Clinical Recommendation and Explanation as already used by Sorin et al., 2023 (9) were used to rate the answers by ChatGPT 3.5 and 4.0. The grading scales are depicted in the Supplementary Figure 1. The scoring transitioned from a baseline of zero points to a maximum of five points in these grading scales. The whole treatment history was considered. The inter-rater reliability was measured using Cohen’s kappa coefficient.

3 Results

The most common response from ChatGPT 3.5 involved the primary treatment modalities available for potential use in a primary or adjuvant setting, including surgery, radiotherapy, and chemotherapy. An exemplary prompt and the answers generated by ChatGPT 3.5 and ChatGPT 4.0 are depicted in Figure 1. When comparing the MDT recommendations with those of ChatGPT 3.5, two independent reviewers reached an agreement measured by Cohen’s κ of 1 for the number of therapy options, of Cohen’s κ of 0.459 for summarization of text, of 0.44 for clinical recommendation, and 0.579 for explanation on the decision made. ChatGPT 4.0 reached a Cohen’s κ of 1 for the number of therapy options, of Cohen’s κ of 0.612 for summarization of text, of 0.455 for clinical recommendation, and 0.452 for explanation on the decision made.

Figure 1
www.frontiersin.org

Figure 1 Depiction of an exemplary prompt and responses by ChatGPT 4.0 and 3.5. The answers were copied from the response page.

In absolute numbers ChatGPT 3.5 suggested on average a significantly higher number of treatment options for the patients in this study (4.85), compared to the MDT (Table 2). A summary and depiction of the results is presented in Table 2. Immunotherapy was recommended for 60% of the patients by ChatGPT 3.5 in this study, even in primary early stage HNSCC, without regard to the PD-L1 status or combined positive score (CPS). Neoadjuvant trials were not recommended by ChatGPT 3.5 or ChatGPT 4.0 as well as other strategies like HPV vaccination. The age of the patients was emphasized by ChatGPT 3.5 with a potential inability to endure surgery or chemotherapy. Surgery was recommended by ChatGPT 3.5 for all patients in this study (100%), also reaching the highest concordance with the MDT (18 out of 20 patients, 90%). Surgery was also recommended in two cases that explicitly had a quite advanced tumor infiltration of the supraglottic structures and into the trachea or on multiple subsites, where the MDT suggested a definitive radiochemotherapy. In these special cases a second thread in ChatGPT 3.5 was opened and it was specifically asked which therapy offers the highest quality of life. Surgery was still the first suggestion of ChatGPT 3.5, but it was mentioned that infiltration into the trachea and “other factors” may lead to surgery not being the sole therapy for this patient. ChatGPT 3.5 in general seemed to be careful not to recommend a single treatment modality and tried to balance the importance of different therapy options. On the other hand, some therapy options, such as adjuvant radiochemotherapy were not mentioned by ChatGPT. ChatGPT 3.5 and 4.0 were able to name different surgical approaches for some cases.

Table 2
www.frontiersin.org

Table 2 Overview of the results and treatment options of ChatGPT 3.5 compared to the MDT.

ChatGPT 4.0 suggested on average a higher number of treatment options than ChatGPT 3.5 for the patients in this study (5.1), significantly more than the MDT, which in our institution mostly recommends a single therapy, or a maximum of two therapies for each case (Table 3; Figure 2). A summary and depiction of the results is presented in Table 3. When analyzing the scales of summarization, clinical recommendation, and explanation, we see that ChatGPT 4.0 achieves better grades for summarization, explanation and clinical recommendation from both Reviewer 1 and 2 (Figure 2). Additionally, ChatGPT 4.0 is able to reference its sources, in some cases it even mentions recent clinical studies. On the other hand, ChatGPT 4.0 also refrained from giving precise recommendations, and mentioned that it is not meant to give medical advice or replace the opinion of a medical doctor. ChatGPT 4.0 recommended surgery for 18 out of 20 patients (90%), which is in concordance with the decisions of the tumor board. Two patients with extensive tumor spread, which the MDT deemed candidates for radiochemotherapy, were allocated for definitive radiochemotherapy. These patients would have received surgery according to ChatGPT 3.5, even though the extent of the tumor growth limits surgical resection, and this therapy was therefore not recommended by the MDT.

Table 3
www.frontiersin.org

Table 3 Overview of the results and treatment options of ChatGPT 4.0 compared to the MDT.

Figure 2
www.frontiersin.org

Figure 2 Rating of the performance of ChatGPT 4.0 and 3.5 by comparison of the number of treatment options per case and the grading of summarization of text, clinical recommendation, and explanation on the decision made by two independent reviewers. Each bar is the average of the two independent reviewers grading. # = number.

4 Discussion

This study represents the first examination of ChatGPT 3.5 and 4.0, an open AI, as an auxiliary tool for MDTs engaged in discussions about patients with a primary diagnosis of HNSCC. The recommendations of ChatGPT 3.5 and 4.0 were compared. NLP, a subset of AI focusing on the analysis of human language, has found applications across various medical specialties, including breast cancer research, rheumatology, medical education, medical exams, and many more (8, 1014). The diagnostic accuracy in complex medical cases was tested in a recent publication, showing that ChatGPT accurately identifies the correct diagnosis in nearly 40% of cases as its primary suggestion (15). These findings, and ChatGPT’s ability to provide logical and contextually appropriate responses to diverse text questions through the application of advanced language modeling techniques and extensive access to large and diverse datasets (10, 11) led to this study.

MDTs deal with a lot of data when reviewing a patient’s case with the goal of tailoring general treatment plans to the patient’s individual needs and therefore improving care (2). Organizing and processing data is one of the strengths of ChatGPT, which is the reason it may become an auxiliary tool for the MDT in the future. In this study ChatGPT 3.5 and 4.0 achieved good grades for clinical recommendation, explanation and summarization in this study. ChatGPT 3.5 and 4.0 listed the different treatment options in a very efficient way, highlighting its potential to enhance efficiency in clinical workflows. A recommendation for immunotherapy as a therapy option was given to 60% of the patients in this study. This rate was significantly higher (600%) than the rate of the MDT. There the recommendation was only given in a very advanced case where the tumor was already unresectable. In Germany and other countries, immunotherapy is currently limited to patients with recurrent/metastatic HNSCC, with the two antibodies that were approved by the US Food and Drug Administration (FDA), nivolumab and pembrolizumab, widely used (1618). The recommendation of immunotherapy in the primary cases of HNSCC in this study are currently not concordant with the FDA approval and treatment guidelines. Another often recommended treatment option of ChatGPT was the definitive radiotherapy, regardless of the tumor localization. This topic is widely discussed in oropharyngeal carcinoma with primary radiotherapy and/or chemoradiotherapy still widely used, especially in settings where transoral surgery is not feasible (19, 20) as well as in hypopharyngeal carcinoma, which is often asymptomatic for a long time until a surgical therapy is not possible anymore (19). The concordance with the MDT in terms of surgical therapy was 90% (9 out of 10 patients). Surgery was recommended by ChatGPT 3.5 for all patients in this study, even for one patient with a very advanced, unresectable tumor, where the MDT suggested a definitive radiochemotherapy or immunotherapy. When ChatGPT 3.5 was in this one case asked for a therapy option with the highest quality of life, surgery was still the first therapy, but the problem of potential surgical limitations was at least addressed. ChatGPT 4.0 recommended a definitive radiochemotherapy for this patient and was therefore superior in terms of concordance with the MDT. ChatGPT 3.5 reached high scores in terms of explanation, clinical recommendation and summarization by both reviewers in this study. The explanation grade was graded the worst. The reason for this result may be that while the on average long answers of ChatGPT might lead some users into thinking the LLMs explain the case in detail, the overall explanation for the therapy of primary head and neck cancer was not convincing to the reviewing head and neck oncologists. ChatGPT 4.0 had the overall better performance with a better rate of concordance with the MDT, probably due to the ability to access more data than ChatGPT 3.5, as demonstrated in a few studies (8, 10, 15).

There have been a few studies investigating the use of ChatGPT in the field of medical and surgical oncology due to the ability to rapidly extract and deal with a large quantity of information (7, 9, 20). ChatGPT is often reduced to an auxiliary tool or assistant rather than a replacement of the MDT in most studies. One of these studies investigated decision making with ChatGPT for ten consecutive patients with primary breast cancer. This study compared the treatment recommendations of ChatGPT with the MDT and found similarities in 7 out of 10 cases (9). Compared to the results for primary head and neck cancer, a similar Cohen’s κ and grades for summarization, explanation and clinical recommendation were found, while our study additionally analyzed the number of therapy options as a potential indicator of the quality of ChatGPT. For primary breast cancer a high reference rate for surgery of 80% was found, similar to the rate of 90% (for ChatGPT 3.5) seen for head and neck cancer in this study. For the use in otolaryngology a recent publication discovered in a clinical case series of laryngology and oncological cases a performance of decision making of 60-68% (15). That study used ChatGPT 4.0 as a tool in the clinical setting focusing on diagnostic examinations and the general clinical workflow without addressing the specific needs and situation of the MDT. Another study also investigated the use of ChatGPT for breast cancer and used prompts in German, leading to a lower concordance with mostly general answers, probably since the literature and studies and therefore the source material of ChatGPT is predominantly in English and may impair the access of ChatGPT (12). In a recent study by Benary et al., the fictional cases of 10 patients with advanced cancer with genetic alterations were submitted to 4 different LLMs and one physician to identify personalized treatment options. The number of treatment options, precision, recall, F1 score of LLMs compared with human experts, recognizability, and usefulness of recommendations were graded, but the results did not reach the quality of human experts (21). The number of therapy options recommended by LLMs was large, similar to the results of our study for head and neck cancer, while the authors summarized that LLMs mainly support idea generation and cannot be used in the current oncological routine. This finding is very similar to the high number of therapy options recommended by ChatGPT for head and neck cancer in our study. ChatGPT’s ability to summarize a vast number of therapy options is at the moment just informative, since the participants of the MDT would potentially also be able to name the therapy options, but name only the therapy option with the best prognosis or the one that is most suitable to the patient’s current situation. The ability to name the therapy options could be used for teaching purposes and to enable the patient to make an informed decision but is currently hindered by a lack of information provided on the prognosis of the patient based on the suggested therapy. Maybe for this reason, ChatGPT explicitly states in most responses that it is not programmed as a medical bot and the literature reveals that there has been no prior oncological training of ChatGPT (7, 20) Therefore other research groups investigated the use of a clinical decision support system based on Bayesian networks (BN) for laryngeal cancer (LC) as a prototype with over 1,000 variables (2224). The TNM classification was the main classifier for the therapeutic recommendations of the Bayesian Network, compared to the ability of ChatGPT 4.0 in this study to address the comorbidities, extent of the tumor and in the case of ChatGPT 4.0 some of the latest studies, including the quality of life of a patient. Limitations of this study of Bayesian Networks include the limited access of other physicians to the software, the integration of only laryngeal carcinoma and a general lack of data on immunotherapy due to insufficient data. On the other hand, ChatGPT accessed and cited recent studies, including immunotherapy, while also referencing some rarely used therapy options, as way of thinking outside the box, similar to the conclusion of Benary et al., 2023 (21).

The major benefits of using AI in the setting of the tumor board tackle some of the biggest challenges of the physical MDT, while most of these positive aspects are connected to the digital interface of it. Like in a virtual tumor board, the participants of the AI-guided MDT may benefit from the exceptional accessibility of ChatGPT without geographical boundaries. While the MDT at the hospital where this study was performed, hosts physical MDTs, an AI guided MDT may boost the number of participating physicians by not limiting them to a specific place. Currently ChatGPT is free of charge in version 3.5, which makes it an accessible tool for physicians in developing countries, where resources in terms of oncological knowledge and surgical expertise are limited to major hospitals. Another potential use is the ability to summarize large datasets of information and extract key content from records of patients to facilitate a more informed decision making process as suggested by some studies (20). There was no direct comparison of the time spent in a physical MDT versus the MDT with ChatGPT, but ChatGPT answered the prompts in this study in an exceptionally fast way, taking mere seconds. On of the other hand, ChatGPT in it’s current version has a few limitations. For ChatGPT 3.5 the source information used for the answer is missing. Every medical professional in the MDT has to discuss the therapy options with other specialists and has to base his opinion on scientific ground (2, 3). Without naming the source of information there can be no discussion based on the answers of ChatGPT. A limitation of this study is that even though this study analyzes one of the larger cohorts of patients compared to the literature for breast cancer, only a small patient cohort (n =20) was analyzed, impacting the study results due to inherent patient heterogeneity. To make up for this, only consecutive primary cases of HNSCC were included in this proof-of-concept study. Future studies may need to investigate ChatGPT-driven MDT decisions in a more homogeneous patient group, focusing on only one specific location of HNSCC. Another drawback is that in this study the MDT of only one institution was investigated, whereas other MDTs might decide cases differently due to historical or resource/local reasons, or simply because of different physicians being present (3). Another limitation is that ChatGPT is based on a large language model and will access information that is based on the prompt (12, 14). Four different prompts were used initially, but only one prompt, identical to the way a case would have been presented at the MDT was used for this study to directly compare the results to the MDT. This is a limitation, that might have influenced the results of this study, since the prompt can have a significant impact on the answers of ChatGPT as seen in other studies. ChatGPT will, at least in the current version, not ask for additional information, whereas the participants of the MDT will demand more information if the available information is insufficient, or the decision-making is difficult. This was also confirmed in other studies (11, 21). In addition, it remains to be elucidated whether a large dataset of clinical data from the MDT may be able to train ChatGPT or a different AI tool to get more tailored recommendations. Within the clinical context of a MDT, ChatGPT 4.0 was significantly superior to ChatGPT 3.5 in terms of transparency of the information, citing sources, including a clinical study on quality of life after laryngeal surgery. For one patient with extensive tumor growth, ChatGPT 3.5 chose a different treatment than the MDT, while ChatGPT 4.0 was able to recommend the right treatment. The most important limitation and one of the reasons for the results in this study is that ChatGPT is not built to think independently, producing texts based on public documents and databases (5). It is therefore not able to make individual decisions, thereby failing to tailor the therapy to the patient’s current situation and needs. The treatment plan of the MDT combines the clinical knowledge of the patient and the results of recent studies, to optimize a patient’s prognosis, while also balancing potential downsides and risks of the therapy (14, 21). When one thinks of the three potential use cases of ChatGPT in the future, a fully AI-guided MDT, an assistant to the MDT, or a validation method, ChatGPT, even in its most advanced version (ChatGPT 4.0) is currently merely able to assist the MDT. The reason for this is, that the ChatGPT was able to reproduce text from online databases to name a vast number of therapy options, but doesn’t address the scientific clinical studies in this study that lead the MDT to recommend a specific therapy for a patient, while in theory ChatGPT 4.0 is able to produce convincing answers in terms of summarization, explanation and clinical recommendation. An ideal AI-guided MDT would be able to evaluate the most recent oncological studies and compare the results to the standard operating procedure to enable a more up to date and faster decision-making process. In the closer future an AI-validation by LLMs could already assess the decisions of the MDT and comment on the scientific basis, with the task of tailoring the treatment to the patient still in the hands of the MDT. Nonetheless it will remain the MDTs participants task to carefully evaluate the recommendations based on their own clinical knowledge and knowledge of the patient’s situation.

5 Conclusions

In the current version ChatGPT is able to list the various treatment options for primary head and neck cancer. Even though there was no prior oncological training, ChatGPT, especially in Version 4.0 reached exceptionally high scores for explanation and clinical recommendation by two independent reviewers. The limitations of ChatGPT, including a lack of personalized treatment planning, the recommendation of incorrect treatment guidelines and in the case of ChatGPT 3.5 a lack of information on the source material, currently allows ChatGPT to merely assist the MDT for head and neck cancer cases.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

BS: Conceptualization, Formal analysis, Funding acquisition, Methodology, Resources, Validation, Visualization, Writing – original draft, Writing – review & editing. ToH: Formal analysis, Methodology, Writing – review & editing. SP: Conceptualization, Investigation, Resources, Supervision, Writing – review & editing. FS: Data curation, Formal analysis, Methodology, Writing – review & editing. CH: Data curation, Formal analysis, Validation, Writing – review & editing. TiH: Conceptualization, Methodology, Supervision, Writing – review & editing. BW: Conceptualization, Methodology, Supervision, Validation, Writing – review & editing. MW: Conceptualization, Resources, Supervision, Validation, Writing – original draft.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2024.1353031/full#supplementary-material

References

1. Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. (2018) 68:394–424. doi: 10.3322/caac.21492

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Luchini C, Lawlor RT, Milella M, Scarpa A. Molecular tumor boards in clinical practice. Trends Cancer. (2020) 6:738–44. doi: 10.1016/j.trecan.2020.05.008

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Berardi R, Morgese F, Rinaldi S, Torniai M, Mentrasti G, Scortichini L, et al. Benefits and limitations of a multidisciplinary approach in cancer patient management. Cancer Manag Res. (2020) 12:9363–74. doi: 10.2147/CMAR.S220976

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Thenappan A, Halaweish I, Mody RJ, Smith EA, Geiger JD, Ehrlich PF, et al. Review at a multidisciplinary tumor board impacts critical management decisions of pediatric patients with cancer. Pediatr Blood Cancer. (2017) 64:254–8. doi: 10.1002/pbc.v64.2

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the feasibility of ChatGPT in healthcare: an analysis of multiple clinical and research scenarios. J Med Syst. (2023) 47:33. doi: 10.1007/s10916-023-01925-4

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Zhou N, Zhang CT, Lv HY, Hao CX, Li TJ, Zhu JJ, et al. Concordance study between IBM watson for oncology and clinical practice for patients with cancer in China. Oncologist. (2019) 24:812–9. doi: 10.1634/theoncologist.2018-0255

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Temsah O, Khan SA, Chaiah Y, Senjab A, Alhasan K, Jamal A, et al. Overview of early ChatGPT's presence in medical literature: insights from a hybrid literature review by ChatGPT and human experts. Cureus. (2023) 15:e37281. doi: 10.7759/cureus.37281

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Huang Y, Gomaa A, Semrau S, Haderlein M, Lettmaier S, Weissmann T, et al. Benchmarking ChatGPT-4 on a radiation oncology in-training exam and Red Journal Gray Zone cases: potentials and challenges for ai-assisted medical education and decision making in radiation oncology. Front Oncol. (2023) 13:1265024. doi: 10.3389/fonc.2023.1265024

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Sorin V, Klang E, Sklair-Levy M, Cohen I, Zippel DB, Lahat Balint N, et al. Large language model (ChatGPT) as a support tool for breast tumor board. NPJ Breast Cancer. (2023) 9:44. doi: 10.1038/s41523-023-00557-8

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Hughes KS, Zhou J, Bao Y, Singh P, Wang J, Yin K. Natural language processing to facilitate breast cancer research and management. Breast J. (2020) 26:92–9. doi: 10.1111/tbj.13718

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Hügle T. The wide range of opportunities for large language models such as ChatGPT in rheumatology. RMD Open. (2023) 9. doi: 10.1136/rmdopen-2023-003105

CrossRef Full Text | Google Scholar

12. Lukac S, Dayan D, Fink V, Leinert E, Hartkopf A, Veselinovic K, et al. Evaluating ChatGPT as an adjunct for the multidisciplinary tumor board decision-making in primary breast cancer cases. Arch Gynecol Obstet. (2023) 308:1831–44. doi: 10.1007/s00404-023-07130-5

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Kanjee Z, Crowe B, Rodman A. Accuracy of a generative artificial intelligence model in a complex diagnostic challenge. JAMA. (2023) 330:78–80. doi: 10.1001/jama.2023.8288

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Chakraborty C, Pal S, Bhattacharya M, Dash S, Lee S-S. Overview of Chatbots with special emphasis on artificial intelligence-enabled ChatGPT in medical science. Front Artif Intell. (2023) 6. doi: 10.3389/frai.2023.1237704

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Lechien JR, Georgescu BM, Hans S, Chiesa-Estomba CM. ChatGPT performance in laryngology and head and neck surgery: a clinical case-series. Eur Arch Otorhinolaryngol. (2023). doi: 10.1007/s00405-023-08282-5

CrossRef Full Text | Google Scholar

16. Cohen EE, LaMonte SJ, Erb NL, Beckman KL, Sadeghi N, Hutcheson KA, et al. American cancer society head and neck cancer survivorship care guideline. CA Cancer J Clin. (2016) 66:203–39. doi: 10.3322/caac.21343

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Hanna GJ, O’Neill A, Shin KY, Wong K, Jo VY, Quinn CT, et al. Neoadjuvant and adjuvant nivolumab and lirilumab in patients with recurrent, resectable squamous cell carcinoma of the head and neck. Clin Cancer Res. (2022) 28:468–78. doi: 10.1158/1078-0432.CCR-21-2635

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Hanna GJ, Villa A, Shi R, O’Neill A, Liu M, Quinn CT, et al. 650O A phase II study of nivolumab for high-risk oral leukoplakia. Ann Oncol. (2022) 33:S839. doi: 10.1016/j.annonc.2022.07.774

CrossRef Full Text | Google Scholar

19. Bozec A, Culié D, Poissonnet G, Demard F, Dassonville O. Current therapeutic strategies in patients with oropharyngeal squamous cell carcinoma: impact of the tumor HPV status. Cancers (Basel). (2021) 13. doi: 10.3390/cancers13215456

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Uprety D, Zhu D, West HJ. ChatGPT-A promising generative AI tool and its implications for cancer care. Cancer. (2023) 129:2284–9. doi: 10.1002/cncr.34827

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Benary M, Wang XD, Schmidt M, Soll D, Hilfenhaus G, Nassir M, et al. Leveraging large language models for decision support in personalized oncology. JAMA Netw Open. (2023) 6:e2343689. doi: 10.1001/jamanetworkopen.2023.43689

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Gaebel J, Wu HG, Oeser A, Cypko MA, Stoehr M, Dietz A, et al. Modeling and processing up-to-dateness of patient information in probabilistic therapy decision support. Artif Intell Med. (2020) 104:101842. doi: 10.1016/j.artmed.2020.101842

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Hikal A, Gaebel J, Neumuth T, Dietz A, Stoehr M. A treatment decision support model for laryngeal cancer based on bayesian networks. Biomedicines. (2023) 11. doi: 10.3390/biomedicines11010110

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Cypko MA, Stoehr M, Oeltze-Jafra S, Dietz A, Lemke HU. A guide for constructing Bayesian network graphs of cancer treatment decisions. Stud Health Technol Inform. (2017) 245:1355.

PubMed Abstract | Google Scholar

Keywords: HNSCC, multidisciplinary tumor board, radiotherapy, artificial intelligence, ChatGPT

Citation: Schmidl B, Hütten T, Pigorsch S, Stögbauer F, Hoch CC, Hussain T, Wollenberg B and Wirth M (2024) Assessing the role of advanced artificial intelligence as a tool in multidisciplinary tumor board decision-making for primary head and neck cancer cases. Front. Oncol. 14:1353031. doi: 10.3389/fonc.2024.1353031

Received: 09 December 2023; Accepted: 26 March 2024;
Published: 24 May 2024.

Edited by:

Dietmar Thurnher, Medical University of Graz, Austria

Reviewed by:

Jared Shenson, Orlando Health Cancer Institute, United States
Wolfgang Freysinger, Innsbruck Medical University, Austria

Copyright © 2024 Schmidl, Hütten, Pigorsch, Stögbauer, Hoch, Hussain, Wollenberg and Wirth. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Markus Wirth, Markus.Wirth@tum.de

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.