- Department of Computer Sciences and Mathematics, University of Québec at Chicoutimi, Chicoutimi, QC, Canada
Artificial Intelligence (AI) has demonstrated exceptional performance in automating critical healthcare tasks, such as diagnostic imaging analysis and predictive modeling, often surpassing human capabilities. The integration of AI in healthcare promises substantial improvements in patient outcomes, including faster diagnosis and personalized treatment plans. However, AI models frequently lack interpretability, leading to significant challenges concerning their performance and generalizability across diverse patient populations. These opaque AI technologies raise serious patient safety concerns, as non-interpretable models can result in improper treatment decisions due to misinterpretations by healthcare providers. Our systematic review explores various AI applications in healthcare, focusing on the critical assessment of model interpretability and accuracy. We identify and elucidate the most significant limitations of current AI systems, such as the black-box nature of deep learning models and the variability in performance across different clinical settings. By addressing these challenges, our objective is to provide healthcare providers with well-informed strategies to develop innovative and safe AI solutions. This review aims to ensure that future AI implementations in healthcare not only enhance performance but also maintain transparency and patient safety.
1 Introduction
In recent years, AI technology has enabled the widespread adoption of machine learning (ML) and deep learning (DL) methods in healthcare. These techniques have automated various processes like screening, diagnostics, and treatment (Zhongqi and Jia, 2022). However, DL algorithms are often considered “black boxes” due to their intricate training and numerous parameters, making it hard to decipher the connection between inputs and outputs (Zhang and Zhang, 2023). Although advanced DL methods can significantly improve speed and accuracy without human intervention, establishing an interpretable AI framework remains imperative. Understanding algorithms’ inner workings and prediction bases is essential, not just competitively but also for stakeholder trust (Stinson and Vlaad, 2024). Discussing trustworthy AI involves the terms accuracy, interpretability, and explainability, which are occasionally used interchangeably. In data mining and machine learning, interpretability is the interface between people and decision models that accurately represents the model and is intelligible to humans (Savindu Herath Pathirannehelage et al., 2024).
2 Research gaps
While current literature highlights the use of AI in healthcare, the research gaps in terms of model generalizability, lack of real-time interpretability, and inclusion of diverse clinical data remain relatively unexplored. Additionally, the integration of uncertainty quantification with interpretability models is an underexamined area that could improve AI adoption in healthcare settings. Gaps related to real-time feedback for clinicians and the role of user-centered design in AI development also need further exploration (Helman et al., 2022).
This paper characterizes the challenges of AI applications in healthcare focused on interpretability and accuracy to ensure accountability and regulatory compliance. Our survey helps healthcare providers develop appropriate strategies to rapidly implement innovative solutions safely. The goal is consolidating existing knowledge on AI systems and the interpretability models to help researchers swiftly grasp the state-of-the-art and determine areas needing more research (Saeed and Omlin, 2023). We also distinguish our review from earlier studies by focusing on both the technical and practical limitations of AI in healthcare, which are often overlooked in previous reviews, such as those focusing solely on specific AI techniques or healthcare applications.
The primary contributions of this work include:
• Summarizing the current state of AI systems in healthcare.
• Identifying and detailing the key limitations encountered when developing AI systems, with a focus on both interpretability and accuracy.
• Discussing the research gaps and future challenges in AI application in healthcare.
• Offering a detailed comparison of how this review differs from previous literature, including works like (Sadeghi et al., 2024).
Additionally, we consider the control strategy presented by Stefanelli et al. (2023) in their work on sensorimotor control in prosthetics, which integrates both force and temperature information to replicate human reflexive behavior during manipulation. This approach provides valuable insights into designing more human-like AI systems in healthcare by considering multimodal sensory inputs.
3 Methodology
To achieve our objectives, we systematically reviewed 61 AI systems for classifying and treating various diseases in healthcare. The selected systems were analyzed by extracting how they address the challenges of assessing AI systems in terms of interpretability and accuracy. Our work defines the black box of healthcare AI applications, emphasizing model interpretability and accuracy.
3.1 Search strategy
We conducted our search using multiple databases, including Google Scholar, PubMed, and IEEE Xplore, to ensure comprehensive coverage of the literature. The search was conducted using the following keywords, combined with Boolean operators:
• “Artificial Intelligence” AND “Healthcare” AND “Accuracy”
• “Deep Learning” AND “Interpretability” AND “Healthcare”
• “Machine Learning” AND “Explainability” AND “Healthcare”
3.2 Inclusion criteria
• Peer-reviewed articles focused on healthcare AI applications.
• Studies published between 2010 and 2024.
• Articles emphasizing interpretability and/or accuracy (both ML and DL).
3.3 Exclusion criteria
• Articles not focused on AI techniques (i.e., non technical).
• Non-healthcare AI applications.
• Legislative or legal discussions that do not contribute to the technical understanding of interpretability.
3.4 Data extraction
For each selected study, we extracted information on:
• The AI model used.
• The metrics employed to assess interpretability and accuracy.
• The context in which the AI model was applied.
• A detailed discussion on the results achieved.
The results of the literature analysis are summarized in Table 1.
The PRISMA diagram shown in Figure 1 illustrates the article selection process:
3.5 Study limitations
Our review primarily focuses on AI models applied in diagnostic imaging tasks. While we have considered various healthcare applications (e.g., screening, and treatment decision-making), our scope does not extend to other domains like genomics or personalized medicine, which could benefit from AI innovations. Additionally, the lack of real-time applicability and the high computational cost of certain models have not been deeply explored in this review.
4 Summary of selected literature
We reviewed 61 papers that applied AI models in various healthcare contexts. Table 1 summarizes the key information extracted from each paper, including the AI model used, the metrics for accuracy and interpretability, and the results achieved.
5 Key metrics in AI model development
Despite the numerous benefits, AI models in healthcare are not without limitations. These include:
5.1 The accuracy
Accuracy is paramount in healthcare applications, especially for tasks like diagnostic imaging and predictive modeling. AI systems can often surpass human experts in detecting abnormalities (Esteva et al., 2017). However, the performance of AI models is typically evaluated using historical data, split into training and testing sets. This method, though widely used, does not always reflect real-world performance (Baeza-Yates, 2022). The trade-off between model complexity and accuracy is an ongoing challenge.
5.2 The interpretability
Interpretability in AI refers to the degree to which a human can understand and trust the decisions made by AI systems. In the context of healthcare, interpretability is paramount because medical professionals rely on these systems to make critical decisions about patient care. Without a clear understanding of how AI systems arrive at their conclusions, healthcare providers may be reluctant to adopt these technologies or may misinterpret their recommendations, potentially leading to adverse patient outcomes (Ennab and Mcheick, 2022b). Techniques like activation visualization enable us to infer which features in an image are represented by each feature map within every layer of the trained convolutional neural network. When activation visualization is needed, it involves inputting sample medical images into the convolutional neural network for analysis and observing the patterns in the results (Kim, 2023).
5.3 The interpretability-accuracy trade-off
As highlighted (Ennab and Mcheick, 2022a), there is often a trade-off between interpretability and accuracy. While simpler models like decision trees are more interpretable, they may not achieve the same level of accuracy as more complex models, such as deep neural networks (Escalante et al., 2018).
6 AI models in healthcare
The field of medicine has employed AI technology to automate various stages of clinical research, offering valuable assistance for clinical decision-making (Wu Yazhou and Chen, 2022). Utilizing AI techniques across diverse medical domains brings advantages such as enhancing diagnostic accuracy and reducing both time and labor requirements. Leveraging the latest AI advancements, typical applications experiencing revolutionary changes include intelligent screening, precise diagnosis, risk prediction, and supportive therapy (Khalifa et al., 2024) as shown in Figure 2.
6.1 Conversational AI
Conversational AI is a technology that leverages conversation-oriented techniques, facilitating dynamic interactions and widespread engagements across various users and platforms (Kaushik, 2023). It refers to a set of technologies that communicate with humans using “natural language” through hardware, software, etc., by combining natural language processing (NLP), ML, DL, and conversation processing technologies into a single platform (Han et al., 2023). Within the healthcare sector, conversational AI is facilitating several significant use cases that have an impact on both healthcare providers and patients. In healthcare, five distinct use cases have been recognized:
(1) Disseminating Information: AI assistants can engage in conversational interactions to provide answers to frequently asked questions (FAQs) concerning specific illnesses, health conditions, or pandemics (Haptik, 2023). It has the potential to increase awareness of a specific health-related issue or disaster by providing quick access to accurate, trustworthy, and timely information, e.g., Dr. LalPathLabs (Gottlieb et al., 2023).
(2) Scheduling appointments: Conversational AI enables patients to easily arrange and reschedule medical appointments (Haptik, 2023).
(3) Patients care: Healthcare organizations have effectively used AI Assistants to automate the answer to frequently asked questions and the resolution of regular, repetitive chores and diagnostics chains (Haptik, 2023).
(4) Managing and tracking patients’ health: conversational AI aids individuals in monitoring their physical wellbeing by detecting symptoms at an early stage and facilitating online consultations with healthcare professionals when necessary (Haptik, 2023). Additionally, patients are equipped with the resources and information required to manage their own health effectively (Han et al., 2023).
(5) Enhancing Patient Assistance: conversational AI fosters a data-driven approach within healthcare, empowering patients and caregivers with the necessary information for informed decision-making (Haptik, 2023).
6.2 Intelligent screening
AI technology has been applied in the screening of various malignant tumors, and it can automatically screen the beginning and malignant areas of suspected cancer. However, it should be noted that the accuracy of the model has a major impact on the clinical decision-making of physicians (Chen et al., 2024). When the model’s predictions are inaccurate, the effectiveness of its supplementary screening is frequently significantly diminished. Moreover, for diseases with low occurrence rates and limited sample sizes, the presence of false positives cannot be disregarded, making it advisable to conduct manual reviews for verification. Consequently, considerable challenges persist in the application of AI models in clinical settings, and it is essential to factor in the potential adverse repercussions of model-assisted screening during the design of AI tools. Furthermore, it is noteworthy that all AI systems operated as black boxes without real-time interpretability techniques (Wu Yazhou and Chen, 2022).
6.3 Screening for digestive tumors
Kiani et al., (2020) developed a deep learning system for liver pathology image analysis; hence, the system can be self-employed in the screening of both hepatocellular carcinoma and cholangiocarcinoma. When tested against the validation set, this system has a performance accuracy of 88.5% and boasts the same accuracy of 88.5% for the independent test set. A similar system was developed by Sinha et al., (2024), which carried out the analysis of colonoscopy images with the help of machine learning. Precisely, the system detects and differentiates the adenomas, which need to be extracted, from nonneoplastic polyps, while having a predictive accuracy of 98.1%. A similar artificial intelligence (AI)-based deep learning system was developed by Wang et al., (2020) for colonoscopy picture analysis; the study showed that the Adenoma Detection Rate (ADR) increased tremendously after AI application compared with that in the traditional group. Therefore, it can be said that with this system, the polyp and tumor detection rate of colonoscopy have been well increased. On the other hand, Chen et al. (2020) and Wu L (2019) have developed a deep learning-based image analysis system for esophagogastroduodenoscopy (EGD), which can classify the stage of duodenal diseases. This system leads to an excellent reduction of the blind spot’s missed diagnosis rate, which is 5.9% and 3.4% missed rate.
6.4 Screening for other tumors
In a study by McGenity et al. (2024), an innovative deep learning-driven system for analyzing slice pathological images was introduced. This system demonstrated the capability to automatically diagnose and classify breast cancer. Impressively, the overall accuracy rate achieved a notable 83.1% when compared to pathological results, which are considered the gold standard. Whereas, Hossain et al. (2024) presented a chest CT processing system based on deep learning, achieving a sensitivity rate of 91% for metastasis detection, thereby enabling the automated screening of metastatic breast cancer. Mohamed et al. (2022) introduced a fully automated breast cancer detection system. Initially, the U-Net network is employed to automatically extract and segregate the breast area from the surrounding body, which is considered noise in the breast cancer detection model. This yielded an accuracy of 99.33%, sensitivity of 100%, and specificity of 98.67%. Lotter et al. (2021) introduced an annotation-efficient DL approach that attains state-of-the-art performance in fields like mammogram classification, showcasing average sensitivity enhancements of up to 14% for AI methods compared to mammography experts. Yoo et al. (2018) presented a deep learning-based ultrasound image analysis system, which elevated the screening sensitivity for thyroid cancer from 84% to 92.0%, and successfully achieved automated screening for both benign and malignant thyroid nodules. Masood et al. (2018) Constructed a lung CT image processing system using the Internet of Medical Things (IoMT) and DL, which successfully predicted the malignant transformation stage of pulmonary nodules, achieving a classification accuracy rate of 84.6%. Wu Yazhou and Chen (2022) and Zadeh Shirazi et al. (2021) employed a Deep Convolutional Neural Network (DCNN) as a semantic segmentation model for segmenting seven distinct tumor regions, achieving a segmentation accuracy of 70%.
6.5 Screening for eye diseases
Hassan et al. (2024) utilized DL techniques to process retinal images, enabling the automated screening and severity grading of diabetic retinopathy. The AI method exhibited a sensitivity and specificity of 100% and 88.4%, respectively, for diagnosing severe lesions. With the ability to obtain the sensitivity of 85.2% and specificity of 92% for lesions. Archana and Jeevaraj (2024) developed a DL-based system for processing the images of the eye, and it provided an outstanding AUC-ROC from 99.3% to 99.7% for cataract classification. This aids in conducting automated screening and collective management of cataract-affected persons. Lombardi et al. (2021) proposed a new deep learning system to predict brain age through native resting-state scalp EEG raw data and reached an accuracy of 89.7%. Wu et al. (2018) found in a study on the diagnosis of fungal keratitis, that automatic hyphae detection technology is sensitive at 89.3% and specific at 95.7%. The AUC value of the ROC curve is 94.6%, which is timely, accurate, objective, and quantitated to provide evaluation criteria for fungal keratitis.
6.6 Intelligent diagnosis
AI systems within healthcare provide patients with reliable and tailored services that go beyond traditional testing. They effectively assist patients in overseeing their individual wellbeing, tracking medical conditions, delivering treatments, offering psychological guidance, and handling dietary considerations. This holistic support contributes to extending patient longevity and enhancing their overall quality of life (Origimid, 2022). However, the effectiveness of the AI model is limited by the size of the training dataset. A model trained on one dataset may not perform optimally when applied to a different dataset with different characteristics. Therefore, it is important to incorporate an external test dataset during model training to evaluate its ability to generalize. Additionally, while many AI-driven diagnostic methods focus on analyzing medical imaging data, it is essential to recognize that clinically meaningful research outcomes require a holistic assessment of various patient indicators by healthcare professionals. Therefore, future research efforts should prioritize the comprehensive use of diverse clinical data to improve the effectiveness and wide-ranging applicability of AI models (Data sharing in the age of deep learning, 2023).
6.7 Diagnosis of infectious diseases
The outbreak of novel coronavirus disease (COVID-19) provides a suitable application opportunity for AI technology. AI technology has made a significant progress in the diagnosis, typing, risk prediction, and adjuvant treatment of COVID-19 (Ennab and Mcheick, 2022b). Shorfuzzaman et al. (2021) introduced a deep learning fusion framework that incorporates the principles of transfer learning, achieving intelligent diagnosis of COVID-19 patients with a classification accuracy of 95.5%. Quiroz et al. (2021) confirmed that the ML approach can be employed to automatically assess the severity of COVID-19, aiding in the classification and diagnosis of COVID-19 patients with an accuracy rate of 96.0%, a sensitivity of 84.5%, and a specificity of 92.9%. This enables the prioritization of follow-up diagnosis and treatment. Salem et al. (2023) developed a machine and deep learning system to identify metabolites and clinical features linked to gallstone disease. Pang et al. (2019) developed an innovative YOLOv3-arch model specifically designed to identify cholelithiasis and classify gallstones in CT images. This model significantly enhanced accuracy, achieving a 92.7% accuracy rate for identifying granular gallstones and an average accuracy of 80.3% for identifying muddy gallstones. However, it is important to note a major limitation of these models: healthcare workers cannot predict COVID-19 in patients with other underlying conditions such as COPD, hypertension, asthma, and more (Ennab and Mcheick, 2022b).
6.8 Diagnosis of medical diseases
As medical imaging technology advances and clinical diagnosis accuracy improves, clinical diagnostic approaches employing DL technology have experienced significant growth (Ennab and Mcheick, 2022b). Arbabshirani et al. (2018) and Kim et al. (2024), each constructed a DL-based brain CT image processing system, with accuracy values of (73%) and (84.6%), realizing automatic detection of acute neurological events such as stroke. Lo et al. (2021) suggested an automated ischemic stroke diagnosis method using deep DL, achieving a sensitivity of 98.1%, specificity of 96.9%, and an accuracy rating of 99.3%. This method effectively offers clinicians valuable diagnostic recommendations for acute ischemic stroke cases. Bibi et al. (2020) created a system that combines DL and the IoMT to achieve swift and secure identification and classification of leukemia, boasting an impressive average accuracy rate of 99.6%. This system facilitates real-time communication between healthcare professionals and patients regarding leukemia detection, diagnosis, and treatment, effectively conserving clinicians’ time and efforts. Jing et al. (2023) developed a DL system aimed at predicting the likelihood of recurrence and disability outcomes in patients who have experienced a transient ischemic attack or ischemic stroke. Eckardt et al. (2022) created a deep learning model specifically tailored for detecting NPM1 mutation and forecasting the presence of acute myeloid leukemia, attaining an impressive accuracy level of 95%. Bukhari et al. (2022) developed a deep learning framework for identifying leukemia cancer in microscopic blood samples by incorporating squeeze and excitation learning techniques, achieving an accuracy rate of 89.72%. Hamedan et al. (2020) employed Expert Systems (ES) to assess chronic kidney disease, integrating AI technology with artificial expert judgments. The results indicated that the ES significantly outperformed all other models in performance for predicting chronic kidney disease when the accuracy of the ES was 92.1%, sensitivity was 95.4%, and specificity was 88.9%.
6.9 Diagnosis of surgical diseases
Deep learning image knee analytical system was designed for knee injury diagnosing, e.g., injured knee estimation, such as anterior cruciate ligament and meniscus tears by Dratsch et al. (2024). Deep learning models in Siouras et al. (2022) related to MRI-associated knee injury estimation with accuracies ranging from 72.5% to 100%. The processing system based on deep learning for X-ray images was developed by Lindsey et al. (2018) to detect and localize fractures. After they integrated AI technology, the sensitivity in fracture detection of the clinicians improved from 80.8% to 91.5%, and the specificity increased from 87.5% to 93.9%, the CT image analysis system, designed by Fu et al. (2019), intended for automatic diagnosis and visualization of interfemoral fractures by identifying the most probable area of the fracture.
6.10 Adjuvant therapy
The goal of is to aid individuals with early invasive cancer in accurately predicting the personalized survival benefits, allowing them to make informed treatment decisions, without supplanting conventional treatments (Zheng et al., 2024). Adjuvant therapy relies on machine learning models and has the potential to enhance patient survival (Howard et al., 2020). Machine learning techniques can accurately predict patient outcomes under different treatment regimens by modeling the complex interactions between risk factors in a data-driven manner (Alaa et al., 2020).
6.11 Treatment decision support
Radiation therapy plays a pivotal role in the treatment of various tumors (Frisch et al., 2024). Within the treatment process, it is imperative to precisely outline the organ at risk (OAR) to provide guidance for radiotherapy and predict the patient’s prognosis (Hall et al., 2022). Bird et al. (2021) constructed ML model using a multi-center dataset, with the goal of establishing precise and broadly applicable radiotherapy protocols for rectal cancer. Through rigorous validation studies and regulatory approvals, this treatment decision support system mitigates prescription errors and augments existing error warning mechanisms (Yang et al., 2023). Rawson et al. (2021) constructed machine learning-driven systems for antimicrobial prescription decision-making, offering clinical decision support in antibiotic management. The AI-generated prescription recommendations have achieved a level of accuracy similar to that of clinicians. Bozyel et al. (2024) introduced an ML-based prescription recognition system to enable automated early detection and correction of prescription errors in heart disease patients, achieving a clinical effectiveness rate of 85%. Hooshmand et al. (2021) utilized DL techniques to explore potential COVID-19 drug candidates that exhibit minimal side effects and promising efficacy. Ke et al. (2020) employed DL methodologies to pinpoint existing drugs with the potential to combat COVID-19, ultimately discovering over 80 promising candidates for coronavirus treatment.
6.12 AI systems for medical surgical robots
Surgical robot technology stands as the foundational element of Computer Integrated Surgery (CIS). It brings quantification to the complete medical process, spanning from diagnosis and procedural planning to surgery, recovery, and observation (Lee et al., 2024). By integrating robotics and IT technology into the medical domain, it aims to enhance surgical procedures, making them safer and more efficient through rigorous objectification (Reddy et al., 2023). Surgical robots typically fall into two primary categories. This is identified to be the first category where robot systems are developed, which are designed to operate on body cavities such as the abdomen, chest, and lumbar regions. An illustrative example is the da Vinci robot system, which conducts surgery using a master-slave approach relying on 3D imaging (Biswas et al., 2023). The second category pertains to surgical robot systems designed for addressing lesions within the brain, bones, or muscles. These systems create a 3D model of the patient’s lesion using pre-existing medical images. The surgery planning, therefore, becomes a pointer of other factors like the lesion location, lesion condition, image information regarding the lesion, and the position of the patient in relation to the operating table, along with the 3D patient coordinate system realized before surgery (Chiou et al., 2022). Employing surgical robots yields several advantages, including reduced surgery duration, lowered risk of bleeding and infection for patients, and decreased fatigue and hand tremors for medical staff during procedures. Besides, it has benefits to hospitals by the patients having short stay lengths in hospital, therefore, increasing bed turnover, and positive patient outcomes through minimally invasive surgical technologies (Lee et al., 2024). At the moment, IR is used in broad clinical fields, such as orthopedics, biliary tract, throat, and liver surgeries, to mention but a few (Gumbs et al., 2021; Zhang and Zhang, 2023) used IR in spinal surgery in whose application brings out the accuracy in screw positioning and uses a few intraoperative fluoroscopies and has a low incidence in postoperative complications. Xie et al. (2021) used da Vinci surgical system in managing biliary cysts in infants below 1 year and brought out that the use of IR is safe in this clinical application. Arora et al. (2024) performed robotic surgery, in this case, transoral, for the extraction of masses in the throat, which was well exposed visually and that one encountered no major adverse reactions. Serednicki et al. (2023) used IR in liver surgery, and one enjoys the benefits of low blood loss and few adhesions, therefore short stay days in hospital and the duration for postoperative recovery is short.
7 The interpretability models
With the widespread adoption of AI solutions in healthcare, directly understanding models’ working mechanisms and opening the black box has become increasingly important (Pavlidis, 2024). Building confidence in machine learning models is now necessary for fully adopting AI systems. Thus, model interpretability is highly sought-after, especially in high-risk domains requiring dependability like healthcare (Ennab and Mcheick, 2022a). AI models in healthcare may provide contradictory results to medical institutions, potentially having disastrous consequences across cultures and situations. Interpretable approaches can clarify how a model choice is made, allowing tracking each output result and controlling model outputs (Longo et al., 2024). Interpretable AI models enable users to query, comprehend, correct, and even advance the learning system. Overall, interpretable learning models allow end-users like doctors to assess the model before taking any further action. By justifying predictions, interpretable machine and deep learning models give users the option to reject or accept the predictions and recommendations (Salih, 2022).
7.1 Types of the interpretability models
7.1.1 Ante-hoc(Intrinsic)
The fastest way to easily interpret a model is designing it with an interpretable structure initially. A simple model like a decision tree is easy for humans to interpret by looking at its structure. A model with a simple structure was termed Intrinsic (originally equipped) because it already secured interpretability itself, also expressed as having transparency (Cao et al., 2024). The advantage of Intrinsic is explaining “how the model works.” However, due to the trade-off, intrinsic models have low accuracy (Carvalho et al., 2019).
7.1.2 Post-hoc
If the model itself lacks explanatory power, its prediction results must be interpreted post hoc. Most interpretability techniques in machine and deep learning are post hoc. Ideally, the model would have high accuracy and explanatory power, but this is rare in reality (Cao et al., 2024). Complex high-performing models are commonly used, with post hoc analysis done. Post-hoc examples are explained with the viewpoints introduced later (Moradi and Samwald, 2021).
7.1.3 Global
The global technique describes all model predictions based on understanding its logic, or at the module level, scopes describing how one module affects predictions (Carvalho et al., 2019).
7.1.4 Local
Local techniques describe specific decisions or predictions. A range describing a prediction group by bundling predictions is also a local technique. Compared to global methods, local techniques have a smaller scope to explain, making them relatively feasible and inexpensive. Additionally, even if the overall prediction trend is unexplained, one or a few predictions can be described nearly perfectly (Moradi and Samwald, 2021).
7.2 The interpretability models
The categorization of the interpretability models may vary slightly depending on specific implementations and use cases as shown in Table 2.
7.2.1 The interpretability-based model using the relative weight of the image features
Ennab and Mcheick (2022a) presented an interpretability-based model using statistics and probability principles to train datasets by determining the relative weights of variables indicating their respective significance in predicting and estimating disease likelihood. The variables are either patient symptoms or characteristics of injured organ regions in medical images. Dividing each variable’s weight by the sum of all weights gives the relative weights. Training the dataset determines infection likelihood. Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022).
7.2.2 LIME (local interpretable model-agnostic explanation)
LIME is a model-agnostic technique explaining which features are most important in a feature space region (Shi et al., 2020). LIME’s core idea is computing a Local Surrogate Model in a region of interest, which is an easily interpretable model like linear or decision tree trained to mimic a complex model’s behavior (Garreau and von Luxburg, 2020). For an explanation, LIME creates new similar data points with slightly altered values. Feeding the perturbed points into the complex model reveals relationships between perturbed features and predictions, captured by the surrogate model (Garreau and von Luxburg, 2020).
7.2.3 SHAP
SHAP uses game theory to measure each attribute’s impact on the prediction process. The Shapley value evenly divides advantages among contributing parties (features) when contributions are unequal (Shapley, 2016). In other words, Shapley values are based on features interacting to influence predictions toward a value. It attempts to evenly distribute contributions across all feature subgroups. Specifically, the Shapley value uniformly distributes the difference between the prediction and average prediction among the instance’s feature values needing explanation (Strumbelj and Kononenko, 2011). SHAP values provide a unique additive feature importance measure satisfying attribution features (local accuracy, missingness, consistency). These features represent intuitive rules for determining the final prediction, translatable to the machine learning problem. However, direct Shapley value computation requires efficient computation, needing to check every permutation combination (Mitchell et al., 2022).
7.2.4 Grad-CAM
Class Activation Mapping (CAM) is a technique widely used in computer vision to create visual representations displaying contributions of different image regions to neural network predictions (Selvaraju et al., 2017). This produces a heatmap-resembling image with each pixel signifying the activation level for a class. Pixel values range from 0 to 1, typically shown as a 0 to 255 grayscale image. Higher scores indicate regions in the original image having a stronger influence on the network’s response or prediction. Overlaying CAM onto the original image enhances visual appeal and informativeness. Unlike CAM requiring model structural changes and retraining, Grad-CAM utilizes pretrained weights to backpropagate gradients to the desired parameter layer (like convolutional) when predicting an image (Schöttl, 2020). This yields a gradient matrix with identical dimensions to the parameter layer’s output feature map. By globally average pooling the gradient matrix across spatial dimensions, a vector with equal length to the feature map’s channel count emerges. This vector contributes to weighing the diverse feature map channels, ultimately creating a heatmap visualization. As Grad-CAM avoids model architectural adjustments and retraining, it offers a more flexible, efficient CAM alternative (Liu et al., 2023). Conventionally, a neural network’s classification module uses a fully connected model processing extracted features, converting them into class probability scores via a softmax layer. The highest scoring class then dictates the ultimate prediction. Grad-CAM takes a different approach by not just discriminating between classes but also pinpointing relevant image regions. This is achieved by exploiting gradients (derivatives) from the final convolutional layer’s feature map (Vinogradova et al., 2020). These gradients serve as a tool to emphasize important areas significantly impacting the eventual prediction.
7.2.5 Bayesian nonparametric model
Guo et al. (2018) created a Bayesian nonparametric approach to build a parameter space with infinite dimensions. In other words, model size can fluctuate in response to data increases or decreases, determined by the number of data parameters used. It requires few assumptions to learn data and perform clustering. Growing data can also be continuously aggregated into proper classifications, this model also makes predictions concurrently. A spatial data model comprises all issue-related properties manageable based on the unique learning problem (Ribeiro et al., 2018). The core of the non-parametric Bayesian statistical model is setting the data probability distribution function to an arbitrary, flexible distribution rather than a specific parametric one and placing a prior distribution on this to perform posterior inference. In Bayesian statistical models, non-parametric modelling allows flexible modelling of data distribution, random effect distribution, or parameter of interest prior distribution, presenting a wider probability model class. Thus, the prior distribution for random distribution is key in nonparametric Bayesian statistical modelling, with the probability most frequently used (Noh et al., 2014).
7.2.6 GAM
Kraus et al. (2024) introduced GAM, a generalized additive global variable weight technique considering neural network swarm forecast patterns. GAM’s global interpretation describes the neural network’s non-linear representation. GAM also enables modifying subpopulation granularity and tracking global interpretations for particular samples. In statistics, a generalized additive model (GAM) is a linear model where predicted variable values are the aggregation of several unknown smooth functions defined for the predictors. The purpose is inferring a predictor smooth function whose aggregate composition approximates it. This structure is easily interpretable, allowing the user to see each variable’s importance, i.e., its effect on the predicted output via its function (Linardatos et al., 2021).
7.2.7 MAPLE
The key difference between using MAPLE as a black-box model explanation versus a predictive model is fitting MAPLE to the black-box model’s prediction in the first case and the response variable in the second. Since MAPLE is a very accurate predictive model providing correct predictions, it avoids trading off performance and interpretability. It finds global trends using local examples and explanations. MAPLE differs from other frameworks by its training distributions (El Shawi et al., 2019). Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022). Maple alters tree ensembles to provide local explanations that detect global trends and example-based explanations. It uses the ensemble to determine the most important training points for a new prediction, building a linear model from those points for prediction and local explanation (Plumb et al., 2018).
7.2.8 Anchors
Anchors is a model-independent, rule-based local explainer approach (Ribeiro et al., 2018). Anchors ensure that projections from the same anchor are roughly equal. Anchors, in other words, identify the features that are sufficient to correct the forecast while adjusting the others that do not influence the prediction. The bottom-up strategy, in which anchors are constructed sequentially, is one type of anchor building (Ribeiro et al., 2018). Data were collected as previously described in Mohammad Ennab and Hamid Mcheick (22 October 2022). Anchor describes individual predictions for black-box classification models by finding decision rules that sufficiently “anchor” the predictions. A rule freezes a prediction if changes in other feature values do not affect the prediction. Anchor leverages reinforcement learning techniques with graph search algorithms to reduce the number of model calls (required running time) to a minimum while still being able to recover from the local optimization (T, 2020).
7.2.9 Perturbation-based methods
Perturbation is the most basic technique for assessing input property modifications on output. This involves eliminating, masking, or altering specific inputs, then forward passing output and comparing to the original output. This is similar to the sensitivity analysis performed in parametric control system models. The input features that have the greatest impact on the outcome are prioritized. Because a forward pass must be performed after perturbing each set of features in the input, it is computationally intensive. In the case of picture data, the perturbation is accomplished by covering areas of the image with a grey patch, thereby obscuring them from the system’s view (Singh et al., 2020). In images, disturbance is done by covering areas with a grey patch, concealing them from the system’s view (Ennab and Mcheick, 2022a). The perturbation-based methods are broadly divided into input sampling-based methods and input optimization methods. RISE (Randomized Input Sample for Explanation) is the most representative algorithm among input sampling-based methods that put a random mask on the input image and saves the AI output. After around 8,000 repetitions, a linear combination of random masks and AI output values is performed. In this way, random masks made with high output values show consistency in feature exposure, allowing for a proper explanation (LG AI Researcher, 2021). However, the problem with this approach is that it requires about 8,000 outputs to illustrate. Another limitation is that the results are different each time it is performed on the same input image because a random mask is used (LG AI Researcher, 2021). The most representative algorithm in the latter input optimization method is an algorithm called extremal perturbation, which finds feature parts showing high output values by the optimization method. The problem with this approach, however, is that it relies only on the optimization method. When an accurate solution is obtained, a very interpretable explanation can be obtained, but when an exact solution is not obtained, features that seem completely unrelated may be displayed. In addition, since it uses a numerical optimization method, it has a computation time problem (LG AI Researcher, 2021).
7.2.10 Attention based
The fundamental concept of attention is motivated by how people pay attention to various areas of an image or other data sources in order to interpret them. The technique employed attention mechanisms, which included an image model and a language model, to show the detection process (Li et al., 2021). The interpretation in attention-based involves combining mechanisms for the selective traits of dominant features with attention towards the reported trait. One part of this will be selecting some of the hidden states over the time steps and on top of that, adding an attention layer to the present deep learning model. An attention score is computed at each important feature or time sequence to denote its importance. The attention mechanism found in the language model was used to learn the mapping between diagnostic reports and sights (Singh et al., 2020).
7.2.11 Testing concept activation vectors
In Kim et al. (2017), Concept Activation Vectors (TCAV) is another innovative approach used to explain the acquired characteristics of successive layers in terms of human-understandable concepts to domain experts without deep learning comprehension. It uses the network’s directional derivative in concept space in a similar way that saliency maps use input feature space. The directional derivative of the network in concept space is treated as a saliency map in TCAV. In this way, TCAV is perfectly suitable as the approach for detection of microaneurysms and aneurysms in the retina when saliency maps are used for the classification of the Diabetic Retinopathy (DR) level prediction and provides understandable reasoning to the physician for the level of DR. This provides justification whether a conceptual or physical structure is present in the image (Singh et al., 2020). However, many clinical concepts in medicine, such as structural texture or tissue shape, cannot be fully described by TCAV directly to prove their existence or non-existence (Kim et al., 2018). Many clinical concepts, such as form, texture, or shape, cannot be effectively described in terms of presence or absence and require a continuous scale of assessment (Ennab and Mcheick, 2022a).
7.2.12 Similar images
Stano et al. (2020) proposed research assessing layers of a 3D-CNN using a Gaussian mixture model (GMM) and binary encoding of training and test pictures based on their GMM components to offer explanations for comparable 3D images. As an explanation for its result, the software returned activation-wise similar training pictures utilizing the atlas. The model includes a perceptual code in binary vector space that defines the CNN’s processing of the input sample (Kim et al., 2018). A collection of the most perceptually similar and dissimilar samples may be retrieved from an existing atlas of labelled samples in order to support and further explain the choice made by the CNN model by calculating distances between pairs of samples in this perceptual encoding space. Applications of this model include Computer-Aided Diagnosis software using Computed Tomography (CT) data from medical imaging tests (Stano et al., 2020). The same imaging model was carried on the 3D MNIST datasets and Magnetic Resonance Imaging (MRI) datasets, and the findings were congenial atrophy conditions. There is an indication in some of the cases that the breakthrough of similarity with activation is invariant to the picture spatial orientation, and impact may be there in choices related to pictures returning (Singh et al., 2020).
7.2.13 Textual justification
The justification for the choice was given in the form of words or phrases and, on general, can communicate directly to both expert and non-expert users (Singh et al., 2020). A diagnostic phase and visual heatmaps for the breast-mass classification was developed using an explanation model that took input from the visual features of a classifier as well as the prediction embeddings (Lee et al., 2018). Justification generator was trained to generate justifications in the existence of a small number of medical reports using a visual word constraint loss (Singh et al., 2020). Data were taken as described earlier in Mohammad Ennab and Hamid Mcheick (22 October 2022). A diagnostic network and a justification generator are the two parts of the overall design of the model. Any general CADx network (classifier of malignant mass and benign mass) can be used as the diagnostic network. A visual characteristic and a diagnosis made by the diagnosis network are used by the justification generator. A visual word constraint loss is developed in the training stage to efficiently train the justification generator by preventing the training set’s sentences from being duplicated (Lee et al., 2019).
7.3 Characteristics of the interpretability models
It is important that when choosing the interpretation models to implement, many-paged features should be carefully and perhaps realized that no one fits all solution may exist:
1. Model complexity vs. model interpretability: Some interpretability method simplifies complex models to be interpretable. Still, in that simplification, the majority of the time, a tradeoff between model accuracy and interpretability are made (Johansson et al., 2011).
2. Suitability to different model types: not all interpretability method has a similar level of suitability for different kinds of machine learning models. Some work well with large deep neural networks, while others work better for decision tress or linear (Escalante et al., 2018).
3. Local explanations: Most of the interpretation methods aim to present local explanations: for each prediction, the user can get an idea of which features that have played into account for that particular output. That is to say that these methods do not aim to have a global perception of the model’s behavior (Linardatos et al., 2020).
4. Consistency: The interpretability methods can explain why the explanations are consistent depend on the way the input samples are chosen and hence might produce different explanations for marginally different sampled data points (Hu et al., 2019).
5. Discrete explanation: Some of them offer differences in terms of interpretation where they provide interpretations in a discrete or rule-based way; therefore, the fine-grained nature of the original model might miss this kind of methods (Gilpin et al., 2018).
6. Interpreting complex models: On the other hand, the interpretation of complex models like deep neural networks might be quite troublesome, and yes, the explanation might be done (Carvalho et al., 2019).
Table 3 summarizes the limitations associated with various interpretability models. Each model offers distinct advantages and challenges, with trade-offs between local and global explanations, computational overhead, and applicability across different architectures. While some models (like LIME and SHAP) provide local insights at the cost of consistency, others (like Grad–CAM and Bayesian Nonparametric) are tailored for specific use cases, such as convolutional neural networks. The table highlights how different interpretability methods suit various needs but may have constraints related to stability, complexity, and applicability beyond specific models or domains.
7.4 Assessment of the interpretability models
The interpretability models can mainly be classified under the qualitative and quantitative models. The models that are applied to the quantitative evaluation are as follows:
1. KAR method (keep and retrain): the modularity of the analysis on how the removal of the least significant N % pixel features in the saliency map is done in respect of the change in retrained model accuracy (Kim et al., 2019).
2. ROAR method (remove and retrain): the removability analysis of how the accuracy of the retrained model is affected when excluding the most important N % pixel features in the saliency map (Kim et al., 2019).
The following are used as measures of qualitative evaluation:
1. Coherence: This means that the input pattern that is closely related to the prediction given by the interpretability method needs to have an attribute that is somehow (Gilpin et al., 2018).
2. Selectivity: in the case that input image, the exclusion of pixels that are rated important in the saliency map by the method, this, in turn leads to reduction of related probability that is related to the model prediction category that corresponds (Fuhrman et al., 2022).
3. Implementation invariance: In the case of two models, for explanation, being similar, that is to say, two models that provide the same input produces the same output, then this, in turn means that interpretability method should give the same for two models (Carvalho et al., 2019).
4. Class sensitivity: The explanation that is generated by interpretability method should be sensitive to the category (Nielsen et al., 2022).
5. Explanation continuity interpretability methods should give similar explanations to similar input (Longo et al., 2024).
Figure 3 explains the categorization of the interpretability models into quantitative and qualitative methods. The quantitative models (KAR and ROAR) focus on performance changes with model retraining based on feature importance. In contrast, qualitative models (such as Coherence, Selectivity, Implementation Invariance, Class Sensitivity, and Explanation Continuity) assess interpretability based on the behavior and consistency of explanations across different inputs and predictions.
Figure 3. Assessment of interpretability models categorized into quantitative and qualitative evaluation methods.
7.5 Recapitulation of main points
The application of AI in healthcare has demonstrated immense potential in automating diagnosis, treatment decisions, and patient monitoring. However, several key challenges persist:
• Advantages: AI systems have shown remarkable accuracy in tasks like diagnostic imaging, risk prediction, and disease screening. The ability of deep learning models to process large datasets quickly and accurately has outperformed traditional methods in several areas (e.g., cancer detection and diagnostic imaging).
• Limitations: The black-box nature of many deep learning models makes it difficult for healthcare providers to trust and interpret their decisions. The lack of transparency and interpretability remains a critical barrier to widespread adoption. Additionally, AI models often struggle to generalize across diverse patient populations, leading to reduced performance in real-world clinical settings.
• Challenges: Another significant challenge is the computational cost and real-time applicability of certain models. Many AI systems require significant computational resources, which limits their deployment in resource-constrained healthcare facilities. Furthermore, the integration of uncertainty quantification with interpretability models remains an underexplored area that could provide more reliable and explainable AI systems.
8 Lessons learned and future directions
Through our systematic review, several trends and common issues have emerged:
• Lack of real-time interpretability: Most high-accuracy AI models operate as black boxes, which limits their clinical applicability.
• Trade-offs between accuracy and interpretability: Models that prioritize accuracy often do so at the expense of interpretability, and vice versa.
• Generalizability issues: The majority of studies focus on specific datasets, making it difficult to generalize their findings to broader clinical populations.
8.1 Future directions
To address these challenges, future research should focus on developing hybrid models that balance accuracy and interpretability. Additionally, incorporating uncertainty quantification methods can improve the reliability of AI models in healthcare. Additionally, as AI continues to evolve, several key areas require further investigation:
• Improved Interpretability: Future research should focus on developing models that balance both accuracy and interpretability. Techniques such as Local Interpretable Model-Agnostic Explanations (LIME) and SHAP (Shapley Additive Explanations) could be further explored to enhance the transparency of deep learning models. Additionally, more effort is needed to integrate uncertainty quantification techniques to improve model reliability in real-time clinical settings.
• Incorporation of Multimodal Data: AI systems that can process diverse types of clinical data, such as medical imaging, patient history, and genetic information, are likely to provide more comprehensive and accurate predictions. Research should explore how multimodal AI systems can improve patient outcomes by integrating various data sources.
• User-Centered Design: A key challenge in AI healthcare applications is the need for user-centered design. Future research should explore how AI models can be developed in collaboration with healthcare providers to ensure that they are user-friendly, interpretable, and aligned with clinical workflows.
• AI in Personalized Medicine: While this review focused primarily on diagnostic imaging, future research should explore how AI can be applied to other domains, such as genomics and personalized medicine. AI models that can predict patient-specific outcomes and suggest personalized treatment plans based on genetic data have the potential to revolutionize healthcare.
• Learning-Based Manipulation: In the context of learning-based manipulation, as discussed in Stefanelli et al. (2023), future research should investigate how AI systems can integrate multiple sensory inputs (e.g., force, temperature) to mimic human-like responses in medical robotics and prosthetics. Testing and refining these methodologies in clinical scenarios could lead to more adaptive and responsive AI-driven prosthetic systems.
9 Conclusion
This review systematically examined AI models in healthcare, focusing on the trade-offs between accuracy and interpretability. We highlighted the most commonly used models and metrics, and identified key challenges, such as the black-box nature of deep learning models and generalizability issues. Our findings suggest that future AI research should prioritize transparency and safety, particularly in high-risk healthcare applications. By addressing these issues, healthcare providers can develop AI solutions that not only enhance performance but also build trust among stakeholders, ultimately leading to safer and more effective patient care. Additionally, incorporating strategies like those presented by Stefanelli et al. (2023), which integrate force and temperature information in prosthetic control, could offer new avenues for enhancing AI systems’ human-like capabilities in healthcare.
Author contributions
ME: Data curation, Formal Analysis, Resources, Validation, Writing–original draft, Writing–review and editing. HM: Conceptualization, Supervision, Writing–original draft.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Afrash, M. R., Shafiee, M., and Kazemi-Arpanahi, H. (2023). Establishing machine learning models to predict the early risk of gastric cancer based on lifestyle factors. BMC Gastroenterol. 23 (1), 6. doi:10.1186/s12876-022-02626-x
Alaa, A., Gurdasani, D., Harris, A., Rashbass, J., and Schaar, M. v. d. (2020) Machine learning to guide the use of adjuvant therapies for breast cancer. doi:10.21203/rs.3.rs-53594/v1
Arbabshirani, M. R., Fornwalt, B. K., Mongelluzzo, G. J., Suever, J. D., Geise, B. D., Patel, A. A., et al. (2018). Advanced machine learning in action: identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. NPJ Digit. Med. 1, 9. doi:10.1038/s41746-017-0015-z
Archana, R., and Jeevaraj, P. S. E. (2024). Deep learning models for digital image processing: a review. Artif. Intell. Rev. 57 (1), 11. doi:10.1007/s10462-023-10631-z
Arora, A., Faulkner, J., Paleri, V., Kapoor, K., Al-Lami, A., Olaleye, O., et al. (2024). New robotic platform for transoral robotic surgery: an IDEAL stage 0 study. BMJ Surg. interventions, & health Technol. 6 (1), e000181. doi:10.1136/bmjsit-2022-000181
Baeza-Yates, R. (2022). “Ethical challenges in AI,” in WSDM ’22: The Fifteenth ACM International Conference on Web Search and Data Mining, Virtual Event, Tempe, AZ, USA, February 21–25, 2022 (ACM), 1–2. doi:10.1145/3488560.3498370
Bibi, N., Ud Din, I., Sikandar, M., Almogren, A., and Ali, S. (2020). IoMT-based automated detection and classification of leukemia using deep learning. J. Healthc. Eng. 6648574, 6648574. doi:10.1155/2020/6648574
Bird, D., McCallum, H., Nix, M. G., Teo, M., Gilbert, A., Casanova, N., et al. (2021). Multicentre, deep learning, synthetic-CT generation for ano-rectal MR-only radiotherapy treatment planning. Radiother. Oncol. 156 (1), 23–28. doi:10.1016/j.radonc.2020.11.027
Biswas, P., Sikander, S., and Kulkarni, P. (2023). Recent advances in robot-assisted surgical systems. Biomed. Eng. Adv. 6, 100109. doi:10.1016/j.bea.2023.100109
Bozyel, S., Şimşek, E., Koçyiğit Burunkaya, D., Güler, A., Korkmaz, Y., Şeker, M., et al. (2024). Artificial intelligence-based clinical decision support systems in cardiovascular diseases. Anatol. J. Cardiol. 28 (2), 74–86. doi:10.14744/AnatolJCardiol.2023.3685
Bukhari, M., et al. (2022). “A deep learning framework for leukemia cancer detection in microscopic blood samples using squeeze and excitation learning,” in Mathematical problems in engineering. Editor N. Ali doi:10.1155/2022/2801227
Cao, S., et al. (2024). A systematic literature review on explainability for machine/deep learning-based software engineering research. ArXiv, abs/2401.14617.
Carvalho, D. V., Pereira, E. M., and Cardoso, J. S. (2019). Machine learning interpretability: a survey on methods and metrics. Electronics 8 (8), 832. doi:10.3390/electronics8080832
Chen, D., Wu, L., Li, Y., Zhang, J., Liu, J., Huang, L., et al. (2020). Comparing blind spots of unsedated ultrafine, sedated, and unsedated conventional gastroscopy with and without artificial intelligence: a prospective, single-blind, 3-parallel-group, randomized, single-center trial. Gastrointest. Endosc. 91 (2), 332–339.e3. doi:10.1016/j.gie.2019.09.016
Chen, K., Shao, A., Burapacheep, J., and Li, Y. (2024). Conversational AI and equity through assessing GPT-3’s communication with diverse social groups on contentious topics. Sci. Rep. 14 (1), 1561. doi:10.1038/s41598-024-51969-w
Chiou, S.-Y., Zhang, Z.-Y., Liu, H.-L., Yan, J.-L., Wei, K.-C., and Chen, P.-Y. (2022). Augmented reality surgical navigation system for external ventricular drain. Basel, Switzerland: Healthcare 10 (10). doi:10.3390/healthcare10101815
Data sharing in the age of deep learning (2023). Nature Biotechnology 41 (4), 433. doi:10.1038/s41587-023-01770-3
Dratsch, T., Zäske, C., Siedek, F., Rauen, P., Hokamp, N. G., Sonnabend, K., et al. (2024). Reconstruction of 3D knee MRI using deep learning and compressed sensing: a validation study on healthy volunteers. Eur. Radiol. Exp. 8 (1), 47. doi:10.1186/s41747-024-00446-0
Eckardt, J.-N., Middeke, J. M., Riechert, S., Schmittmann, T., Sulaiman, A. S., Kramer, M., et al. (2022). Deep learning detects acute myeloid leukemia and predicts NPM1 mutation status from bone marrow smears. Leukemia 36 (1), 111–118. doi:10.1038/s41375-021-01408-w
El Shawi, R., et al. (2019). “Interpretability in HealthCare A comparative study of local machine learning interpretability techniques,” in 2019 IEEE 32nd international symposium on computer-based medical systems (CBMS), 275–280. doi:10.1109/CBMS.2019.00065
Ennab, M., and Mcheick, H. (2022a). Designing an interpretability-based model to explain the artificial intelligence algorithms in healthcare. Diagn. (Basel, Switzerland) 12 (7), 1557. doi:10.3390/diagnostics12071557
Ennab, M., and Mcheick, H. (2022b). Survey of COVID-19 prediction models and their limitations. Inf. Syst. 11 (2), 14–21. doi:10.11648/j.ijiis.20221102.11
Escalante, H. J., Escalera, S., Guyon, I., Baró, X., Güçlütürk, Y., and Güçlü, U. (2018). “Explainable and interpretable models in computer vision and machine learning,” in Springer Verlag (The Springer Series on Challenges in Machine Learning). Available at: https://inria.hal.science/hal-01991623.
Esteva, A, Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., et al. (2017). Erratum: corrigendum: Dermatologist-level classification of skin cancer with deep neural networks. Nature 546 (7660), 686. doi:10.1038/nature22985
Frisch, S., Libuschewski, H., Peters, S., Gerß, J., von Hoff, K., Kortmann, R. D., et al. (2024). Radiation therapy plays an important role in the treatment of atypical teratoid/rhabdoid tumors: analysis of the EU-RHAB cohorts and their precursors. Int. J. Radiat. Oncol. Biol. Phys. 119 (4), 1147–1157. doi:10.1016/j.ijrobp.2024.01.200
Fu, Y., Liu, R., Liu, Y., and Lu, J. (2019). Intertrochanteric fracture visualization and analysis using a map projection technique. Med. Biol. Eng. Comput. 57 (3), 633–642. doi:10.1007/s11517-018-1905-1
Fuhrman, J. D., Gorre, N., Hu, Q., Li, H., El Naqa, I., and Giger, M. L. (2022). A review of explainable and interpretable AI with applications in COVID-19 imaging. Med. Phys. 49 (1), 1–14. doi:10.1002/mp.15359
Garreau, D., and von Luxburg, U. (2020). “Explaining the explainer: a first theoretical analysis of LIME,” in Proceedings of the twenty third international Conference on artificial Intelligence and statistics. PMLR (proceedings of machine learning research). Editors S. Chiappa, and R. Calandra, 1287–1296. Available at: https://proceedings.mlr.press/v108/garreau20a.html.
Gilpin, L. H., et al. (2018). “Explaining explanations: an overview of interpretability of machine learning,” in 2018 IEEE 5th International Conference on data science and advanced analytics (DSAA) (IEEE), 80–89.
Gottlieb, M., Kline, J. A., Schneider, A. J., and Coates, W. C. (2023). ChatGPT and conversational artificial intelligence: friend, foe, or future of research? Am. J. Emerg. Med. 70, 81–83. doi:10.1016/j.ajem.2023.05.018
Gumbs, A. A., Frigerio, I., Spolverato, G., Croner, R., Illanes, A., Chouillard, E., et al. (2021). Artificial intelligence surgery: how do we get to autonomous actions in surgery? Sensors (Basel, Switzerland) 21 (16), 5526. doi:10.3390/s21165526
Guo, W., et al. (2018). Explaining deep learning models--A bayesian non-parametric approach. Adv. neural Inf. Process. Syst. 31.
Hall, W. A., Paulson, E., Li, X. A., Erickson, B., Schultz, C., Tree, A., et al. (2022). Magnetic resonance linear accelerator technology and adaptive radiation therapy: an overview for clinicians. CA a cancer J. Clin. 72 (1), 34–56. doi:10.3322/caac.21707
Hamedan, F., Orooji, A., Sanadgol, H., and Sheikhtaheri, A. (2020). Clinical decision support system to predict chronic kidney disease: a fuzzy expert system approach. Int. J. Med. Inf. 138 (104134), 104134. doi:10.1016/j.ijmedinf.2020.104134
Hamm, C. A., Baumgärtner, G. L., Biessmann, F., Beetz, N. L., Hartenstein, A., Savic, L. J., et al. (2023). Interactive explainable deep learning model informs prostate cancer diagnosis at MRI. Radiology 307, 222276. doi:10.1148/radiol.222276
Han, T., et al. (2023) MedAlpaca – an open-source collection of medical conversational AI models and training data. doi:10.48550/arXiv.2304.08247
Haptik, T. (2023). “haptik”, conversational AI in healthcare. Spring. Available at: https://www.haptik.ai/blog/conversational-ai-healthcare-use-cases/.
Hassan, S. A., Akbar, S., and Khan, H. U. (2024). Detection of central serous retinopathy using deep learning through retinal images. Multimedia Tools Appl. 83 (7), 21369–21396. doi:10.1007/s11042-023-16206-y
Helman, S., Terry, M. A., Pellathy, T., Williams, A., Dubrawski, A., Clermont, G., et al. (2022). Engaging clinicians early during the development of a graphical user display of an intelligent alerting system at the bedside. Int. J. Med. Inform. 159, 104643. doi:10.1016/j.ijmedinf.2021.104643
Hooshmand, S. A., Zarei Ghobadi, M., Hooshmand, S. E., Azimzadeh Jamalkandi, S., Alavi, S. M., and Masoudi-Nejad, A. (2021). A multimodal deep learning-based drug repurposing approach for treatment of COVID-19. Mol. Divers 25 (3), 1717–1730. doi:10.1007/s11030-020-10144-9
Hossain, M. M., Walid, M. A. A., Galib, S. S., Azad, M. M., Rahman, W., Shafi, A., et al. (2024). COVID-19 detection from chest CT images using optimized deep features and ensemble classification. Syst. Soft Comput. 6, 200077. doi:10.1016/j.sasc.2024.200077
Howard, F. M., Kochanny, S., Koshy, M., Spiotto, M., and Pearson, A. T. (2020). Machine learning-guided adjuvant treatment of head and neck cancer. JAMA Netw. open 3 (11), e2025881. doi:10.1001/jamanetworkopen.2020.25881
Hu, Y., et al. (2019). “Multi-modal probabilistic prediction of interactive behavior via an interpretable model,” in 2019 IEEE intelligent vehicles symposium (IV) (IEEE), 557–563.
Jing, J., Liu, Z., Guan, H., Zhu, W., Zhang, Z., Meng, X., et al. (2023). A deep learning system to predict recurrence and disability outcomes in patients with transient ischemic attack or ischemic stroke. Adv. Intell. Syst. 5 (4), 2200240. doi:10.1002/aisy.202200240
Johansson, U., Sönströd, C., Norinder, U., and Boström, H. (2011). Trade-off between accuracy and interpretability for predictive in silico modeling. Future Med. Chem. 3 (6), 647–663. doi:10.4155/fmc.11.23
Kaushik, P. (2023). Artificial intelligence accelerated transformation in the healthcare industry. Amity J. Prof. Pract. 3. doi:10.55054/ajpp.v3i01.630
Ke, Y.-Y., Peng, T.-T., Yeh, T. K., Huang, W. Z., Chang, S. E., Wu, S. H., et al. (2020). Artificial intelligence approach fighting COVID-19 with repurposing drugs. Biomed. J. 43 (4), 355–362. doi:10.1016/j.bj.2020.05.001
Khalifa, M., Albadawy, M., and Iqbal, U. (2024). Advancing clinical decision support: the role of artificial intelligence across six domains. Comput. Methods Programs Biomed. Update 5, 100142. doi:10.1016/j.cmpbup.2024.100142
Kiani, A., Uyumazturk, B., Rajpurkar, P., Wang, A., Gao, R., Jones, E., et al. (2020). Impact of a deep learning assistant on the histopathologic classification of liver cancer. NPJ Digit. Med. 3, 23. doi:10.1038/s41746-020-0232-8
Kim, B., et al. (2017) Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). doi:10.48550/ARXIV.1711.11279
Kim, B., et al. (2019). “Why are saliency maps noisy? Cause of and solution to noisy saliency maps,” in 2019 IEEE/CVF international conference on computer vision workshop (ICCVW), Seoul, Korea (South), October 27–28, 2019 (IEEE), 4149–4157.
Kim, H., Yoo, S. K., Kim, J. S., Kim, Y. T., Lee, J. W., Kim, C., et al. (2024). Clinical feasibility of deep learning-based synthetic CT images from T2-weighted MR images for cervical cancer patients compared to MRCAT. Sci. Rep. 14 (1), 8504. doi:10.1038/s41598-024-59014-6
Kim, K. (2023). “COGNEX”, Overview of interpretable machine learning: (2) main interpretation methods of deep learning models in image recognition problems, 21 February. Available at: https://www.cognex.com/ko-kr/blogs/deep-learning/research/overview-interpretable-machine-learning-2-interpreting-deep-learning-models-image-recognition (Accessed February 21, 2023).
Lee, A., Baker, T. S., Bederson, J. B., and Rapoport, B. I. (2024). Levels of autonomy in FDA-cleared surgical robots: a systematic review. npj Digit. Med. 7 (1), 103. doi:10.1038/s41746-024-01102-y
Lee, G., Jeong, J., Seo, S., Kim, C., and Kang, P. (2018). Sentiment classification with word localization based on weakly supervised learning with a convolutional neural network. Knowledge-Based Syst. 152 (15), 70–82. doi:10.1016/j.knosys.2018.04.006
Lee, H., Kim, S. T., and Ro, Y. M. (2019). Generation of multimodal justification using visual word constraint model for explainable computer-aided diagnosis, abs/1906.03922. CoRR, 21–29. doi:10.1007/978-3-030-33850-3_3
LG AI Researcher (2021). New algorithm SISE of “explainable artificial intelligence” (XAI), Naver. Available at: https://m.post.naver.com/viewer/postView.naver?volumeNo=30627518&memberNo=52249799. (Accessed May 20, 2024)
Li, A., Xiao, F., Zhang, C., and Fan, C. (2021). Attention-based interpretable neural network for building cooling load prediction. Appl. Energy 299, 117238. doi:10.1016/j.apenergy.2021.117238
Linardatos, P., Papastefanopoulos, V., and Kotsiantis, S. (2020). Explainable ai: a review of machine learning interpretability methods. Entropy 23 (1), 18. doi:10.3390/e23010018
Lindsey, R., Daluiski, A., Chopra, S., Lachapelle, A., Mozer, M., Sicular, S., et al. (2018). Deep neural network improves fracture detection by clinicians. Proc. Natl. Acad. Sci. U. S. A. 115 (45), 11591–11596. doi:10.1073/pnas.1806905115
Liu, B., Guo, J., Chen, C. L. P., Wu, X., and Zhang, T. (2023). Fine-grained interpretability for EEG emotion recognition: concat-aided grad-CAM and systematic brain functional network. IEEE Trans. Affect. Comput. 15, 671–684. doi:10.1109/taffc.2023.3288885
Lo, C.-M., Hung, P.-H., and Lin, D.-T. (2021). Rapid assessment of acute ischemic stroke by computed tomography using deep convolutional neural networks. J. Digit. Imaging 34 (3), 637–646. doi:10.1007/s10278-021-00457-y
Lombardi, A., Monaco, A., Donvito, G., Amoroso, N., Bellotti, R., and Tangaro, S. (2021). Brain age prediction with morphological features using deep neural networks: Results from predictive analytic competition 2019. Front. Psychiatry. 11, 619629.
Longo, L., Brcic, M., Cabitza, F., Choi, J., Confalonieri, R., Ser, J. D., et al. (2024). Explainable Artificial Intelligence (XAI) 2.0: a manifesto of open challenges and interdisciplinary research directions. Inf. Fusion 106, 102301. doi:10.1016/j.inffus.2024.102301
Lotter, W., Haslam, B., Diab, A. R., Kim, J. G., Grisot, G., Wu, E., et al. (2021). Robust breast cancer detection in mammography and digital breast tomosynthesis using an annotation-efficient deep learning approach. Nat. Med. 27 (2), 244–249. doi:10.1038/s41591-020-01174-9
Masood, A., Sheng, B., Li, P., Hou, X., Wei, X., Qin, J., et al. (2018). Computer-assisted decision support system in pulmonary cancer detection and stage classification on CT images. J. Biomed. Inf. 79, 117–128. doi:10.1016/j.jbi.2018.01.005
McGenity, C., Clarke, E. L., Jennings, C., Matthews, G., Cartlidge, C., Freduah-Agyemang, H., et al. (2024). Artificial intelligence in digital pathology: a systematic review and meta-analysis of diagnostic test accuracy. npj Digit. Med. 7 (1), 114. doi:10.1038/s41746-024-01106-8
Mella, M. H., Chabrillac, E., Dupret-Bories, A., Mirallie, M., and Vergez, S. (2023). Transoral robotic surgery for head and neck cancer: advances and residual knowledge gaps. J. Clin. Med. 12 (6), 2303. doi:10.3390/jcm12062303
Mitchell, R., Cooper, J., Frank, E., and Holmes, G. (2022). Sampling permutations for shapley value estimation. J. Mach. Learn. Res. 23 (43), 1–46.
Mohamed, E. A., Rashed, E. A., Gaber, T., and Karam, O. (2022). Deep learning model for fully automated breast cancer detection system from thermograms. PLOS ONE 17 (1), 1–20. doi:10.1371/journal.pone.0262349
Moradi, M., and Samwald, M. (2021). Post-hoc explanation of black-box classifiers using confident itemsets. Expert Syst. Appl. 165, 113941. doi:10.1016/j.eswa.2020.113941
Nicolae, A., Semple, M., Lu, L., Smith, M., Chung, H., Loblaw, A., et al. (2020). Conventional vs machine learning-based treatment planning in prostate brachytherapy: results of a Phase I randomized controlled trial. Brachytherapy 19 (4), 470–476. doi:10.1016/j.brachy.2020.03.004
Nielsen, I. E., Dera, D., Rasool, G., Ramachandran, R. P., and Bouaynaya, N. C. (2022). Robust explainability: a tutorial on gradient-based attribution methods for deep neural networks. IEEE Signal Process. Mag. 39 (4), 73–84. doi:10.1109/msp.2022.3142719
Noh, HS, Park, J., Sim, G., Yu, J. E., and Chung, Y. (2014). Nonparametric bayesian statistical models in biomedical research. Korean J. Appl. Statistics 27 (6), 867–889. doi:10.5351/kjas.2014.27.6.867
Origimid (2022). Artificial intelligence assisted diagnosis. Available at: https://www.origimed.com/Product/ai (Accessed May 20, 2024).
Pang, S., Ding, T., Qiao, S., Meng, F., Wang, S., Li, P., et al. (2019). A novel YOLOv3-arch model for identifying cholelithiasis and classifying gallstones on CT images. PLoS one 14 (6), e0217647. doi:10.1371/journal.pone.0217647
Pavlidis, G. (2024). Unlocking the black box: analysing the EU artificial intelligence act’s framework for explainability in AI. Law, Innovation Technol. 16, 293–308. doi:10.1080/17579961.2024.2313795
Plumb, G., Molitor, D., and Talwalkar, A. S. (2018). “Model agnostic supervised local explanations,” in Advances in neural information processing systems. Editor S. Bengioet al. (Curran Associates, Inc.). Available at: https://proceedings.neurips.cc/paper/2018/file/b495ce63ede0f4efc9eec62cb947c162-Paper.pdf (Accessed January 6, 2024).
Quiroz, J. C., Feng, Y. Z., Cheng, Z. Y., Rezazadegan, D., Chen, P. K., Lin, Q. T., et al. (2021). Development and validation of a machine learning approach for automated severity assessment of COVID-19 based on clinical and imaging data: retrospective study. JMIR Med. Inf. 9 (2), e24572. doi:10.2196/24572
Rawson, T. M., Hernandez, B., Moore, L. S. P., Herrero, P., Charani, E., Ming, D., et al. (2021). A real-world evaluation of a case-based reasoning algorithm to support antimicrobial prescribing decisions in acute care. Clin. Infect. Dis. 72 (12), 2103–2111. doi:10.1093/cid/ciaa383
Reddy, K., Gharde, P., Tayade, H., Patil, M., Reddy, L. S., and Surya, D. (2023). Advancements in robotic surgery: a comprehensive overview of current utilizations and upcoming frontiers. Cureus. 15 (12), e50415. doi:10.7759/cureus.50415
Ribeiro, M. T., Singh, S., and Guestrin, C. (2018). Anchors: high-precision model-agnostic explanations. Proc. AAAI Conf. Artif. Intell. 32 (1). doi:10.1609/aaai.v32i1.11491
Sadeghi, Z., Alizadehsani, R., Cifci, M. A., Kausar, S., Rehman, R., Mahanta, P., et al. (2024). A review of explainable artificial intelligence in healthcare. Comput. Electr. Eng. 118, 109370. doi:10.1016/j.compeleceng.2024.109370
Saeed, W., and Omlin, C. (2023). Explainable AI (XAI): a systematic meta-survey of current challenges and future opportunities. Knowledge-Based Syst. 263, 110273. doi:10.1016/j.knosys.2023.110273
Salem, N. M., Jack, K. M., Gu, H., Kumar, A., Garcia, M., Yang, P., et al. (2023). Machine and deep learning identified metabolites and clinical features associated with gallstone disease. Comput. Methods Programs Biomed. Update 3, 100106. doi:10.1016/j.cmpbup.2023.100106
Salih, S. (2022). Understanding machine learning interpretability. Available at: https://towardsdatascience.com/understanding-machine-learning-interpretability-168fd7562a1a (Accessed February 6, 2024).
Savindu Herath Pathirannehelage, Y. R. S., Krogh, G. von, and von Krogh, G. (2024). Design principles for artificial intelligence-augmented decision making: an action design research study. Eur. J. Inf. Syst. 0 (0), 1–23. doi:10.1080/0960085X.2024.2330402
Schöttl, A. (2020). “A light-weight method to foster the (Grad) CAM interpretability and explainability of classification networks,” in 2020 10th international conference on advanced computer information technologies (ACIT), Deggendorf, Germany, September 16–18, 2020 (IEEE), 348–351.
Selvaraju, R. R., et al. (2017). “Grad-cam: visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE international conference on computer vision, Venice, Italy, October 22–29, 2017 (IEEE), 618–626.
Serednicki, W. A., Hołówko, W., Major, P., Małczak, P., and Pędziwiatr, M. (2023). Minimizing blood loss and transfusion rate in laparoscopic liver surgery: a review. Wideochir. Inne. Tech. Maloinwazyjne 18 (2), 213–223. doi:10.5114/wiitm.2022.124088
Shapley, L. S. (2016) “17. A value for n-person games,” in Contributions to the theory of games (AM-28), eds H. W. Kuhn, and A. W. Tucker Vol. II. (Princeton University Press), pp. 307–318. doi:10.1515/9781400881970-018
Shi, S., Du, Y., and Fan, W. (2020) An extension of LIME with improvement of interpretability and fidelity.
Shorfuzzaman, M., Masud, M., Alhumyani, H., Anand, D., and Singh, A. (2021). Artificial neural network-based deep learning model for COVID-19 patient detection using X-ray chest images. J. Healthc. Eng. 2021, 1–16. doi:10.1155/2021/5513679
Singh, A., Sengupta, S., and Lakshminarayanan, V. (2020). Explainable deep learning models in medical image analysis. arXiv 6, 52. doi:10.3390/jimaging6060052
Sinha, S., Vegesna, R., Mukherjee, S., Kammula, A. V., Dhruba, S. R., Wu, W., et al. (2024). PERCEPTION predicts patient response and resistance to treatment using single-cell transcriptomics of their tumors. Nat. Cancer 5, 938–952. doi:10.1038/s43018-024-00756-7
Siouras, A., Moustakidis, S., Giannakidis, A., Chalatsis, G., Liampas, I., Vlychou, M., et al. (2022). Knee injury detection using deep learning on MRI studies: a systematic review. Diagn. (Basel, Switzerland) 12 (2), 537. doi:10.3390/diagnostics12020537
Stano, M., Benesova, W., and Martak, L. S. (2020). “Explainable 3D convolutional neural network using GMM encoding,” in Twelfth international conference on machine vision (ICMV 2019). Editors W. Osten, and D. P. Nikolaev (SPIE), 507–513. doi:10.1117/12.2557314
Stefanelli, E., Cordella, F., Gentile, C., and Zollo, L. (2023). Hand prosthesis sensorimotor control inspired by the human somatosensory system. Robotics 12 (5), 136. doi:10.3390/robotics12050136
Stinson, C., and Vlaad, S. (2024). A feeling for the algorithm: diversity, expertise, and artificial intelligence. Big Data & Soc. 11. doi:10.1177/20539517231224247
Strumbelj, E., and Kononenko, I. (2011) ‘A general method for visualizing and explaining black-box regression models, , 21–30. doi:10.1007/978-3-642-20267-4_3
Subramani, S., Varshney, N., Anand, M. V., Soudagar, M. E. M., Al-Keridis, L. A., Upadhyay, T. K., et al. (2023). Cardiovascular diseases prediction by machine learning incorporation with deep learning. Front. Med. 10, 1150933. doi:10.3389/fmed.2023.1150933
T, J. (2020). [Interpretable machine learning (5-8)] scope rule (anchor). Tistory. Available at: https://eair.tistory.com/27 (Accessed May 12, 2024).
Vinogradova, K., Dibrov, A., and Myers, G. (2020). “Towards interpretable semantic segmentation via gradient-weighted class activation mapping (student abstract),” in Proceedings of the AAAI conference on artificial intelligence, New York, United States, February 2020, 13943–13944.
Wang, P., Liu, X., Berzin, T. M., Glissen Brown, J. R., Liu, P., Zhou, C., et al. (2020). Effect of a deep-learning computer-aided detection system on adenoma detection during colonoscopy (CADe-DB trial): a double-blind randomised study. Lancet Gastroenterol. Hepatol 5 (4), 343–351. doi:10.1016/S2468-1253(19)30411-X
Wu, X., Tao, Y., Qiu, Q., and Wu, X. (2018). Application of image recognition-based automatic hyphae detection in fungal keratitis. Australas. Phys. Eng. Sci. Med. 41 (1), 95–103. doi:10.1007/s13246-017-0613-8
Wu Yazhou, Y. D., and Chen, X. (2022). Research progress and prospect of artificial intelligence in the clinical field. J. Army Med. Univ. 44, 89–102. doi:10.16016/j.1000-5404.202112092
Xie, X., Li, K., Wu, Y., Ai, C., Wang, Q., Wang, C., et al. (2021). Preliminary experiences with robot-assisted choledochal cyst excision using the da vinci surgical system in children below the age of one. Front. Pediatr. 23 (9), 741098. doi:10.3389/fped.2021.741098
Yang, W., Wei, Y., Wei, H., Chen, Y., Huang, G., Li, X., et al. (2023). Survey on explainable AI: from approaches, limitations and applications aspects. Human-Centric Intell. Syst. 3 (3), 161–188. doi:10.1007/s44230-023-00038-y
Yoo, Y. J., Ha, E. J., Cho, Y. J., Kim, H. L., Han, M., and Kang, S. Y. (2018). Computer-aided diagnosis of thyroid nodules via ultrasonography: initial clinical experience. Korean J. Radiol. 19 (4), 665–672. doi:10.3348/kjr.2018.19.4.665
Zadeh Shirazi, A., McDonnell, M. D., Fornaciari, E., Bagherian, N. S., Scheer, K. G., Samuel, M. S., et al. (2021). A deep convolutional neural network for segmentation of whole-slide pathology images identifies novel tumour cell-perivascular niche interactions that are associated with poor survival in glioblastoma. Br. J. Cancer 125 (3), 337–350. doi:10.1038/s41416-021-01394-x
Zhang, J., and Zhang, Z. (2023). Ethics and governance of trustworthy medical artificial intelligence. BMC Med. Inf. Decis. Mak. 23 (1), 7. doi:10.1186/s12911-023-02103-9
Zheng, S., Rammohan, N., Sita, T., Teo, P. T., Wu, Y., Lesniak, M., et al. (2024). GlioPredictor: a deep learning model for identification of high-risk adult IDH-mutant glioma towards adjuvant treatment planning. Sci. Rep. 14 (1), 2126. doi:10.1038/s41598-024-51765-6
Zhongqi, H., and Jia, Z. C. (2022). Development and application of software testing under artificial intelligence, 21ks. Available at: https://www.21ks.net/lunwen/rgznlw/180797.html (Accessed May 23, 2024).
Keywords: artificial intelligence, machine learning, deep learning, healthcare, interpretability, explainability, accuracy
Citation: Ennab M and Mcheick H (2024) Enhancing interpretability and accuracy of AI models in healthcare: a comprehensive review on challenges and future directions. Front. Robot. AI 11:1444763. doi: 10.3389/frobt.2024.1444763
Received: 06 June 2024; Accepted: 27 September 2024;
Published: 28 November 2024.
Edited by:
Sanja Dogramadzi, The University of Sheffield, United KingdomReviewed by:
Christian Tamantini, National Research Council (CNR), ItalySadiq Hussain, Dibrugarh University, India
Copyright © 2024 Ennab and Mcheick. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Mohammad Ennab, bWVubmFiQGV0dS51cWFjLmNh
†ORCID: Mohammad Ennab, orcid.org/0000-0001-7117-6271; Hamid Mcheick, orcid.org/0000-0002-2589-8609