ORIGINAL RESEARCH article

Front. Neurol., 12 January 2026

Sec. Artificial Intelligence in Neurology

Volume 16 - 2025 | https://doi.org/10.3389/fneur.2025.1579206

Assessing the effectiveness of machine learning and deep learning in differentiating neuroimmunological diseases: a systematic review and meta-analysis

  • 1. Faculty of Medicine, Vilnius University, Vilnius, Lithuania

  • 2. Clinic of Neurology and Neurosurgery, Institute of Clinical Medicine, Vilnius University, Vilnius, Lithuania

Article metrics

View details

1k

Views

78

Downloads

Abstract

Objective:

The differential diagnosis of neuroimmunological disorders remains a significant challenge in clinical practice, even with advancements in diagnostic techniques. Recently, the use of artificial intelligence (AI) for diagnosing and distinguishing between various neuroimmunological disorders has gained traction. Our objective was to conduct a systematic review and meta-analysis to evaluate the diagnostic performance of Machine Learning (ML) and Deep Learning (DL) techniques in differentiating these disorders. We aimed to identify the most effective approaches, compare their diagnostic outcomes, and offer recommendations for improving their applicability across multiple clinical centers and for future research.

Methods:

Following the PRISMA 2020 guidelines, a systematic search in PubMed and Web of Science was conducted to identify relevant articles published between 2000 and 2024 that fell within the scope of our research. QUADAS-2 tool was assessed to evaluate the risk of bias and applicability concerns. The performed meta-analysis allowed us to estimate the overall accuracy, sensitivity, and specificity of the developed models providing quantitative insights from this analysis.

Results:

Of 4,470 articles identified, 19 met inclusion criteria: 9 (47.4%) used ML and 10 (52.6%) used DL. Most models relied on MRI data to differentiate multiple sclerosis from neuromyelitis optica spectrum disorders. Pooled accuracy, sensitivity, and specificity were 0.87, 0.86, and 0.84, respectively. Substantial heterogeneity was observed, which decreased in a sensitivity analysis excluding larger-sample studies and varied between ML and DL models, with ML showing lower heterogeneity.

Conclusion:

New AI tools, primarily utilizing MRI data, are emerging and demonstrate the potential to differentiate between various neuroimmunological disorders. While most neuroimmunological conditions have accessible antibody tests with strong diagnostic performance, AI efforts should concentrate on seronegative diseases. This approach should incorporate clinical and epidemiological data into diagnostic algorithms for improved accuracy.

1 Introduction

Differential diagnosis of neuroimmunological disorders remains challenging in clinical practice despite evolving diagnostic techniques (13). Application of artificial intelligence (AI) to diagnose (4) and differentiate between multiple sclerosis (MS), neuromyelitis optica spectrum disorders (NMOSD), myelin oligodendrocyte glycoprotein antibody-associated disease (MOGAD), and autoimmune encephalitis (AE) have been increasingly exploited. On the one hand, AI techniques may benefit standard clinical care by processing large amounts of information, including clinical data and radiological images, to identify patterns undetectable by conventional means (5) aiding decision-making and reducing the risk of human error (6).

On the other hand, despite the benefits mentioned, only a small proportion of AI tools are applied internationally (7). Similarly, the use of AI tools in neuroimmunological disorders is limited, with only a few studies published to date (8). To assess advancements in AI techniques within the field of neuroimmunology, we conducted a systematic review and meta-analysis to evaluate the application of ML and DL techniques in differentiating neuroimmunological diseases. Our goals include identifying commonly used approaches, analyzing their diagnostic performance, and providing recommendations to enhance their applicability across various clinical centers and future research.

A previous systematic review and meta-analysis focused primarily on MS versus NMOSD and reported substantial heterogeneity (9). Our study extends these findings by examining contributors to heterogeneity, such as study size, dataset composition, and model architecture, and emphasizes the need for improved methodological rigor.

2 Material and methods

2.1 Information sources and search

A systematic search was conducted to select articles that fall within the scope of our research. We reviewed publications in PubMed and Web of Science databases published from 2000 to 2024. The temporal search range was preregistered to ensure comprehensive coverage of earlier literature. The search strategy that was applied: (((((((((Multiple sclerosis) OR (Autoimmune encephalitis)) OR (Neuromyelitis optica)) OR (NMOSD)) OR (NMO)) OR (Devic's disease)) OR (Myelin oligodendrocyte glycoprotein)) OR (MOG)) OR (MOGAD)) AND ((((Artificial intelligence) OR (Machine Learning)) OR (Deep Learning)) OR (Neural network)). Search results were included or excluded in the final analysis based on the criteria shown below, accordingly.

2.2 Eligibility criteria

Studies were considered eligible for inclusion if they met all of the following criteria: (1) investigated the differentiation between neuroimmunological diseases (e.g., MS, NMOSD, AE, MOGAD), (2) utilized Machine Learning or Deep Learning techniques for classification or diagnostic purposes, (3) involved human subjects, (4) were available as full-text articles, (5) were published from the year 2000 onward, and (6) were written in English.

2.3 Screening

Two reviewers (D.P. and M.V.) independently performed the screening process. After excluding articles based on the titles and abstracts that are out of our scope, the rest were sought for retrieval as potentially eligible and were assessed full text.

2.4 Data collection and items

Two reviewers (D.P. and M.V.) independently performed data extraction to ensure accuracy and reduce bias. Any discrepancies between reviewers were resolved through discussion, and, if necessary, a third reviewer (N.G.) was consulted to reach consensus. Data was extracted into the spreadsheet we created to include the relevant data. Further metadata was extracted: 1. first author; 2. year of article; 3. neuroimmunological diseases; 4. objective of the study; 5. used parameters, e.g. clinical data, MRI images; 6. data source; 7. AI technique, e.g. Machine Learning or Deep Learning; 8. model performance, e.g. accuracy, specificity, sensitivity, area under the curve (AUC).

2.5 Data synthesis

Extracted relevant data was categorized and tabulated to facilitate a comprehensive analysis. Categorization was primarily based on AI techniques, distinguishing whether the classification task was performed with Machine Learning or Deep Learning models. By conducting a meta-analysis we chose a random-effects model due to variations in parameter characteristics, patient populations (data sources), and AI algorithms. I2 and τ statistics were used to assess the degree of data heterogeneity (10). The meta-analysis was conducted using the meta package in R 4.2.2 (11).

Among the included studies, MRI-based MS versus NMOSD comparisons represented the largest and most methodologically comparable group, and were therefore included in the quantitative meta-analysis. For studies reporting multiple ML or DL models, the model with the highest reported performance metrics was selected for inclusion in the quantitative meta-analysis. Studies using imaging modalities other than MRI or investigating neuroimmunological diseases other than MS versus NMOSD were planned to be synthesized narratively if they were few in number or methodologically heterogeneous, rather than included in the quantitative meta-analysis.

For studies that did not provide full 2 × 2 tables, counts were reconstructed from the reported sensitivity, specificity, accuracy, and corresponding sample sizes, and these reconstructed counts were used for the meta-analysis. Meta-analyses of sensitivity, specificity, and accuracy were performed using the metaprop() function. A univariate random-effects model with inverse-variance weighting was applied, and logit transformation (sm = “PLOGIT”) was used to stabilize variances. In the meta-analysis, all counts—both reported and reconstructed—were sufficient for analysis without requiring a continuity correction. Univariate models were chosen rather than bivariate or HSROC models because the number of included studies was limited and several studies did not report complete 2 × 2 tables, making estimation of between-study correlation unreliable.

Publication bias was assessed using Deeks' funnel plot asymmetry test. A p-value > 0.05 was considered to indicate no significant small-study effects.

2.6 Quality assessment and risk of bias

Risk of bias and applicability were assessed using the QUADAS-2 tool (12), which evaluates four domains: patient selection, index test, reference standard, and flow and timing. Two reviewers (D.P. and M.V.) independently conducted the assessments. Discrepancies were resolved through discussion and consensus, and overall inter-reviewer agreement was high. Each domain was rated as low, high, or unclear risk based on standard criteria. High risk was assigned for patient selection with non-transparent or non-representative sampling; index tests lacking sufficient methodological detail; reference standards without clearly defined diagnostic criteria; and flow or timing concerns regarding the application of the diagnostic process.

3 Results

3.1 Study selection and characteristics

Our search strategy identified a total of 4,470 publications. After screening, results from 19 articles met the inclusion criteria and were included in the systematic review. Figure 1 presents a diagram depicting the flow of study selection.

Figure 1

Flowchart detailing study identification and screening process. It begins with 4,470 records from Pubmed and Web of Science. After removing duplicates and ineligible entries, 3,249 records are screened, excluding 3,190 as off-topic or retracted. Fifty-nine reports are sought, twelve not retrieved due to lack of full text. Forty-seven assessed for eligibility, excluding specific studies for not meeting criteria. Nineteen studies included in the review, eleven in the meta-analysis.

PRISMA flow diagram depicting the flow of study selection.

All the studies included in this systematic review and meta-analysis were published between 2020 and 2024. Nine publications (47.4%) implemented ML algorithms, while ten (52.6%) utilized DL techniques.

As shown in Table 1, among the articles applying ML, the most common application (n = 5) was differentiating between MS and NMOSD. Other studies developed AI techniques mainly between antibody-associated demyelinating disease (NMOSD and MOGAD, respectively).

Table 1

Study Diseases Data source Model Imaging modality Parameters Performance Training / validation set Test set
El Khoury et al. (24) MS vs. NMOSD 60 MS, 60 NMOSD Random forest Fourier-transform infrared spectra of serum samples AUC: 100%, Sensitivity: 100%, Specificity: 100%, Precision: 100% 108 12
Yan et al. (25) MS vs. NMOSD 47 MS, 36 NMOSD Logistic regression MRI Brain radiomics signatures and demographic information Combined model: AUC of 0.927 (95% CI: 0.871–0.984), Demographic information-only model: AUC of 0.733 (95% CI: 0.639–0.818), Sensitivity 0.511, Specificity 0.861, Accuracy 0.663, Radiomics-only model: AUC of 0.902 (95% CI: 0.840–0.955), Sensitivity 0.851, Specificity 0.889, Accuracy 0.867 83
Clarke et al. (26) MS vs. NMOSD 100 MS, 66 NMOSD Decision tree MRI Brain, spine, orbits T1, T2, FLAIR sequences TP: 60, FP: 4, TN: 96, FN: 6, TP rate: 0.929, FP rate: 0.060, Precision 0.939, F-measure: 0.934, AUC: 0.935
Huang et al. (27) MS vs. NMOSD 78 MS, 38 NMOSD Random forest MRI Brain radiomic features (extracted from T1-MPRAGE and T2 sequences), clinical features Multi-parametric MRI: AUC 0.902 ± 0.027, Sensitivity 0.873 ± 0.083, Specificity 0.869 ± 0.051, Accuracy 0.871 ± 0.044 86 30
Gharaibeh et al. (28) MS vs. NMOSD 424 MS, 261 NMOSD KNN (VGG16, VGG19, InceptionV3 for feature extraction) MRI Brain features extracted from FLAIR and T2W sequences VGG16:
KNN: Precision 0.98, Recall 0.99, F1-score 0.99, Accuracy 0.99
VGG19:
KNN: Precision 0.96, Recall 0.98, F1-score 0.97, Accuracy 0.97
InceptionV3:
KNN: Precision 0.92, Recall 0.95, F1-score 0.93, Accuracy 0.93
548 137
Ciftci Kavaklioglu et al. (29) MS vs. (NMOSD and MOGAD) 57 MS, 11 NMOSD, 27 MOGAD Random forest OCT OCT features Accuracy: 0.68, Sensitivity: 0.69, Specificity: 0.67, AUC: 0.73 76 19
Luo et al. (30) MS vs. (NMOSD and MOGAD); NMOSD vs. MOGAD 63 MS, 87 NMOSD, 45 MOGAD Random forest, logistic regression MRI Brain radiomics and spatial distribution features of brain lesions extracted from T1, T2-FLAIR sequences 1. MS vs. (NMOSD and MOGAD)
Joint model: AUC 0.927, Accuracy 0.863, Sensitivity 0.858, Specificity 0.868
2. MOGAD vs NMOSD
Joint model: AUC 0.871, Accuracy 0.805, Sensitivity 0.808, Specificity 0.805
(1) 195 (2) 132
Ding et al. (31) MOGAD vs. non-MOGAD 66 MOGAD, 66 non-MOGAD Support vector machine MRI Radiomic features extracted from T1WI, T2WI, T2W-FLAIR, DWI sequences Internal test set (AUC 0.844, Accuracy 83.33%, Sensitivity 85.71%, Specificity 81.25%)
External test set (AUC 0.846, Accuracy 80.65%, Sensitivity 93.75%, Specificity 66.67%)
101 31
Wei et al. (32) ADEM vs. MOGAD 49 ADEM, 21 MOGAD Multilayer perceptron, support vector machine MRI Brain radiomic features extracted from FLAIR sequence 0–6y Female:
MLP: Accuracy 0.784, F1 0.556, Specificity 0.774, Sensitivity 0.833, AUC 0.903
0–6y Male:
SVM: Accuracy: 0.805, F1 0.638, Specificity 0.821, Sensitivity 0.750, AUC 0.890
0–14y Female:
SVM: Accuracy: 0.891, F1 0.759, Specificity 0.885, Sensitivity 0.917, AUC 0.981
0–6y Male: SVM: Accuracy: 0.971, F1 0.857, Specificity 1.000, Sensitivity 0.750, AUC 0.992
70

Studies differentiating neuroimmunological diseases using machine learning.

MS, Multiple Sclerosis; NMOSD, Neuromyelitis Optica Spectrum Disorder; MOGAD, Myelin Oligodendrocyte Glycoprotein Antibody Disease; ADEM, Acute Disseminated Encephalomyelitis; F1, F1 Score; MRI, Magnetic resonance imaging; OCT, Optical Coherence Tomography.

Similarly, most studies using DL applied neural networks to distinguish MS vs NMOSD. Other studies heterogeneously used models to differentiate between antibody-associated nervous system disorders. Table 2 lists the studies employing DL techniques.

Table 2

Study Diseases Data source Model Imaging modality Parameters Performance Training /validation set Test set
Cacciaguerra et al. (33) MS vs. NMOSD 95 MS, 85 NMOSD ResNet MRI Brain T2- and T1-weighted sequences Accuracy: 0.95, MAE of 0.21, and MSE of 0.07 180
Seok et al. (34) MS vs. NMOSD 86 MS, 70 NMOSD ResNet18 MRI Brain FLAIR sequences Accuracy: 76.1%, Sensitivity: 77.3%, Specificity: 74.8%, PPV: 76.9%, NPV: 78.6%, AUC: 0.85 156
Kim et al. (35) MS vs. NMOSD 213 MS, 125 NMOSD ResNeXt MRI Brain 2D FLAIR sequences, clinical data Accuracy: 71.1%, Sensitivity: 87.8%, Specificity, 61.6%, AUC:0.82 203 135
Hagiwara et al. (36) MS vs. NMOSD 35 MS, 18 NMOSD SqueezeNet MRI Brain multi-dynamic multi-echo sequence AUC:0.859. MS Sensitivity: 80.0%, NMOSD Sensitivity: 83.3%. Accuracy: 81.1% 53
Zhuo et al. (37) MS vs. NMOSD 134 MS, 186 NMOSD MultiResUNet, DenseNet121 MRI Spine T2-weighted sequence Accuracy: 79.5%, Sensitivity: 80.0%, Specificity: 78.8%, PPV: 83.7%, NPV: 74.3%, Precision: 83.7%, Recall: 80.0%, AUC:0.85 242 78
Wang et al. (38) MS vs. NMOSD 41 NMOSD, 47 MS Pre-trained ResNet18 MRI Brain T2-FLAIR sequence Accuracy: 0.750, Sensitivity: 0.707, Specificity: 0.759 88
Huang et al. (39) MS vs. NMOSD 69 MS, 62 NMOSD ResNet MRI Brain T2-FLAIR sequence Accuracy: 92.16%, Sensitivity: 95.60%, Specificity: 92.60%, AUC: 96.33% 131
Huang et al. (40) MS vs. (MOGAD and NMOSD) NMOSD vs. (MS and MOGAD) MOGAD vs. (MS and NMOSD) 67 MS, 162 NMOSD, 61 MOGAD MIL-CoaT MRI Brain T2WI, brain T2-FLAIR, cervicothoracic T2WI, and thoracolumbar T2WI sequences MS vs. (MOGAD and NMOSD) Brain MRI (AUC:0.936, Accuracy: 88.9%, Sensitivity: 78.6%, Specificity: 92.5%, PPV: 78.6%, NPV: 92.5%, F1: 0.786)
NMOSD vs. (MS and MOGAD) Combined brain and spinal cord MRI (AUC: 0.942, Accuracy: 88.1%, Sensitivity: 87.9%, Specificity: 88.5%, PPV: 90.6%, NPV: 85.2%, F1: 0.892)
MOGAD vs. (MS and NMOSD) Combined brain and spinal cord MRI (AUC: 0.803, Accuracy: 72.9%, Sensitivity: 83.3%, Specificity: 70.2%, PPV: 41.7%, NPV: 94.3%, F1: 0.556)
231 59
Zhou et al. (41) NMOSD vs. ADEM 16 NMOSD, 174 ADEM M-DDC MRI Brain MRI images Precision: 96.96%, Recall: 96.96%, Accuracy: 99.19%,
AUC: 96.66, Fβ: 96.96%
152 38
Pan et al. (42) AE(LGI1) vs. AE(GABAB) 64 AE(LGI1), 17 AE(GABAB) ResNet18 PET/CT Brain PET/CT images AUC: 0.98, Accuracy: 96.30%, Sensitivity: 94.12%,
Specificity: 96.88%
81

Studies differentiating neuroimmunological diseases using deep learning.

MS, Multiple Sclerosis; NMOSD, Neuromyelitis Optica Spectrum Disorder; MOGAD, Myelin Oligodendrocyte Glycoprotein Antibody Disease; ADEM, Acute Disseminated Encephalomyelitis; and AE, Autoimmune Encephalitis associated with Leucine-Rich Glioma-Inactivated 1 (LGI1) or GABAB, Gamma-Aminobutyric Acid Receptor B; MAE, Mean Absolute Error; MSE, Mean Squared Error; PPV, Positive Predictive Value; NPV, Negative Predictive Value; F1, F1 Score; Fβ, F-beta Score; MRI, Magnetic Resonance Imaging; PET/CT, Positron Emission Tomography/Computed Tomography.

Number of MRI image samples.

Reporting of seronegative patients was limited: only a small subset of the included studies explicitly stated whether seronegative cases were part of their cohorts.

3.2 Risk of bias and applicability concerns

The quality assessment using QUADAS-2 revealed several methodological limitations across the included studies. Many models were developed using single-center datasets with relatively small sample sizes, increasing the potential for bias and limiting generalizability. Case selection procedures were often insufficiently described, making it unclear whether participants were enrolled consecutively or randomly, and whether study populations were representative of the broader clinical cohorts. Limited reporting of performance metrics and validation methods, such as cross-validation or external testing, further raised concerns regarding the robustness of the reported diagnostic performance. Additionally, a few studies relied on parameters not routinely available in standard clinical settings (e.g., PET/CT), which may restrict reproducibility and wider applicability. Supplementary Table 2 provides full QUADAS-2 ratings for all included studies, summarizing domain-level risk-of-bias and applicability assessments.

The highest risk of bias was observed in the patient selection domain (n = 9, 47.4% of all studies). A high risk of bias was assessed for models based on a limited number of disease subtypes, such as relapsing-remitting multiple sclerosis (RRMS) or seropositive NMOSD, not including others. The lack of transparency of data inclusion also increases the risk, raising concerns about the further applicability of such models. In contrast, index test, reference standard, and flow and timing domains had low risk in 73.7%, 89.5%, and 94.7% of studies, respectively. Despite the quite significant risk of bias in the patient selection domain, in terms of applicability, most studies were rated as less high risk, as illustrated in Figure 2.

Figure 2

Two horizontal bar charts display risks of bias and applicability concerns across QUADAS-2 domains. The left chart shows risks of bias with “Flow and Timing” mostly low risk, “Reference Standard” low to unclear risk, “Index Test” varied, and “Patient Selection” split between low and high risk. The right chart shows applicability concerns with most domains at low risk, except “Patient Selection,” which shows low, unclear, and high risk proportions. Color coding: green for low, yellow for unclear, red for high risk, with proportions labeled on bars.

The proportion of studies assessed as having high, low, or unclear risk of bias and applicability concerns.

3.3 Meta-analysis

To perform a meta-analysis, we estimated the pooled accuracy, sensitivity, and specificity to provide a comprehensive understanding of the diagnostic performance of ML and DL models in differentiating neuroimmunological diseases. We included 11 studies that investigated the differentiation between MS and NMOSD based on brain and/or spinal MRI data. Studies using other parameters, such as optical coherence tomography (OCT) or serum samples, were excluded, as including them would increase heterogeneity, especially given that only two such studies were available.

After removing outlier studies—specifically, studies that used modalities other than MRI or investigated neuroimmunological diseases outside MS vs. NMOSD, as their outcomes were not directly comparable—we performed a random-effects meta-analysis to estimate pooled diagnostic performance (see Supplementary Table 1 for the full list of included and excluded studies). Models classifying between MS and NMOSD achieved a pooled accuracy of 0.87, indicating strong overall performance. The pooled sensitivity and specificity were 0.86 and 0.84, respectively. Substantial heterogeneity was found across studies for accuracy (I2 = 84.2%) and specificity (I2 = 73.9%), while heterogeneity for sensitivity was moderate (I2 = 65.1%; Figure 3a).

Figure 3

Three panels of forest plots labeled a, b, and c, display study data on accuracy, sensitivity, and specificity. Each plot shows individual study estimates, proportions, confidence intervals, and weights, along with a summary random-effects model. Heterogeneity statistics with p-values are included at the bottom of each plot. Panel a, b, and c plots have varying studies with summary estimates: Panel a - 0.87 accuracy, 0.86 sensitivity, 0.84 specificity; Panel b - 0.83 accuracy, 0.79 sensitivity, 0.81 specificity; Panel c - 0.90 sensitivity, 0.93 specificity.

Forest plots of pooled diagnostic performance. (a) All included studies. (b) Studies with sample sizes < 100. (c) Studies stratified by model type (ML and DL).

To account for heterogeneity, we conducted a secondary analysis by excluding studies with sample sizes greater than 100, thereby including only smaller sample studies. As illustrated in Figure 3b, the pooled accuracy, sensitivity, and specificity in this subset were 0.83, 0.79, and 0.81, respectively. Notably, heterogeneity was markedly reduced in this analysis (accuracy: I2 = 53.7%; sensitivity: I2 = 0.0%; specificity: I2 = 0.0%).

We also performed subgroup analyses by model type (Figure 3c). In the ML group, the pooled sensitivity was 0.90 and the pooled specificity was 0.93. Heterogeneity was low to moderate (sensitivity: I2 = 24.2%; specificity: I2 = 32.0%). In the DL group, pooled sensitivity and specificity were 0.83 and 0.78, respectively. Heterogeneity was higher in this group (sensitivity: I2 = 64.7%; specificity: I2 = 64.7%).

3.4 Publication bias

The visual inspection of the funnel plot revealed a symmetrical distribution of the included studies around the regression line, suggesting the absence of small-study effects (Figure 4). In addition, Deeks' asymmetry test yielded a non-significant result (p = 0.4904), indicating no statistically significant evidence of publication bias. These findings suggest that the likelihood of publication bias influencing the pooled diagnostic accuracy estimates is low.

Figure 4

Scatter plot illustrating the relationship between inverse square root of effective sample size and natural log of diagnostic odds ratio (ln(DOR)). Points represent studies and a red line shows the regression line. Deeks' test p-value is 0.4904, indicating non-significant asymmetry.

Funnel plot assessing publication bias in the included diagnostic accuracy studies. Deeks' asymmetry test showed p > 0.05.

4 Discussion

In this review, we synthesized current evidence on AI applications for differentiating neuroimmunological disorders and performed a meta-analysis to evaluate the diagnostic performance of these models. Although individual studies frequently reported solid diagnostic accuracy, their results varied substantially, reflecting differences in study design, dataset characteristics, and modeling approaches.

Our meta-analysis demonstrated strong overall performance of AI-based models in distinguishing MS from NMOSD, with pooled accuracy, sensitivity, and specificity of 0.87, 0.86, and 0.84, respectively. Heterogeneity was substantial for accuracy and specificity and moderate for sensitivity; however, it decreased markedly after excluding large-sample studies, indicating that dataset size contributed significantly to variability. Subgroup analyses showed that ML models achieved higher pooled sensitivity (0.90) and specificity (0.93)—with lower heterogeneity—compared with DL models (0.83 and 0.78). While ML models showed higher pooled sensitivity and specificity than DL models, these comparisons are exploratory and should not be interpreted as definitive evidence of superiority. These ML–DL comparisons should be interpreted cautiously, as DL generally requires larger and more diverse datasets, which were often lacking in the included studies. Overall, results suggest that differences in dataset composition, sample size, and model architecture influenced the robustness of pooled estimates.

Methodological limitations identified through risk-of-bias assessment—particularly single-center design and unclear case selection—may affect the reliability and generalizability of reported model performance. Studies with narrowly defined or non-random samples can inflate accuracy estimates because models are trained on relatively homogeneous populations that may not reflect real-world clinical variability. In contrast, the greater heterogeneity observed in large-sample studies and DL models likely reflects increased variability in patient characteristics and technical factors, such as MRI acquisition protocols, preprocessing, and network design. Despite these sources of variation, Deeks' funnel plot asymmetry test did not indicate publication bias.

Limitations in the imaging modalities used across studies may further influence diagnostic performance. Most studies relied solely on cranial MRI, although spinal MRI provides critical diagnostic information—such as longitudinally extensive transverse myelitis or conus lesions—that strongly supports antibody-mediated demyelinating diseases. In contrast, optic nerve involvement, common across multiple neuroimmunological disorders, may be less clearly characterized on cranial imaging (13, 14). Comprehensive neuraxial imaging and analysis of larger, clinically representative datasets are therefore essential.

Beyond imaging, disease-specific antibody testing remains central to diagnosing autoimmune encephalitis and antibody-associated demyelinating diseases (15). However, a proportion of patients remain seronegative, requiring diagnosis based on clinical assessment and non-specific ancillary tests (16). Brain biopsy can increase diagnostic accuracy in selected cases but is used infrequently due to procedural risk (17). Because most AI studies have focused on seropositive cases, incorporating clinical parameters into future models may aid in identifying seronegative neuroimmunological disorders. Given that only a few studies included seronegative patients, there is a clear need for future AI research to focus on developing diagnostic models that can accurately identify seronegative cases.

To improve model performance and reduce variability, methodological strategies such as transfer learning and feature-attribution techniques are recommended, particularly for small datasets (18). Appropriate selection of classification algorithms and rigorous validation approaches, including external testing, can enhance model reliability and reduce bias (19, 20). Pre-trained architectures like ResNet have shown strong generalization (21), and interpretability tools such as Grad-CAM can enhance transparency by highlighting relevant MRI regions (22, 23).

Nevertheless, our work has limitations. Despite extensive screening, relatively few studies evaluated autoimmune encephalitis, ADEM, or MOGAD, limiting conclusions about AI performance in these disorders. External validation remained limited, and our meta-analysis was constrained by the predominance of MRI-based models due to the scarcity of research incorporating other modalities.

Future research should prioritize multicenter datasets, integration of clinical variables, and development of interpretable models to enhance diagnostic precision. While traditional diagnostic tools remain indispensable, AI has strong potential to support and augment neuroimmunological assessment in clinical practice.

5 Conclusion

AI approaches show promising potential for differentiating neuroimmunological disorders, with most substantial progress to date in distinguishing MS from NMOSD. Although individual studies often report high performance, our meta-analysis reveals significant heterogeneity driven by differences in study size, dataset composition, and model architecture. Future work should emphasize stronger methodological rigor, consistent external validation, and the integration of clinical and epidemiological variables into diagnostic algorithms. Because antibody testing enables accurate diagnosis for many conditions, AI applications may be particularly valuable for seronegative disorders, where current tools are limited. Overall, our findings offer practical guidance for developing more robust and clinically applicable AI models in neuroimmunology.

Statements

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

DP: Conceptualization, Methodology, Software, Visualization, Writing – original draft. NG: Project administration, Supervision, Writing – review & editing. RK: Project administration, Supervision, Writing – review & editing. DJ: Project administration, Supervision, Writing – review & editing. GK: Project administration, Supervision, Writing – review & editing. MV: Conceptualization, Methodology, Project administration, Writing – original draft.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2025.1579206/full#supplementary-material

References

  • 1.

    Kim SM Kim SJ Lee HJ Kuroda H Palace J Fujihara K . Differential diagnosis of neuromyelitis optica spectrum disorders. Ther Adv Neurol Disord. (2017) 10:26589. doi: 10.1177/1756285617709723

  • 2.

    Solomon AJ Marrie RA Viswanathan S Correale J Magyari M Robertson NP et al . Global barriers to the diagnosis of multiple sclerosis. Neurology. (2023) 101:e62435. doi: 10.1212/WNL.0000000000207481

  • 3.

    Li A Guo K Liu X Gong X Li X Zhou D et al . Limitations on knowledge of autoimmune encephalitis and barriers to its treatment among neurologists: a survey from western China. BMC Neurol. (2023) 23:99. doi: 10.1186/s12883-023-03139-0

  • 4.

    Gaetani M Mazwi M Balaci H Greer R Maratta C . Artificial intelligence in medicine and the pursuit of environmentally responsible science. Lancet Digit Health. (2024) 6:e43840. doi: 10.1016/S2589-7500(24)00090-6

  • 5.

    Umapathy VR Rajinikanth B S Samuel Raj RD Yadav S Munavarah SA Anandapandian PA et al . Perspective of artificial intelligence in disease diagnosis: a review of current and future endeavours in the medical field. Cureus. 15:e.(45684). doi: 10.7759/cureus.45684

  • 6.

    Miller DD Brown EW . Artificial intelligence in medical practice: the question to the answer?Am J Med. (2018) 131:12933. doi: 10.1016/j.amjmed.2017.10.035

  • 7.

    Al-Kawaz M Primiani C Urrutia V Hui F . Impact of RapidAI mobile application on treatment times in patients with large vessel occlusion. J Neurointerventional Surg. (2022) 14:2336. doi: 10.1136/neurintsurg-2021-017365

  • 8.

    Demuth S Paris J Faddeenkov I De Sèze J Gourraud PA . Clinical applications of deep learning in neuroinflammatory diseases: a scoping review. Rev Neurol (Paris). (2024). doi: 10.1016/j.neurol.2024.04.004

  • 9.

    Etemadifar M Norouzi M Alaei SA Karimi R Salari M . The diagnostic performance of AI-based algorithms to discriminate between NMOSD and MS using MRI features: a systematic review and meta-analysis. Mult Scler Relat Disord. (2024) 87:105682. doi: 10.1016/j.msard.2024.105682

  • 10.

    Higgins JPT Thompson SG . Quantifying heterogeneity in a meta-analysis. Stat Med. (2002) 21:153958. doi: 10.1002/sim.1186

  • 11.

    Viechtbauer W . Conducting meta-analyses in R with the metafor package. J Stat Softw. (2010) 36:148. doi: 10.18637/jss.v036.i03

  • 12.

    Whiting PF Rutjes AWS Westwood ME Mallett S Deeks JJ Reitsma JB et al . QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. (2011) 155:52936. doi: 10.7326/0003-4819-155-8-201110180-00009

  • 13.

    Banwell B Bennett JL Marignier R Kim HJ Brilot F Flanagan EP et al . Diagnosis of myelin oligodendrocyte glycoprotein antibody-associated disease: international MOGAD Panel proposed criteria. Lancet Neurol. (2023) 22:26882. doi: 10.1016/S1474-4422(22)00431-8

  • 14.

    Darakdjian M Chaves H Hernandez J Cejas C MRI . pattern in acute optic neuritis: comparing multiple sclerosis, NMO and MOGAD. Neuroradiol J. (2023) 36:26772. doi: 10.1177/19714009221124308

  • 15.

    Prüss H . Autoantibodies in neurological disease. Nat Rev Immunol. (2021) 21:798813. doi: 10.1038/s41577-021-00543-w

  • 16.

    MojŽišová H Krýsl D Hanzalová J Dargvainiene J Wandinger KP Leypoldt F et al . Antibody-negative autoimmune encephalitis. Neurol Neuroimmunol Neuroinflammation. (2023) 10:e200170. doi: 10.1212/NXI.0000000000200170

  • 17.

    Cellucci T Van Mater H Graus F Muscal E Gallentine W Klein-Gitelman MS et al . Clinical approach to the diagnosis of autoimmune encephalitis in the pediatric patient. Neurol Neuroimmunol Neuroinflammation. (2020) 7:e663. doi: 10.1212/NXI.0000000000000663

  • 18.

    Shin HC Roth HR Gao M Lu L Xu Z Nogues I et al . Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging. (2016) 35:128598. doi: 10.1109/TMI.2016.2528162

  • 19.

    Jordan MI Mitchell TM . Machine learning: trends, perspectives, and prospects. Science. (2015) 349:25560. doi: 10.1126/science.aaa8415

  • 20.

    Talaei Khoei T Kaabouch N . Machine learning: models, challenges, and research directions. Future Internet. (2023) 15:332. doi: 10.3390/fi15100332

  • 21.

    He F Liu T Tao D . Why ResNet works? residuals generalize. IEEE Trans Neural Netw Learn Syst. (2020) 31:534962. doi: 10.1109/TNNLS.2020.2966319

  • 22.

    Castelvecchi D . Can we open the black box of AI?Nat News. (2016) 538:20. doi: 10.1038/538020a

  • 23.

    Selvaraju RR Cogswell M Das A Vedantam R Parikh D Batra D . Grad-CAM: visual explanations from deep networks via gradient-based localization. In: 2017 IEEE International Conference on Computer Vision (ICCV). (2017). p. 61826. doi: 10.1109/ICCV.2017.74

  • 24.

    El Khoury Y Gebelin M de Sèze J Patte-Mensah C Marcou G Varnek A et al . Rapid discrimination of neuromyelitis optica spectrum disorder and multiple sclerosis using machine learning on infrared spectra of sera. Int J Mol Sci. (2022) 23:2791. doi: 10.3390/ijms23052791

  • 25.

    Yan Z Liu H Chen X Zheng Q Zeng C Zheng Y et al . Quantitative susceptibility mapping-derived radiomic features in discriminating multiple sclerosis from neuromyelitis optica spectrum disorder. Front Neurosci. (2021) 3:15. doi: 10.3389/fnins.2021.765634

  • 26.

    Clarke L Arnett S Bukhari W Khalilidehkordi E Jimenez Sanchez S O'Gorman C et al . MRI patterns distinguish AQP4 antibody positive neuromyelitis optica spectrum disorder from multiple sclerosis. Front Neurol. (2021) 9:12. doi: 10.3389/fneur.2021.722237

  • 27.

    Huang J Xin B Wang X Qi Z Dong H Li K et al . Multi-parametric MRI phenotype with trustworthy machine learning for differentiating CNS demyelinating diseases. J Transl Med. (2021) 19:377. doi: 10.1186/s12967-021-03015-w

  • 28.

    Gharaibeh M Abedalaziz W Alawad NA Gharaibeh H Nasayreh A El-Heis M et al . Optimal integration of machine learning for distinct classification and activity state determination in multiple sclerosis and neuromyelitis optica. Technologies. (2023) 11:131. doi: 10.3390/technologies11050131

  • 29.

    Ciftci Kavaklioglu B Erdman L Goldenberg A Kavaklioglu C Alexander C Oppermann HM et al . Machine learning classification of multiple sclerosis in children using optical coherence tomography. Mult Scler J. (2022) 28:225362. doi: 10.1177/13524585221112605

  • 30.

    Luo X Li H Xia W Quan C ZhangBao J Tan H et al . Joint radiomics and spatial distribution model for MRI-based discrimination of multiple sclerosis, neuromyelitis optica spectrum disorder, and myelin-oligodendrocyte-glycoprotein-IgG-associated disorder. Eur Radiol. (2024) 34:436475. doi: 10.1007/s00330-023-10529-y

  • 31.

    Ding S Zheng H Wang L Fan X Yang X Huang Z et al . Classification of myelin oligodendrocyte glycoprotein antibody-related disease and its mimicking acute demyelinating syndromes in children using MRI-based radiomics: from lesion to subject. Acad Radiol. (2024) 31:208596. doi: 10.1016/j.acra.2023.11.011

  • 32.

    Wei S Xu L Zhou D Wang T Liu K Gao F et al . Differentiation of MOGAD in ADEM-like presentation children based on FLAIR MRI features. Mult Scler Relat Disord. (2023) 70:104496. doi: 10.1016/j.msard.2022.104496

  • 33.

    Cacciaguerra L Storelli L Radaelli M Mesaros S Moiola L Drulovic J et al . Application of deep-learning to the seronegative side of the NMO spectrum. J Neurol. (2022) 269:154656. doi: 10.1007/s00415-021-10727-y

  • 34.

    Seok JM Cho W Chung YH Ju H Kim ST Seong JK et al . Differentiation between multiple sclerosis and neuromyelitis optica spectrum disorder using a deep learning model. Sci Rep. (2023) 13:11625. doi: 10.1038/s41598-023-38271-x

  • 35.

    Kim H Lee Y Kim YH Lim YM Lee JS Woo J et al . Deep learning-based method to differentiate neuromyelitis optica spectrum disorder from multiple sclerosis. Front Neurol. (2020) 30:11. doi: 10.3389/fneur.2020.599042

  • 36.

    Hagiwara A Otsuka Y Andica C Kato S Yokoyama K Hori M et al . Differentiation between multiple sclerosis and neuromyelitis optica spectrum disorders by multiparametric quantitative MRI using convolutional neural network. J Clin Neurosci. (2021) 87:558. doi: 10.1016/j.jocn.2021.02.018

  • 37.

    Zhuo Z Zhang J Duan Y Qu L Feng C Huang X et al . Automated classification of intramedullary spinal cord tumors and inflammatory demyelinating lesions using deep learning. Radiol Artif Intell. (2022) 4:e210292. doi: 10.1148/ryai.210292

  • 38.

    Wang Z Yu Z Wang Y Zhang H Luo Y Shi L et al . 3D Compressed convolutional neural network differentiates neuromyelitis optical spectrum disorders from multiple sclerosis using automated white matter hyperintensities segmentations. Front Physiol. (2020) 23:11. doi: 10.3389/fphys.2020.612928

  • 39.

    Huang L Shao Y Yang H Guo C Wang Y Zhao Z et al . A joint model for lesion segmentation and classification of MS and NMOSD. Front Neurosci. (2024) 27:18. doi: 10.3389/fnins.2024.1351387

  • 40.

    Huang C Chen W Liu B Yu R Chen X Tang F et al . Transformer-based deep-learning algorithm for discriminating demyelinating diseases of the central nervous system with neuroimaging. Front Immunol. (2022) 14:13. doi: 10.3389/fimmu.2022.897959

  • 41.

    Zhou D Xu L Wang T Wei S Gao F Lai X et al . M-DDC: MRI based demyelinative diseases classification with U-Net segmentation and convolutional network. Neural Netw. (2024) 169:10819. doi: 10.1016/j.neunet.2023.10.010

  • 42.

    Pan J Lv R Wang Q Zhao X Liu J Ai L . Discrimination between leucine-rich glioma-inactivated 1 antibody encephalitis and gamma-aminobutyric acid B receptor antibody encephalitis based on ResNet18. Vis Comput Ind Biomed Art. (2023) 6:17. doi: 10.1186/s42492-023-00144-5

Summary

Keywords

artificial intelligence, deep learning, differential diagnosis, machine learning, neuroimmunology

Citation

Petrosian D, Giedraitiene N, Kizlaitiene R, Jatuzis D, Kaubrys G and Vaisvilas M (2026) Assessing the effectiveness of machine learning and deep learning in differentiating neuroimmunological diseases: a systematic review and meta-analysis. Front. Neurol. 16:1579206. doi: 10.3389/fneur.2025.1579206

Received

18 February 2025

Revised

07 December 2025

Accepted

15 December 2025

Published

12 January 2026

Volume

16 - 2025

Edited by

Roberta Simeoli, University of Naples Federico II, Italy

Reviewed by

Rohan Gupta, Galgotias University, India

Shinya Sonobe, Tohoku University, Japan

Updates

Copyright

*Correspondence: David Petrosian,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics