SYSTEMATIC REVIEW article

Front. Neurol., 03 August 2023

Sec. Stroke

Volume 14 - 2023 | https://doi.org/10.3389/fneur.2023.1211733

Machine learning in the prediction of post-stroke cognitive impairment: a systematic review and meta-analysis

  • 1. Department of Rehabilitation Medicine, The Affiliated Hospital of Yunnan University, Kunming, China

  • 2. Department of Research and Teaching, Lijiang People’s Hospital, Lijiang, China

  • 3. Department of Rehabilitation Medicine, The Third People’s Hospital of Yunnan Province, Kunming, China

  • 4. Key Laboratory of Traditional Chinese Medicine for Prevention and Treatment of Neuropsychiatric Diseases, Yunnan University of Chinese Medicine, Kunming, China

Abstract

Objective:

Cognitive impairment is a detrimental complication of stroke that compromises the quality of life of the patients and poses a huge burden on society. Due to the lack of effective early prediction tools in clinical practice, many researchers have introduced machine learning (ML) into the prediction of post-stroke cognitive impairment (PSCI). However, the mathematical models for ML are diverse, and their accuracy remains highly contentious. Therefore, this study aimed to examine the efficiency of ML in the prediction of PSCI.

Methods:

Relevant articles were retrieved from Cochrane, Embase, PubMed, and Web of Science from the inception of each database to 5 December 2022. Study quality was evaluated by PROBAST, and c-index, sensitivity, specificity, and overall accuracy of the prediction models were meta-analyzed.

Results:

A total of 21 articles involving 7,822 stroke patients (2,876 with PSCI) were included. The main modeling variables comprised age, gender, education level, stroke history, stroke severity, lesion volume, lesion site, stroke subtype, white matter hyperintensity (WMH), and vascular risk factors. The prediction models used were prediction nomograms constructed based on logistic regression. The pooled c-index, sensitivity, and specificity were 0.82 (95% CI 0.77–0.87), 0.77 (95% CI 0.72–0.80), and 0.80 (95% CI 0.71–0.86) in the training set, and 0.82 (95% CI 0.77–0.87), 0.82 (95% CI 0.70–0.90), and 0.80 (95% CI 0.68–0.82) in the validation set, respectively.

Conclusion:

ML is a potential tool for predicting PSCI and may be used to develop simple clinical scoring scales for subsequent clinical use.

Systematic Review Registration:

https://www.crd.york.ac.uk/prospero/display_record.php?RecordID=383476.

1. Introduction

Stroke is a serious condition and a leading cause of death and long-term disability, which places a huge burden worldwide (1). Post-stroke cognitive impairment (PSCI) is a prevalent prognosis and cause of death following a stroke. Stroke patients have a higher incidence of 1-year cognitive impairment than non-stroke populations (2, 3). As society and economy progress, more emphasis is placed on disease and health, especially cognitive impairment. Early identification and diagnosis of PSCI, as well as early prophylaxis and treatment, can help improve stroke patient’s prognosis and reduce social and economic burdens.

Clinical tools for early PSCI diagnosis in stroke patients are currently lacking. Researchers have tried to apply existing cognitive impairment risk prediction models constructed based on the general population to the prediction of PSCI, but their predictive performance was not ideal in stroke patients (4). As a result, researchers have shifted their focus to machine learning (ML) in the hopes of developing more accurate PSCI prediction models. ML is an emerging field in medicine that utilizes computer science and statistics to solve healthcare problems (5). In recent years, ML has been increasingly applied to stroke research, and it was shown that ML-based stroke image prediction can outperform existing prediction tools (6). However, the diversity in mathematical modeling and sensitivity of ML algorithms to factors such as patient sampling, missing data and sample size continue to fuel debates over the accuracy of these models in disease prediction.

The performance of existing stroke prediction models has been inconsistent due to the use of different types of ML (e.g., logistic regression or other alternative) and modeling variables. In these predictive models, there are differences in the types of machine learning utilized, with most researchers using logistic regression while some may consider it lacking and opt for alternative models. Furthermore, we note discrepancies in the selection of modeling variables, which ultimately contributes to the uncertainty of their results. Unfortunately, evidence-based studies investigating the efficiency of ML in the prediction of PSCI are still lacking. As a result, the aim of this study is to examine the predictive accuracy of ML in PSCI and comprehensively summarize the modeling variables included in this prediction model in order to provide a useful reference for the subsequent development of simple clinical prediction tools.

2. Materials and methods

This study was conducted following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines (Supplementary material) (7).

This study has been registered in PROSPERO (CRD42022383476).

2.1. Eligibility criteria

2.1.1. Inclusion criteria

  • Patients diagnosed with ischemic stroke or hemorrhagic stroke.

  • Randomized-controlled trials (RCTs), case–control studies, cohort studies, and case-cohort studies.

  • Complete construction of a ML prediction model for PSCI.

  • Studies without external validation are also included.

  • Different studies published using the same data set.

  • Studies reported in English.

2.1.2. Exclusion criteria

  • Meta-analysis, review, guidelines, and expert opinions.

  • Only risk factor analysis was performed and lacks a complete ML model.

  • Missing outcome measures (ROC, c-statistic, c-index, sensitivity, specificity, accuracy, recovery rate, accuracy rate, confusion matrix, diagnosis table, F1 score, and calibration curve).

  • Validation of the maturity scale only.

  • Study on the accuracy of single-factor prediction models.

2.2. Search strategy

Relevant articles were systematically searched in Cochrane, Embase, PubMed, and Web of Science from the inception of each database to 5 December 2022 using MeSH and entry terms without restriction on language or region. The detailed retrieval process is outlined in Supplementary material.

2.3. Literature screening and data extraction

Retrieved articles were imported into Endnote for management, and duplications were deleted. The titles and abstracts were screened to exclude irrelevant studies, and the full texts of the remaining records were downloaded and checked for eligibility. Data were collected from the included studies using a customized data extraction form. The collected data comprised title, first author, year of publication, author country, type of study, source of patient, type of stroke, diagnostic criteria for cognitive impairment, length of follow-up, number of PSCI cases, total subject number, training set, validation set, type of model used, imputation method for missing value, variable screening, and modeling variables. Two independent researchers (YY and HY) performed the literature screening and data extraction, and subsequently cross-checked their results. Any disagreement was resolved by a third researcher (XSL).

2.4. Risk of bias assessment

The Prediction model Risk Of Bias Assessment Tool (PROBAST) was employed to evaluate the quality of the included studies. The PROBAST consists of four domains, namely participants, predictors, outcome, and analysis (8). The four domains contain 2, 3, 6, and 9 specific questions, respectively. Each question has three options: yes/probably yes (Y/PY), no/probably no (N/PN), and no information (NI). If a domain has at least one N/PN, it is rated as high risk. To be graded as low risk, a given domain must have Y/PY for all questions. When all domains are at low risk, the overall risk of bias is low; alternatively, when at least one domain is assessed as high risk, the overall risk of bias is high (9). Two researchers (XSL and DDY) independently evaluated the risk of bias in the included studies and subsequently cross-checked their results. Any disagreement was resolved by a third researcher (BYW).

2.5. Outcome measures

The primary outcome was the C-index, which can be used to reflect the overall accuracy of ML models. However, this indicator alone may not fully reflect the predictive accuracy of ML models in PSCI because the percentage of PSCI patients and non-PSCI patients in the included literature is severely unbalanced. Therefore, sensitivity and specificity were included as complementary outcome measures to evaluate the predictive accuracy of ML in PSCI.

2.6. Data synthesis and statistical analysis

The c-index and accuracy of ML models were meta-analyzed. If a 95% confidence interval (CI) and standard error were missing for the c-index, they were estimated using the methods by Debray (10). Given the differences in modeling variables and parameters, the c-index was pooled using a random effects model while sensitivity and specificity were pooled by a bivariate mixed effects model. In systematic reviews based on machine learning, heterogeneity is difficult to avoid. According to the Cochrane tool, percentages of around 25% (I2 = 25), 50% (I2 = 50), and 75% (I2 = 75) are deemed to represent low, medium, and high levels of heterogeneity, respectively (11). The sensitivity and robustness of the results were evaluated using the leave-one-out method. Publication bias was qualitatively assessed using a funnel plot and quantitatively assessed by Egger’s regression test (value of p). All meta-analyses were conducted in R4.2.0 (R development Core Team, Vienna, http://www.R-project.org). A p < 0.05 was considered statistically significant.

3. Results

3.1. Study selection

The literature screening process is illustrated in Figure 1. We identified a total of 5, 053 unique records. After reviewing the full texts of 41 reports, 21 studies were ultimately included (12–32).

Figure 1

3.2. Characteristics of included studies

Of the 21 eligible studies, 10 were conducted in China (12, 13, 17, 19, 21, 25, 26, 28, 29, 32), 2 in Norway (14, 27), 1 in Australia (20), 2 in German (15, 16), 2 in the Republic of Korea (18, 30), 1 in the Netherlands (14), 1 in France (23), 1 in the UK (24), and 1 in Singapore (31). These studies were published between 2015 and 2023, predominantly in 2021–2023 (n = 17) (12–27, 32).

Of the 7,822 subjects in the included studies, 2,876 developed PSCI. The subject cohort size ranged from 72 to 22,950. The diagnostic criteria used in the included studies were Mini-Mental State Examination (MMSE) (33), Montreal Cognitive Assessment (MoCA) (34) (n = 5) (20, 21, 28, 29, 31), MMSE (n = 6) (12–14, 18, 24, 32), MoCA (n = 6) (17, 19, 22, 25–27), Center of Cancelation (CoC) (35) (n = 1) (15), Global Deterioration Scale (GDS) (36) (n = 1) (16), Informant Questionnaire on Cognitive Decline in the Elderly (IQCODE) (37) (n = 1) (23), and vascular dementia criteria of the AHA/ASA scientific statement (38) (n = 1) (30). The duration of follow-up was predominantly 3 to 12 months, and was 36 years in only one study (23). The incidence of PSCI during follow-up was 12.5%–66.1%.

3.3. Characteristics of included prediction models

There were 31 models in the included studies, which were constructed based on logistic regression (LR) nomogram (n = 18) (12, 14–16, 18, 19, 22, 25, 26, 29–32), random forest (n = 1) (20), ridge regression (n = 1) (23), LASSO regression (n = 1) (21), mixed effects model (n = 3) (24), support vector machine (SVM) classifier (n = 3) (27), and decision trees (n = 3) (28) (Table 1). Modeling variables were selected using a multivariate approach. In the training set, 15 models reported c-index and 13 models reported sensitivity and specificity. In the validation set, 10 models reported c-index and 7 models reported sensitivity and specificity. The main modeling variables used in the included studies were age, gender, education level, stroke history, stroke severity, lesion volume, lesion site, stroke subtype, and vascular risk factors (Table 2).

Table 1

No.First authorYear of publicationAuthor countryStudy typeSouce of patientsStroke typeDiagnostic criteria for cognitive impairment
1Yinwei Zhu (12)2022ChinaRCTMulticenterAcute ischemic strokeMMSE < 25
2Fei Zha (13)2022ChinaRetrospective cohort studySingle centerCerebral strokeMMSE score ≤ 19 (illiteracy), ≤ 22 (primary education), ≤ 26 (Secondary school and above)
3Georgios Vlachos (14)2023NorwayRetrospective cohort studyMulticenterMild acute strokethe Barthel ADL index and the modified Rankin Scale (mRS).
4Lisa R¨ohrig (15)2022GermanyProspective cohort studyMulticenterRight hemisphere strokeletter cancelation test; bells cancelation test the Center of Cancelation [CoC; (35)]; The CoC
5Ragnhild (16)2022NorwayProspective cohort studyMulticenterCerebral strokePremorbid cognitive status based on GDS
6Zhao-Yin Ma (17)2022ChinaRCTSingle centerAcute ischemic strokeMoCA score ≥ 26 indicates normal cognitive function; < 26 indicates MCI; < 20 indicates CI
7Reeree Lee (18)2021Republic of KoreaProspective cohort studyMulticenterCerebral strokeObjective neuropsychology tests, including MMSE and CDR
8Yongzhe Gu (19)2022ChinaRetrospective cohort studyMulticenterIschemic strokeMoCA score < 26
9Nacim Betroun (20)2022AustraliaProspective cohort studyMulticenterCerebral strokeMMSE score < 27 or MoCA score < 25
10Xueling Yuan (21)2021ChinaRetrospective cohort studySingle centerCerebral strokeMMSE score ≤ 17 (illiteracy), ≤ 20 (Primary education), ≤ 24 (Secondary school and above), MOCA score ≤ 26
11Nick A Weaver (22)2021NetherlandsRetrospective cohort studyMulticenterCerebral strokePerformance below the fifth percentile of local normative data in at least one cognitive domain on the Multidomain Neuropsychological Assessment or the Montreal Cognitive Assessment
12Renaud Lopes (23)2021francRetrospective cohort studyMulticenterCerebral strokeIQCODE 49 ± 2
13Youssef Hbid (24)2021UKProspective cohort studySingle centerFirst occurrence of cerebral strokeMMSE score < 24 or AMT < 8
14Li Gong (25)2021ChinaProspective cohort studyMulticenterMild acute strokeMoCA score < 22
15Yi Dong (26)2021ChinaRetrospective cohort studyMulticenterAcute ischemic strokeMoCA score < 22
16Eva Birgitte Aamodt (27)2021NorwayProspective cohort studyMulticenterCerebral strokeTMT A and B, CERAD, COWAT, MoCA, AD-8, GDS (36), NPI Q, HADS, and the Cornell scale
17Yueli Zhu (28)2020ChinaProspective cohort studyMulticenterFirst occurrence of cerebral strokeMMSE score < 27 and MoCA score < 21
18Zhengbao Zhu (29)2019ChinaProspective cohort studyMulticenterIschemic stroke with elevated blood pressureMMSE score < 27 or MoCA score < 25
19Jae-Sung Lim (30)2017Republic of KoreaProspective cohort studyMulticenterCerebral strokeAHA-ASA Criteria, at least two cognitive defects
20Nagaendran Kandiah (31)2015SingaporeRetrospective cohort studyMulticenterMild acute ischemic strokeMMSE score ≤ 2 5 or MoCA score ≤ 22
21Sheng Ye (32)2022ChinaRetrospective cohort studySingle centerLacunar infarctionMMSE score < 24
Follow-up durationNumber of patients with cognitive impairment after strokeTotal sample sizeNumber of patients with cognitive impairment in training setTotal sample size of training set (deriving set and modeling cohort)Validation set generation method [internal validation (k-fold cross-validation, leave-one-out method, random sampling), external validation (prospective, institutional)]Number of patients with cognitive impairment in validation setTotal sample size of validation setVariable screening/feature selection methodModel type
3 m228599228599No validation setMultivariateLogistic regression
3 m8736758245Random sampling at a ratio of 2:129122MultivariateLogistic regression
12 m2111721117No validation setMultivariateLogistic regression
2710327103No validation setMultivariateLogistic regression
3 m9158991589No validation setNoNoMultivariateLogistic regression
9416194161No validation setNoNoMultivariateLogistic regression
6 m1911019110Random internal validation1270MultivariateLogistic regression
6 m6912369123External validation3860MultivariateLogistic regression
1y77327622625-fold cross validation1565Multivariaterandom forest
1y118376118376External validation75125MultivariateLASSO regression
10633810-fold cross validation1238
15 m1,2862,9501,2862,950External validation107246MultivariateLogistic regression
1,1791,704
12-fold cross validation
36y972972External validation40MultivariateRidge regression
5y1,0002,4681,0002,468External validation204940MultivariateMixed effects model
Internal validation
1y112228112228External validationNo1,000MultivariateMultivariate logistic regression
6 m131383131383External validation281MultivariateMultivariate logistic regression
Internal validation102
3 m125227125227Leave-one-out cross validation227MultivariateClassifier model
6 m6610466104Internal cross validation66104MultivariateDecision tree
3 m340 MMSE scoring638NoMultivariateLogistic regression
422 MoCA scoring
3 m50308NoMultivariateLogistic regression
6 m78209708810-fold cross validation821MultivariateLogistic regression
78209External validation35185
12 m5231338219Random sampling at a ratio of 7:31494MultivariateLogistic regression

General characteristics of included studies.

MMSE, mini-mental state examination; CDR, clinical dementia rating; GDS, global deterioration scale; MoCA, Montreal cognitive assessment; IQCODE, informant questionnaire on cognitive decline in the elderly; AMT, abbreviated mental test; TMT A and B, neuropsychological test battery included Trail making A and B; CERAD, 10 word memory and recall test; COWAT, the controlled oral word association test; AD-8, ascertain dementia 8-item informant questionnaire; NPI Q, neuropsychiatric inventory; HADS, hospital anxiety and depression scale.

Table 2

VariablesFrequency
Age29
Gender26
Educational level24
Diabetes15
Smoking15
Hypertension14
History of stroke11
Stroke severity11
Lesion size10
Hyperlipemia9
Drinking9
Stroke subtype8
Body mass index8
NIHSS7
Site of lesion7
Race7
Coronary heart disease6
NIHSS score6
Atrial fibrillation6
MoCA (baseline)5
MMSE (baseline)5
Antiplatelet before stroke5
Glycosylated hemoglobin (HbA1c)5
Information about cognitive decline in the elderly4
homocysteine4
Number of days after stroke4
APOE ε4 positive3
Fazekas3
mRS (baseline)3
sTREM23
WMH size3
Fluvastatin3
Hypercholesteremia3
Family history of stroke3
Transient ischemic attack, a prestroke vascular risk factor3
Uroclepsia3
Cortical thickness3
Cognitive examination3
Socioeconomic group3
Use of antihypertensive drugs3
Cardiovascular risk factors3
aCL2
aPS2
DS-WMH2
HS2
PV-WMH2
RF2
β2-GPi2
Imaging time (days after the event)2
The time from disease onset to randomization2
Red blood cell distribution width2
Mean corpuscular volume2
Diastolic blood pressure2
Random treatment2
Hemoglobin2
Scan sequence or pattern for infarct segmentation2
Hypoglycemic drugs1
Comorbidity21
Fazekas score1
FBG1
FDG PET DL-based cognitive assessment1
hsCRP1
IQCODE score1
LDL-C1
MTA pathology1
NINDS-CSN 5-minute protocol score1
NLR1
OCSP classification1
SAA1
Imaging features of T1-weighted (T1w) image texture analysis1
OCSP classification and functional level in advanced TOAST1
TOAST classification1
White blood cell count1
White matter hyperintensity1
Premorbid cognitive decline1
At discharge (NIHSS, mRS, Barthel scores)1
Low density lipoprotein1
Triglyceride1
Cysteine proteinase inhibitor1
Country1
Marital status1
Progression of acute stroke1
Memory1
Employment situation1
Anticoagulant drugs1
Infections treated with antibiotics1
1 point for six or more correct answers1
Hexagonal orientation1
Number of intracranial atherosclerotic stenosis1
Incranial volume1
Chronic lacunes1
Uric acid1
Global cortical atrophy and stenosis of large intracranial vessels1
Presence of any APOE-e4 allele1
Visual space function1
Affected vascular area1
Pentaterial memory1
Pre-existing depression1
Fibrinogen1
Myocardial infarction1
Serum albumin1
Language1
Speech fluency raw score1
Executive function1
Stroke classification1
Stroke feature1
Neutrophil-lymphocyte ratio (NLR)1
NCD1

Modeling variables in the included studies.

3.4. Risk of bias assessment

The high risk of bias in the included studies was attributed to the limited sample size, retrospective cohort study, and lack of validation set. Therefore, these attributes should be improved in subsequent model construction. The results of the risk of bias assessment are summarized in Figure 2.

Figure 2

3.5. Meta-analysis

Meta-analysis showed that the training set had a c-index of 0.82 (95% CI 0.77–0.87, n = 15), sensitivity of 0.77 (95% CI 0.72–0.80, n = 13), and specificity of 0.80 (95% CI 0.71–0.86, n = 13). Subgroup analysis of the training set showed that the c-index was 0.81 for LR (95%CI 0.74–0.88, n = 12), 0.80 for mixed effects model (95%CI 0.76–0.81 n = 1), 0.88 for SVM classifier (95%CI 0.84–0.92 n = 1), and 0.84 for decision trees (95%CI 0.77–0.92 n = 1; Figures 3, 4).

Figure 3

Figure 4

The c-index, sensitivity, and specificity of the validation set were 0.82 (95% CI 0.77–0.87, n = 10), 0.82 (95% CI 0.70–0.90, n = 7), and 0.76 (95% CI 0.68–0.82, n = 7), respectively. Subgroup analysis of the validation set showed that the c-index was 0.80 for LR (95%CI 0.75–0.85, n = 8) and 0.89 for LASSO regression (95%CI 0.84–0.93 n = 2; Figures 5, 6).

Figure 5

Figure 6

The follow-up period or the meta-regression based on study design showed that there were no significant differences in the c-index between the training and validation sets, even considering the variations due to different study designs or changes in follow-up time (Figures 7–10; Tables 3, 4).

Figure 7

Figure 8

Figure 9

Figure 10

Table 3

Meta-regressionREML estimate of between-study variance% residual variation due to heterogeneityProportion of between-study variance explained
Number of obstau2I-squared_resAdj R-squared
Training setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation sets
14100.006490.00326793.53%63.72%−1.07%−2.04%
With Knapp-Hartung modification
CindexCoefficientStd. errtP > t[95% conf. interval]
Training setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation sets
Follow_time0.00460610.00680740.00605510.00707850.760.960.4620.364−0.0085867−0.00951570.01779890.0231306
_cons0.78108260.76230260.04200740.061947418.5912.310.0000.0000.68955630.61945160.87260880.9051536

Meta-regression results (follow-up time) in the training and validation sets.

Table 4

Meta-regressionREML estimate of between-study variance% residual variation due to heterogeneityProportion of between-study variance explained
Number of obstau2I-squared_resAdj R-squared
Training setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation sets
15100.0079510.00336395.86%63.50%−3.08%−5.03%
With Knapp-Hartung modification
CindexCoefficientStd. errtP > t[95% conf. interval]
Training setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation setsTraining setsValidation sets
Design0.02650530.01957410.03530830.02979130.750.660.4660.530−0.0497738−0.04912470.10278430.088273
_cons0.75952760.76803370.08286550.07854839.179.780.0000.0000.58050770.5869010.93854760.9491663

Meta-regression results (study design) in the training and validation sets.

3.6. Sensitivity analysis and publication bias

Sensitivity analysis indicated that the results of both the training and validation sets were robust (Supplementary Figures S1, S2). However, the asymmetry in the funnel plot and the results of Egger’s regression test suggest that publication bias may be present in the training set (p = 0.056 for Egger’s regression test), and publication bias is clearly present in the validation set (p = 0.005 for Egger’s regression test; Supplementary Figures S3–S6). There were fewer independent validation cohorts in the included literature, and the presence of multiple independent validation cohorts in the same study may have contributed to publication bias.

4. Discussion

Our meta-analysis of 21 original studies demonstrated that ML may be an ideal tool for predicting PSCI. The training set had a c-index of 0.82 (95% CI 0.77–0.87) and sensitivity and specificity of >70%, indicating considerable predictive accuracy in PSCI. Furthermore, the accuracy of the validation set was not significantly lower than that of the training set, indicating that the ML model has good applicability. Currently, LR is the preferred model in clinical practice because it is simple for generating highly accessible nomograms, such as the nomogram on lymph node metastasis developed by the Sloan-Kettering Cancer Center (39–41). In our study, LR was also the preferred model among researchers as it exhibited comparable c-index performance to other ml algorithms while achieving higher sensitivity and specificity. As a result, we conclude that LR demonstrates satisfactory predictive ability for PSCI in this study.

We found that LR is the primary type of model utilized for predicting stroke. LR is a classification algorithm that aims to establish the relationship between features and probability of specific outcomes (42). ML is commonly used to address issues encountered in clinical practice, with supervised learning and unsupervised learning being the most common approaches. Supervised learning primarily focuses on diagnosing and predicting disease prognosis or progression, which involves the process of training, validation, and testing. The training process involves inserting predictive factors into the model and using the model’s inherent parameter calculation rules (e.g., maximum likelihood estimation, iteration) to estimate the optimal model parameters. Selection of modeling variables (feature selection methods) is crucial for the training process and has been a subject of ongoing debate due to their diversification. Furthermore, validation and testing are crucial for a completed model as they reflect the model’s robustness. Unfortunately, in actual research, most studies lacked effective external validation. The original studies included in our analysis predominantly utilized a supervised ML process with single-factor + multi-factor LR model selection method and performed internal validation through random sampling (43–45).

In our study, the c-index of LR did not significantly lag behind other types of ML models, which demonstrates relatively high sensitivity and specificity. Hence, we believe that LR exhibits promising predictive potential for PSCI.

In addition, we found that the major modeling variables for the ML-based PSCI prediction models were age, gender, education level, white matter hyperintensity (WMH), stroke history, stroke severity, lesion volume, lesion site, stroke subtype, and vascular risk factors. These modeling variables were still mainly based on past identified risk factors (race, age, gender, education level, vascular risk factor, stroke severity, and stroke lesion site and volume) (46), and very few or no newly identified risk factors were used for modeling, such as blood proteins [homocysteine (Hcy), C-reactive protein (CRP), low-density lipoprotein cholesterol (LDL-C), total cholesterol (TC)] that have been recognized as effective biomarkers for PSCI (47), cognitive reserve (CR) (48), activity and participation of stroke survivors (49), and intestinal dysbiosis (50). Therefore, the newly identified risk factors should be prioritized for further validation as their efficacy as modeling variables remains uncertain.

It was reported that common cognitive screening tools have similar predictive accuracy in PSCI. Although the MoCA has significantly better sensitivity in PSCI prediction than other cognitive screening tools, its specificity is less than desirable (51, 52). This demonstrates that there is a lack of effective prediction models for the early screening of PSCI. However, our findings showed that ML has considerably high predictive accuracy (c-index, sensitivity, and specificity) in PSCI and is a promising tool for predicting PSCI.

A recent systematic review indicated that although PSCI has unique risk factors (e.g., Vascular risk factors, lifestyle, overweight, and obesity), it is currently unclear whether the intervention of these risk factors can effectively reduce the incidence of PSCI. Most approaches for lowering PSCI incidence are still dependent on effective prophylaxis for stroke (46). Therefore, effective prediction tools for the early identification and diagnosis of PSCI are urgently needed. Despite the uncertainty in the intervention measures for post-stroke cognitive functions, some researchers found that physical activity intervention and noninvasive brain stimulation can improve post-stroke cognitive functions compared with conventional care (53). Though, the ≥2-year improvement in PSCI after intervention was small (54). Moreover, patients with cognitive impairment have significantly increased risks of subsequent ischemic and fatal stroke (55, 56). Hence, early identification of appropriate treatment and rehabilitation measures are critical for improving the health and life expectancy of PSCI patients. Our study demonstrated the feasibility of ML in the development of PSCI prediction tools and that ML is also an important means for PSCI prediction.

Given the low number of PSCI prediction models for hemorrhagic stroke included in this study (n = 3) (19), the predictive accuracy of ML vs. common cognitive screening tools in PSCI in hemorrhagic stroke patients remains unclear and warrants further investigation.

ML plays an important role in the clinical management of stroke and improvement of the accuracy and efficiency of stroke prediction, diagnosis, personalized treatment, and prognosis assessment (57). For prediction of stroke risk, ML algorithms can be trained using patient data to establish predictive models and estimate the risk of stroke based on individual patient information, clinical indicators, and biomarkers. As for stroke diagnosis, ML can learn and identify radiological features of stroke and assist physicians with early and accurate diagnosis. In addition, ML can predict the efficacy and safety of different treatment options based on the patient’s personal information, medical history, and clinical manifestations, enabling physicians to develop personalized treatment strategies. Furthermore, ML algorithms can predict post-stroke recovery and long-term prognosis based on patients’ clinical and biomarker data. ML has been extensively used in stroke diagnosis, particularly in brain imaging, with SVM being the optimal model for stroke imaging (6, 44, 57). However, in our study, SVM exhibited inferior sensitivity to LR despite higher c-index, and the model size was limited (n = 1). Therefore, further exploration and development of SVM in predicting PSCI are warranted. We can attempt to optimize the accuracy of PSCI prediction by using different SVM models and parameter settings. SVM has various variants, such as non-linear SVM, multi-kernel SVM, and support vector regression, which are selected based on specific circumstances. Additionally, the combination of SVM with other ML methods can be explored for PSCI prediction. For instance, integrating SVM with deep learning techniques can improve the accuracy and robustness of predictions when analyzing images or text data. Moreover, extensive clinical validation studies are required to assess the actual effectiveness of SVM in PSCI prediction. The application value of SVM in PSCI prediction can be comprehensively assessed by collecting more data from PSCI patients and evaluating the models on independent validation sets. In conclusion, SVM, as a widely used ML method, has untapped potential in PSCI prediction. Continuous learning and research efforts can further refine and optimize the application of SVM in PSCI prediction, providing more accurate diagnostic and treatment decision support for clinical practitioners.

For this systematic review, the literature search was performed up until December 2022 and additional studies on this topic may become available subsequently. Hence, a regular review of the literature is recommended to obtain the most updated progress on this research topic.

4.1. Strengths and limitations

This systematic review is the first to demonstrate the feasibility of ML in PSCI prediction. The included models were highly consistent and were predominantly logistic regression nomograms, which minimized heterogeneity.

Despite a comprehensive literature search, the number of included studies and models was still relatively low, and bias may be present in model construction.

5. Conclusion

ML has considerable predictive accuracy and is a promising prediction tool for PSCI. Therefore, future studies should concentrate on constructing ML models based on multi-racial, multi-center, and large-cohort samples and transforming them into simple clinical scoring tools with wide application. This will undoubtedly help with the development of follow-up strategies or rehabilitation measures for stroke patients to reduce their risk of developing cognitive impairment.

Funding

This study was supported by the National Natural Science Foundation of China (31960178 and 82160923), Applied Basic Research Programs of Science and Technology Commission Foundation of Yunnan Province (2019FA007), Key Laboratory of Traditional Chinese Medicine for Prevention and Treatment of Neuropsychiatric Diseases, Yunnan Provincial Department of Education, Scientific Research Projects for High-level Talents of Yunnan University of Chinese Medicine (2019YZG01), Young Top-Notch Talent in 10,000 Talent Program of Yunnan Province (YNWR-QNBJ-2019-235), National Science and Technology Innovation 2030 Major Program (2021ZD0200900), Yunnan Key Research and Development Program (202103AC100005), and Yunnan Province Fabao Gao Expert Workstation Construction Project (202105AF150037).

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Statements

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2023.1211733/full#supplementary-material

Supplementary Figure S1

Sensitivity analysis of the training set.

Supplementary Figure S2

Sensitivity analysis of the validation set.

Supplementary Figure S3

The funnel plot of the training set.

Supplementary Figure S4

The funnel plot of the validation set.

Supplementary Figure S5

Egger’s regression test of the training set.

Supplementary Figure S6

Egger’s regression test of the validation set.

References

  • 1.

    RothGAMensahGAJohnsonCOAddoloratoGAmmiratiEBaddourLMet al. Global burden of cardiovascular diseases and risk factors, 1990-2019: update from the GBD 2019 study. J Am Coll Cardiol. (2020) 76:2982–3021. doi: 10.1016/j.jacc.2020.11.010

  • 2.

    PendleburySTRothwellPM. Incidence and prevalence of dementia associated with transient ischaemic attack and stroke: analysis of the population-based Oxford vascular study. Lancet Neurol. (2019) 18:248–58. doi: 10.1016/S1474-4422(18)30442-3

  • 3.

    LoJWCrawfordJDDesmondDWBaeHJLimJSGodefroyOet al. Long-term cognitive decline after stroke: an individual participant data meta-analysis. Stroke. (2022) 53:1318–27. doi: 10.1161/STROKEAHA.121.035796

  • 4.

    TangEYHPriceCIRobinsonLExleyCKöhlerSStaalsJet al. Assessing the predictive validity of simple dementia risk models in harmonized stroke cohorts. Stroke. (2020) 2095:102. doi: 10.1161/STROKEAHA.120.027473

  • 5.

    DeoRC. Machine learning in medicine. Circulation. (2015) 132:1920–30. doi: 10.1161/CIRCULATIONAHA.115.001593

  • 6.

    MouridsenKThurnerPZaharchukG. Artificial intelligence applications in stroke. Stroke. (2020) 51:2573–9. doi: 10.1161/STROKEAHA.119.027479

  • 7.

    PageMJMcKenzieJEBossuytPMBoutronIHoffmannTCMulrowCDet al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. (2021) 372:n71. doi: 10.1136/bmj.n71

  • 8.

    WolffRFMoonsKGMRileyRDWhitingPFWestwoodMCollinsGSet al. PROBAST: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. (2019) 170:51–8. doi: 10.7326/M18-1376

  • 9.

    BellouVBelbasisLKonstantinidisAKTzoulakiIEvangelouE. Prognostic models for outcome prediction in patients with chronic obstructive pulmonary disease: systematic review and critical appraisal. BMJ. (2019) 367:l5358. doi: 10.1136/bmj.l5358

  • 10.

    DebrayTPDamenJARileyRDSnellKReitsmaJBHooftLet al. A framework for meta-analysis of prediction model studies with binary and time-to-event outcomes. Stat Methods Med Res. (2019) 28:2768–86. doi: 10.1177/0962280218785504

  • 11.

    Huedo-MedinaTBSánchez-MecaJMarín-MartínezFBotellaJ. Assessing heterogeneity in meta-analysis: Q statistic or I2 index?Psychol Methods. (2006) 11:193–206. doi: 10.1037/1082-989X.11.2.193

  • 12.

    ZhuYWZhaoYLuYLFangCQZhangQZhangJTet al. The association between plasma soluble triggering receptor expressed on myeloid cells 2 and cognitive impairment after acute ischemic stroke. J Affect Disord. (2022) 299:287–93. doi: 10.1016/j.jad.2021.12.011

  • 13.

    ZhaFZhaoJJChenCJiXQLiMWuYYet al. A high neutrophil-to-lymphocyte ratio predicts higher risk of poststroke cognitive impairment: development and validation of a clinical prediction model. Front Neurol. (2022) 12:5011. doi: 10.3389/fneur.2021.755011

  • 14.

    VlachosGIhle-HansenHWyllerTBBrækhusAMangsetMHamreCet al. Predictors of cognitive and emotional symptoms 12 months after first-ever mild stroke. Neuropsychol Rehabil. (2023) 33:662–79. doi: 10.1080/09602011.2022.2038211

  • 15.

    RöhrigLSperberCBonilhaLRordenCKarnathHO. Right hemispheric white matter hyperintensities improve the prediction of spatial neglect severity in acute stroke. NeuroImage Clin. (2022):36. doi: 10.1016/j.nicl.2022.103265

  • 16.

    Munthe-KaasRAamSSaltvedtIWyllerTBPendleburySTLydersenSet al. Is frailty index a better predictor than pre-stroke modified Rankin scale for neurocognitive outcomes 3-months post-stroke?BMC Geriatr. (2022) 22:139. doi: 10.1186/s12877-022-02840-y

  • 17.

    MaZYWuYYCuiHYLYaoGYBianH. Factors influencing post-stroke cognitive impairment in patients with type 2 diabetes mellitus. Clin Interv Aging. (2022) 17:653–64. doi: 10.2147/CIA.S355242

  • 18.

    LeeRChoiHParkKYKimJMSeokJW. Prediction of post-stroke cognitive impairment using brain FDG PET: deep learning-based approach. Eur J Nucl Med Mol Imaging. (2022) 49:1254–62. doi: 10.1007/s00259-021-05556-0

  • 19.

    GuYZWangFGongLFangMLiuXY. A nomogram incorporating red blood cell indices to predict post-stroke cognitive impairment in the intracerebral hemorrhage population. Front Aging Neurosci. (2022):14. doi: 10.3389/fnagi.2022.985386

  • 20.

    BetrouniNJiangJYDueringMGeorgakisMKOestreichLSachdevPSet al. Texture features of magnetic resonance images predict poststroke cognitive impairment: validation in a Multicenter study. Stroke. (2022) 53:3446–54. doi: 10.1161/STROKEAHA.122.039732

  • 21.

    YuanXLZhangLSuiRBWangZ. A risk prediction model of post-stroke cognitive impairment based on magnetic resonance spectroscopy imaging. Neurol Res. (2021) 43:642–52. doi: 10.1080/01616412.2021.1908659

  • 22.

    WeaverNAKuijfHJAbenHPAbrigoJBaeHJBarbayMet al. Strategic infarct locations for post-stroke cognitive impairment: a pooled analysis of individual patient data from 12 acute ischaemic stroke cohorts. Lancet Neurol. (2021) 20:448–59. doi: 10.1016/S1474-4422(21)00060-0

  • 23.

    LopesRBournonvilleCKuchcinskiGDondaineTMendykAMViardRet al. Prediction of long-term cognitive functions after minor stroke, using functional connectivity. Neurology. (2021). doi: 10.1186/s12883-021-02350-1

  • 24.

    HbidYFaheyMWolfeCDAObaidMDouiriA. Risk prediction of cognitive decline after stroke. J Stroke Cerebrovasc Dis. (2021) 30:105849. doi: 10.1016/j.jstrokecerebrovasdis.2021.105849

  • 25.

    GongLWangHZhuXDongQYuQMaoBet al. Nomogram to predict cognitive dysfunction after a minor ischemic stroke in hospitalized-population. Front Aging Neurosci. (2021) 13:637363. doi: 10.3389/fnagi.2021.637363

  • 26.

    DongYDingMCuiMFangMGongLXuZet al. Development and validation of a clinical model (DREAM-LDL) for post-stroke cognitive impairment at 6 months. Aging. (2021) 13:21628–41. doi: 10.18632/aging.203507

  • 27.

    AamodtEBSchellhornTStageESanjayABLoganPESvaldiDOet al. Predicting the emergence of major neurocognitive disorder within three months after a stroke. Front Aging Neurosci. (2021) 13:705889. doi: 10.3389/fnagi.2021.705889

  • 28.

    ZhuYZhaoSFanZLiZHeFLinCet al. Evaluation of the mini-mental state examination and the Montreal cognitive assessment for predicting post-stroke cognitive impairment during the acute phase in Chinese minor stroke patients. Front Aging Neurosci. (2020) 12:236. doi: 10.3389/fnagi.2020.00236

  • 29.

    ZhuZZhongCGuoDBuXXuTGuoLet al. Multiple biomarkers covering several pathways improve predictive ability for cognitive impairment among ischemic stroke patients with elevated blood pressure. Atherosclerosis. (2019) 287:30–7. doi: 10.1016/j.atherosclerosis.2019.05.028

  • 30.

    LimJSOhMSLeeJHJungSKimCJangMUet al. Prediction of post-stroke dementia using NINDS-CSN 5-minute neuropsychology protocol in acute stroke. Int Psychogeriatr. (2017) 29:1–8. doi: 10.1017/S1041610216002520

  • 31.

    KandiahNChanderRJLinXNgAPohYYCheongCYet al. Cognitive impairment after mild stroke: development and validation of the SIGNAL2 risk score. J Alzheimers Dis. (2015) 49:1169–77. doi: 10.3233/JAD-150736

  • 32.

    YeSPanHQLiWJWangBXingJJXuL. High serum amyloid a predicts risk of cognitive impairment after lacunar infarction: development and validation of a nomogram. Front Neurol. (2022) 13:972771. doi: 10.3389/fneur.2022.972771

  • 33.

    FolsteinMFFolsteinSEMcHughPR. "mini-mental state". A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. (1975) 12:189–98. doi: 10.1016/0022-3956(75)90026-6

  • 34.

    ChanEAltendorffSHealyCWerringDJCipolottiL. The test accuracy of the Montreal cognitive assessment (MoCA) by stroke lateralisation. J Neurol Sci. (2017) 373:100–4. doi: 10.1016/j.jns.2016.12.028

  • 35.

    RordenCKarnathHO. A simple measure of neglect severity. Neuropsychologia. (2010) 48:2758–63. doi: 10.1016/j.neuropsychologia.2010.04.018

  • 36.

    ReisbergBFerrisSHde LeonMJCrookT. The global deterioration scale for assessment of primary degenerative dementia. Am J Psychiatry. (1982) 139:1136–9. doi: 10.1176/ajp.139.9.1136

  • 37.

    JormAF. A short form of the informant questionnaire on cognitive decline in the elderly (IQCODE): development and cross-validation. Psychol Med. (1994) 24:145–53. doi: 10.1017/S003329170002691X

  • 38.

    GorelickPBScuteriABlackSEDecarliCGreenbergSMIadecolaCet al. Vascular contributions to cognitive impairment and dementia: a statement for healthcare professionals from the american heart association/american stroke association. Stroke. (2011) 42:2672–713. doi: 10.1161/STR.0b013e3182299496

  • 39.

    KuoYLChenWCYaoWJChengLHsuHPLaiHWet al. Validation of Memorial Sloan-Kettering Cancer Center nomogram for prediction of non-sentinel lymph node metastasis in sentinel lymph node positive breast cancer patients an international comparison. Int J Surg. (2013) 11:538–43. doi: 10.1016/j.ijsu.2013.05.005

  • 40.

    VieniSGraceffaGLa MendolaRLatteriSCordovaALatteriMAet al. Application of a predictive model of axillary lymph node status in patients with sentinel node metastasis from breast cancer. A retrospective cohort study. Int J Surg. (2016) 35:58–63. doi: 10.1016/j.ijsu.2016.09.015

  • 41.

    CagiannosIKarakiewiczPEasthamJAOhoriMRabbaniFGerigkCet al. A preoperative nomogram identifying decreased risk of positive pelvic lymph nodes in patients with prostate cancer. J Urol. (2003) 170:1798–803. doi: 10.1097/01.ju.0000091805.98960.13

  • 42.

    LandwehrNHallMFrankE. Logistic model trees. Mach Learn. (2005) 59:161–205. doi: 10.1007/s10994-005-0466-3

  • 43.

    ChoiRYCoynerASKalpathy-CramerJChiangMFCampbellJP. Introduction to machine learning, neural networks, and deep learning. Transl Vis Sci Technol. (2020) 9:14. doi: 10.1167/tvst.9.2.14

  • 44.

    JiangFJiangYZhiHDongYLiHMaSet al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol. (2017) 2:230–43. doi: 10.1136/svn-2017-000101

  • 45.

    HandelmanGSKokHKChandraRVRazaviAHLeeMJAsadiH. eDoctor: machine learning and the future of medicine. J Intern Med. (2018) 284:603–19. doi: 10.1111/joim.12822

  • 46.

    RostNSBrodtmannAPaseMPvan VeluwSJBiffiADueringMet al. Post-stroke cognitive impairment and dementia. Circ Res. (2022) 130:1252–71. doi: 10.1161/CIRCRESAHA.122.319951

  • 47.

    KimKYShinKYChangKA. Potential biomarkers for post-stroke cognitive impairment: a systematic review and meta-analysis. Int J Mol Sci. (2022) 23:602. doi: 10.3390/ijms23020602

  • 48.

    ContadorIAlzolaPSternYde la Torre-LuqueABermejo-ParejaFFernández-CalvoB. Is cognitive reserve associated with the prevention of cognitive decline after stroke? A systematic review and meta-analysis. Ageing Res Rev. (2023) 84:101814. doi: 10.1016/j.arr.2022.101814

  • 49.

    StolwykRJMihaljcicTWongDKChapmanJERogersJM. Poststroke cognitive impairment negatively impacts activity and participation outcomes: a systematic review and meta-analysis. Stroke. (2021) 52:748–60. doi: 10.1161/STROKEAHA.120.032215

  • 50.

    KoszewiczMJarochJBrzeckaAEjmaMBudrewiczSMikhalevaLMet al. Dysbiosis is one of the risk factor for stroke and cognitive impairment and potential target for treatment. Pharmacol Res. (2021) 164:105277. doi: 10.1016/j.phrs.2020.105277

  • 51.

    ShiDChenXLiZ. Diagnostic test accuracy of the Montreal cognitive assessment in the detection of post-stroke cognitive impairment under different stages and cutoffs: a systematic review and meta-analysis. Neurol Sci. (2018) 39:705–16. doi: 10.1007/s10072-018-3254-0

  • 52.

    LeesRSelvarajahJFentonCPendleburySTLanghornePStottDJet al. Test accuracy of cognitive screening tests for diagnosis of dementia and multidomain cognitive impairment in stroke. Stroke. (2014) 45:3008–18. doi: 10.1161/STROKEAHA.114.005842

  • 53.

    O'DonoghueMLeahySBolandPGalvinRMcManusJHayesS. Rehabilitation of cognitive deficits poststroke: systematic review and meta-analysis of randomized controlled trials. Stroke. (2022) 53:1700–10. doi: 10.1161/STROKEAHA.121.034218

  • 54.

    SaaJPTseTBaumCMCummingTJosmanNRoseMet al. Cognitive recovery after stroke: a meta-analysis and Metaregression of intervention and cohort studies. Neurorehabil Neural Repair. (2021) 35:585–600. doi: 10.1177/15459683211017501

  • 55.

    LeeMSaverJLHongKSWuYLLiuHCRaoNMet al. Cognitive impairment and risk of future stroke: a systematic review and meta-analysis. CMAJ Can Med Assoc J. (2014) 186:E536–46. doi: 10.1503/cmaj.140147

  • 56.

    KwonHSLeeDLeeMHYuSLimJSYuKHet al. Post-stroke cognitive impairment as an independent predictor of ischemic stroke recurrence: PICASSO sub-study. J Neurol. (2020) 267:688–93. doi: 10.1007/s00415-019-09630-4

  • 57.

    SirsatMSFerméECâmaraJ. Machine learning for brain stroke: a review. J Stroke Cerebrovasc Dis. (2020) 29:105162. doi: 10.1016/j.jstrokecerebrovasdis.2020.105162

Summary

Keywords

cognitive impairment, prediction, machine learning, stroke, meta-analysis

Citation

Li X, Chen Z, Jiao H, Wang B, Yin H, Chen L, Shi H, Yin Y and Qin D (2023) Machine learning in the prediction of post-stroke cognitive impairment: a systematic review and meta-analysis. Front. Neurol. 14:1211733. doi: 10.3389/fneur.2023.1211733

Received

29 April 2023

Accepted

20 July 2023

Published

03 August 2023

Volume

14 - 2023

Edited by

Giulia Abate, University of Brescia, Italy

Reviewed by

Lei Wang, The Ohio State University, United States; Shubham Misra, Yale University, United States

Updates

Copyright

*Correspondence: Hongling Shi, Yong Yin, Dongdong Qin,

†These authors have contributed equally to this work

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics