- 1Department of Neurosurgery, West China Hospital, Sichuan University, Chengdu, China
- 2Department of Neurosurgery, West China Hospital, Sichuan University/West China School of Nursing, Sichuan University, Chengdu, China
- 3Department of Cerebrovascular Disease, The People’s Hospital of Leshan, Leshan, Sichuan, China
Background: Post-stroke depression (PSD) is a prevalent neuropsychological consequence of stroke, associated with cognitive decline, disability, and increased mortality. Early prediction of PSD is critical for timely interventions and better outcomes. This study evaluates the effectiveness of various clinical prediction models, particularly machine learning methods, in forecasting PSD.
Methods: A systematic review and meta-analysis were conducted to evaluate predictive models for post-stroke depression (PSD) using data from 16 studies. The databases searched included PubMed, Embase, Cochrane Library, and Web of Science, covering publications from the year 2000 to the present. The risk of bias in the predictive models was assessed using the Prediction model Risk of Bias Assessment Tool+AI (PROBAST+AI). The review encompassed both traditional statistical methods and machine learning algorithms. Predictive accuracy was analyzed through the area under the curve (AUC) values of these models, considering various data sources, such as clinical, cognitive, and biomarker data. R packages, including ‘metafor,’ ‘meta,’ and ‘forestplot,’ were used for the analysis.
Results: 16 studies were included. Neural Network models yielded the highest pooled AUC (0.88, 95% CI: 0.45–0.98), although this estimate was based on only two studies and exhibited wide confidence intervals. Logistic Regression, Decision Tree, and K-Nearest Neighbor models showed comparable predictive ability, with pooled AUC values ranging from 0.77 to 0.83. Support Vector Machine models demonstrated the lowest predictive performance (AUC = 0.68) but exhibited the lowest heterogeneity. Among different data sources, functional, physical, and cognitive assessments yielded the highest predictive accuracy (AUC = 0.86, 95% CI: [0.81, 0.90]), followed by biomarker-based models (AUC = 0.80, 95% CI: [0.71, 0.86]). Within retrospective studies, biomarker-based data sources demonstrated significantly superior predictive performance (AUC = 0.94, 95% CI: 0.92–0.96).
Conclusions: Machine learning models, particularly Neural Networks, show potential for predicting post-stroke depression, although current evidence is limited by small sample sizes and high heterogeneity. Traditional approaches such as Logistic Regression and Decision Tree models also demonstrate stable and competitive performance. Among data sources, functional, physical, and cognitive assessments provide the strongest predictive value, while biomarker-based models appear particularly effective in retrospective analyses. Despite these findings, the limited number of high-quality studies and methodological inconsistencies highlight the need for rigorous, prospective, and multicenter validation to establish reliable and generalizable predictive models for post-stroke depression.
Systematic Review Registration: https://www.crd.york.ac.uk/prospero/, identifier CRD42025635227.
Introduction
Depression is a frequent and severe neuropsychological consequence of stroke, strongly linked to cognitive decline, higher levels of disability, and increased risk of death (1–3). Post-stroke depression (PSD) may arise weeks, months, or even years after the onset of a stroke. Its progression is dynamic, with higher rates observed in the acute phase and a gradual decline over time. The incidence of early-onset PSD (occurring within two weeks of stroke) varies between 17.6% and 59.2%, likely due to differences in study populations, PSD definitions, and evaluation methods (4–7). One meta-analysis showed that a hazard ration for PSD and all-cause mortality was 1.59 (3). Hence, the early prediction of the risk of developing PSD holds crucial clinical significance, as it can assist healthcare professionals in assessing a patient’s condition and planning timely interventions. However, it remains unclear which clinical prediction model and data source offer the best performance for this purpose.
In recent years, the advancement of clinical prediction models, including machine learning, has been extensively applied in personalized medicine and decision support systems (8). For example, a recent study showed machine learning analysis successfully demonstrated the ability to predict clinical outcomes after first-ever ischemic stroke and identified the leading prognostic factors that contribute to this prediction (9).
Numerous studies have successfully demonstrated the ability of these models, particularly those incorporating machine learning, to accurately predict specific clinical outcomes following a stroke. Research on clinical prediction models for Post-Stroke Depression (PSD) has explored various model types, including Decision Trees, XGBoost, and Random Forest, as well as data sources such as sociological data and liver function test (7, 9–12). the predictive performance among these different models has not been fully compared. Therefore, we carried out this systematic review and meta-analysis to estimate the clinical prediction models for PSD.
Methods
This systematic review was conducted in accordance with the guidelines outlined in the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA 2020) statement (13). A protocol was not prepared.
Information sources and search strategy
For this systematic review and meta-analysis, we conducted a comprehensive and systematic search across four major databases: PubMed, Embase, Cochrane Library, and Web of Science (Publication year: 2000-present). The final search was conducted on January 5, 2025. Our search strategy combined subject headings and free-text keywords, ensuring no restrictions on region or language. The search terms were specifically designed to include keywords and medical subject headings (MeSH) related to post-stroke depression, clinical prediction models, and meta-analysis. The search results were refined and combined using the “AND” logical operator to create the final retrieval dataset. Detailed information on the search strategy of Pubmed was shown as example, and was provided in Supplementary Table S1.
We used the PICOTS system for the systematic review, which allows framing of the review’s aim, search strategy, and study inclusion and exclusion criteria, as described below.
P (Population): Patients with PSD.
I (Index models): Clinical prediction models including machine learning models and Logistic Regression.
C (Comparator): Other Clinical prediction models.
O (Outcome): Prediction accuracy for PSD.
T (Timing): The outcome was predicted after evaluating basic information upon admission, clinical scoring scale results, and laboratory indicators.
S (Setting): The intended use of the risk prediction models was to individualize the prediction of PSD.
Inclusion and exclusion criteria
Inclusion criteria
Studies were included if they focused on patients with stroke, including ischemic stroke, hemorrhagic stroke, or mixed stroke types, and utilized prediction models (traditional statistical methods or machine learning) to predict the occurrence of PSD. Eligible studies included cohort studies, case-control studies, case-cohort studies, and nested case-control studies. Models could focus on predicting PSD at any time point (acute, subacute, or chronic phases), and studies without independent validation datasets were also included. Additionally, studies utilizing machine learning approaches published on the same dataset were eligible, provided they reported at least one model performance metric, such as ROC, AUC, C-index, sensitivity, specificity, accuracy, calibration curves, or F1 score. Only original studies published in English were included, regardless of whether the exact timing or classification of PSD onset was specified.
Exclusion criteria
Studies were excluded if they were meta-analyses, reviews, guidelines, editorials, or expert opinions, or if they solely analyzed risk factors without developing or validating prediction models. Research lacking essential predictive accuracy indicators, such as ROC, C-index, sensitivity, specificity, or calibration curves, was excluded, as were studies with small sample sizes (<50 cases) or insufficient effective events (EPV < 10). Studies focusing primarily on the validation of assessment scales, the prediction of single factors, case series, case reports, randomized controlled trials, descriptive studies, or research related to pediatric populations (patients under 18 years of age) were also excluded.
Study selection and data extraction
To ensure the integrity of the data, H.W. and Y.X. independently conducted the data extraction process. Also, PROBAST+AI tool was applied independently by the two authors. Discrepancies were resolved through discussion or by consulting L.M. In detail, The retrieved articles were imported into EndNote X9, and duplicate records were removed. Titles and abstracts were reviewed to exclude studies that did not meet the eligibility criteria. Full-text articles were then downloaded and assessed to identify studies eligible for inclusion in the systematic review. If studies reported missing summary statistics (such as missing values for AUC, sensitivity, or specificity), the reviewers attempted to contact the authors to request the missing data. If contacting authors was not feasible, studies with missing essential information were excluded from the analysis. The key characteristics of the included studies are summarized in Supplementary Table S2.
Quality assessment and publication bias
We assessed the risk of bias in predictive models using the Prediction model Risk of Bias Assessment Tool +AI (PROBAST+AI) (14). PROBAST+AI is an updated tool for evaluating the quality, risk of bias, and applicability of prediction models, including AI/machine learning models. It follows four steps: defining the model’s purpose using PICOTS criteria, classifying the study type (model development, evaluation, or both), assessing quality or risk of bias for each domain (e.g., study design, predictors, outcomes), and making an overall judgment on the model’s quality, risk of bias, and applicability. This tool improves upon PROBAST-2019 and can be applied to both traditional and AI-based models, helping in the critical appraisal of predictive modeling studies. The publication bias were evaluated by Begg test.
Data synthesis and statistical analyses
We conducted a meta-analysis to evaluate the AUC indicators of predictive models, including SVM, Decision Tree, Logistic Regression, Neural Network, K-nearest Neighbor. We also evaluate the AUC indictors of different data source including sociological data and clinical data, liver function test, Functional, physical and cognitive tests. A random-effects model was prioritized for the meta-analysis of AUC. The analysis was performed using R version 4.2.0 (R Development Core Team, Vienna, http://www.R-project.org). The R packages employed for this process included ‘metafor,’ ‘meta,’ and ‘forestplot.’ The possible causes of heterogeneity among study results will be explored by subgroup analysis. A sensitivity analysis was performed using the leave-one-out method.
Results
Study characteristics
Major data sources were PubMed, Embase, Cochrane, and Web of Science databases. The literature screening procedure is shown in Figure 1. There were 20 studies included in this meta-analysis, which incorporated both machine learning models and the Logistic Regression Model. Specifically, the machine learning models included Support Vector Machine (SVM) (9, 15, 16); Gradient Boosting (9), Light Gradient Boosting (LightGBM) (17, 18), Extreme Gradient Boosting (XGBoost) (15, 17–19), CatBoost (15, 17, 18), Gradient Boosting Decision Tree (GBDT) (15, 17), Random Forest (15–17, 20), AdaBoost (17), decision tree (17), Artificial Neural Network (ANN) (21), Log-Linearized Gaussian Mixture Network (22) and K-nearest neighbor (16); and linear regression methods and extensions such as Stepwise Multiple Linear Regression (22)and Partial Least Squares (PLS) regression (22). Logistic regression model included logistic regression analysis (12, 17, 18, 22–30), binary logistic regression analysis (31) and multivariate logistic regression model (32).
Figure 1. PRISMA flow diagram depicting the study selection process. This diagram illustrates the identification, screening, eligibility, and inclusion of studies in the systematic review and meta-analysis.
In these included studies, multiple data sources were utilized to predict the occurrence of PSD. According to the data sources, we categorized the articles into five groups: sociological and clinical data (3, 9, 12, 17, 19, 23, 25–27, 29, 32, 33), liver function tests (15), functional, physical, and cognitive assessments (16, 18, 21, 22), and biomarkers (17, 20, 24, 25, 28, 30, 31, 33). The detail of included studies were listed in Supplementary Table S2.
Risk of bias in included studies
The risk of bias in prediction models was assessed using the PROBAST tool (34), with a focus on four key domains: predictors, participants, outcomes, and analysis. The detailed findings from the assessment are presented in Supplementary Table S3. Five studies (highlighted in red) were excluded due to unclear risk of bias and applicability concerns. The result of the Begg test (z = 0.87, p-value = 0.3847; Bias estimate = 77.0000, SE = 88.5818) indicates that there is no statistically significant evidence of publication bias.
Meta-analysis
We first compared the predictive performance of different models using a random-effects model. The analysis revealed the following AUC values with corresponding 95% confidence intervals (CI): the Support Vector Machine (SVM) group achieved an AUC of 0.68 (95% CI: [0.66, 0.70]), the Decision Tree group had an AUC of 0.83 (95% CI: [0.76, 0.88]), the Logistic Regression group showed an AUC of 0.79 (95% CI: [0.74, 0.82]), the Neural Network group achieved the highest AUC of 0.88 (95% CI: [0.45, 0.98]), and the K-Nearest Neighbor group demonstrated an AUC of 0.80 (95% CI: [0.76, 0.83]). Among these models, the Neural Network group appeared to demonstrate the highest predictive performance based on the AUC, although this finding was derived from only two studies and was accompanied by wide confidence intervals. Logistic Regression, Decision Tree, and K-Nearest Neighbor models demonstrated comparable performance, with AUC values ranging from 0.77 to 0.78. In contrast, the SVM group showed the lowest predictive performance with an AUC of 0.68. However, the SVM group have the lowest I2. These findings suggest that Neural Network models may provide superior predictive accuracy, while traditional models such as Logistic Regression and Decision Trees also perform well in this context (Figure 2). The stability and generalizability of this result still need further verification due to the limited number of studies included and the significant heterogeneity observed.
Figure 2. Forest plot of the predictive performance of different models for PSD. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
We then a comprehensive analysis of the predictive performance of models for PSD based on different data sources, including sociological and clinical data, liver function tests, functional, physical, and cognitive assessments, and biomarkers. The predictive performance of each data source group is expressed through AUC values and their corresponding 95% confidence intervals (CI), with heterogeneity analysis conducted. A random-effects model is preferred for the subgroup analysis. The results indicate the following predictive performances: the sociological and clinical data group (AUC = 0.76, 95% CI: [0.68, 0.82]), the liver function test group (AUC = 0.73, 95% CI: [0.70, 0.75]), the functional, physical, and cognitive assessment group (AUC = 0.86, 95% CI: [0.85, 0.86]), and the biomarker group (AUC = 0.80, 95% CI: [0.71, 0.83]). Moreover, these findings highlight the variations in predictive performance across different data sources, with the functional, physical, and cognitive assessment group showing the highest AUC, indicating superior predictive ability (Figure 3).
Figure 3. Forest plot of the predictive performance of different data sources for PSD. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
Stratified meta-analysis
In the included studies, there were three studies (9, 19, 21), involved patients with new-onset stroke. These studies were excluded from the subsequent analyses. Among the remaining models, the Neural Network group demonstrated the highest predictive performance based on AUC, although this result was derived from a single study. The model with the second-highest predictive performance was the Decision Tree (Figure 4). Similarly, after excluding the three studies mentioned above, we evaluated the predictive performance of models for PSD across different data sources and found no statistically significant differences among the groups (Figure 5).
Figure 4. Forest plot illustrating the predictive performance of different models for PSD (excluding studies involving new-onset stroke). The plot presents the area under the curve (AUC) values along with their 95% confidence intervals for each included study.
Figure 5. Forest plot illustrating the predictive performance of different data sources for PSD (excluding studies involving new-onset stroke). The plot presents the area under the curve (AUC) values along with their 95% confidence intervals for each included study.
Next, we categorized all included studies into retrospective and prospective designs and excluded the cross-sectional study (22). We then analyzed the predictive performance of PSD models based on study design (retrospective vs. prospective) and data source (retrospective vs. prospective). In the retrospective studies, no statistically significant differences were observed among the model groups (Figure 6). Similarly, in the prospective studies, no statistically significant differences were found among the model groups (Figure 7). In the retrospective studies, the biomarker-based data source group demonstrated the best predictive performance, with a common-effect AUC of 0.94 (95% CI: [0.92, 0.96]), showing a slight but notable statistical difference (Figure 8). However, in the prospective studies, no statistically significant differences were observed among the data source groups (Figure 9).
Figure 6. Predictive performance of different model types in retrospective studies. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
Figure 7. Predictive performance of different model types in prospective studies. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
Figure 8. Predictive performance of different data source groups in retrospective studies. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
Figure 9. Predictive performance of different data source groups in prospective studies. The plot shows area under the curve (AUC) values with their 95% confidence intervals for each included study.
Discussion
This meta-analysis compared the predictive performance of various machine learning models and data sources for PSD using a random-effects model. The findings revealed significant variations in predictive accuracy across different models, highlighting both the strengths and limitations of current methodologies. This study may provide evidence for clinicians to choose the proper model or data source to predict PSD. To the best of our knowledge, this is the first meta-analysis combining AI and traditional models in predicting PSD.
Usually, AUCs of 0.5–0.7 were considered with low diagnostic accuracy, 0.7–0.9 were considered with moderate accuracy, and >0.9 indicates high accuracy. The comparison of machine learning models demonstrated that neural networks achieved the highest predictive performance with an AUC of 0.88 (95% CI: [0.94, 0.96]) under the random-effects model, suggesting its potential as a superior predictive tool for PSD. Despite the promising results of neural networks, the limited number of studies and study type (cross-sectional study) underscores the need for further research to enhance the stability and generalizability of these models.
In contrast, the Support Vector Machine (SVM) group exhibited the lowest predictive performance (AUC = 0.68, 95% CI: [0.66, 0.70]), suggesting limitations in its application to PSD prediction. Interestingly, the SVM groups showed a low heterogeneity (I2 = 0%), indicating that the studies included in this subgroup produced very similar effect sizes, suggesting consistency in the SVM model’s performance across different studies.
These findings highlight the importance of selecting appropriate machine learning models for PSD prediction, with neural networks demonstrating the most potential but requiring further refinement and validation.
Regarding data sources, the functional, physical, and cognitive assessment group demonstrated the highest predictive performance (AUC = 0.86, 95% CI: [0.85, 0.86]), emphasizing its critical role in predicting PSD. This strong performance may stem from the direct relationship between functional and cognitive impairments and PSD. Biomarker-based studies, while less predictive overall (AUC = 0.77, 95% CI: [0.76, 0.78]), revealed the potential value of inflammatory and metabolism-related indicators as predictors of depression. These findings further underscore the importance of exploring inflammatory and metabolic biomarkers in PSD prediction. Future research should focus on identifying novel biomarkers that can serve as tools for early screening and prediction of PSD, which may improve the efficiency and accuracy of clinical interventions.
In the stratified meta-analysis, the biomarker-based data source group in retrospective studies demonstrated a significantly higher predictive performance, with a common-effect AUC of 0.94 (95% CI: 0.92–0.96), indicating a slight but notable statistical advantage over other data source groups (Figure 8). Regarding validation methods, all machine learning studies employed internal cross-validation, whereas none of the regression-based studies used cross-validation. Only one study applied nested cross-validation (9), however, its AUC was relatively low.
Limitations
This study has several limitations. The number of studies evaluating the use of machine learning for PSD prediction is relatively small, which may limit the generalizability and robustness of the findings. Additionally, differences in participant characteristics, such as age, comorbidities, and stroke severity, may have contributed to the observed heterogeneity in results. Variations in data quality, measurement techniques, and the timing of assessments could also influence the predictive accuracy of the models. Lastly, the use of diverse machine learning algorithms, with varying degrees of calibration and validation, might further explain the inconsistency in model performance across studies.
The predictive performance of the Support Vector Machine (SVM) was relatively low, indicating that it may not be the best choice for PSD prediction in the current dataset and model setup. However, SVM could demonstrate different performance in other studies or datasets, and future research should explore its applicability in various conditions. Similarly, the predictive performance of the biomarker group was relatively low, suggesting limited predictive ability in this study. Nevertheless, certain inflammatory and metabolism-related biomarkers showed predictive potential, and further research should validate their application value in PSD prediction. The risk of bias in several included studies is unclear, which may weaken the strength of the evidence.
Conclusion
In conclusion, this meta-analysis highlights the promise of machine learning models, particularly neural networks, in predicting PSD, while also emphasizing the importance of functional, physical, and cognitive assessments as key data sources. The role of biomarkers, such as inflammatory and metabolic factors, also warrants further exploration as potential predictive tools. Moving forward, more robust and diverse studies are needed to refine machine learning models, identify novel biomarkers, and enhance the predictive accuracy and generalizability of PSD prediction tools in clinical settings. This is the first time that a meta-analysis has found varying results for PSD prediction depending on the models and data used, as previous studies have not conducted similar research.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding authors.
Author contributions
HW: Writing – original draft, Data curation, Formal Analysis. YX: Data curation, Formal Analysis, Software, Writing – original draft. XH: Data curation, Formal Analysis, Visualization, Writing – review & editing. CY: Methodology, Writing – original draft. NZ: Conceptualization, Writing – original draft, Writing – review & editing. LM: Conceptualization, Writing – original draft, Writing – review & editing.
Funding
The author(s) declared that financial support was not received for this work and/or its publication.
Conflict of interest
The authors declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2025.1629023/full#supplementary-material
Supplementary Table 1 | Details of the search strategy used for PubMed, Embase, Cochrane Library, and Web of Science. This table provides the list of keywords, MeSH terms, and Boolean operators applied for retrieving studies on PSD prediction models.
Supplementary Table 2 | Summary of the included studies. This table includes study characteristics such as design, data sources, prediction models, and performance metrics (e.g., AUC, sensitivity, specificity).
Supplementary Table 3 | Risk of bias assessment results for each included study based on the PROBAST tool. The table highlights individual ratings across four domains: participants, predictors, outcomes, and statistical analysis.
References
1. Guo J, Wang J, Sun W, and Liu X. The advances of post-stroke depression: 2021 update. J Neurol. (2022) 2022:1–14.
2. Jørgensen TS, Wium-Andersen IK, Wium-Andersen MK, Jørgensen MB, Prescott E, Maartensson S, et al. Incidence of depression after stroke, and associated risk factors and mortality outcomes, in a large cohort of Danish patients. JAMA Psychiatry. (2016) 73:1032–40. doi: 10.1001/jamapsychiatry.2016.1932
3. Cai W, Mueller C, Li Y-J, Shen W-D, and Stewart R. Post stroke depression and risk of stroke recurrence and mortality: a systematic review and meta-analysis. Ageing Res Rev. (2019) 50:102–9. doi: 10.1016/j.arr.2019.01.013
4. Shi Y, Xiang Y, Yang Y, Zhang N, Wang S, Ungvari GS, et al. Depression after minor stroke: prevalence and predictors. J Psychosomatic Res. (2015) 79:143–7. doi: 10.1016/j.jpsychores.2015.03.012
5. He L, Wang J, Zhang L, Wang F, Dong W, and Yang H. Admission heart rate variability is associated with poststroke depression in patients with acute mild-moderate ischemic stroke. Front Psychiatry. (2020) 11:696. doi: 10.3389/fpsyt.2020.00696
6. He J-R, Zhang Y, Lu W-J, Liang H-B, Tu X-Q, Ma F-Y, et al. Age-related frontal periventricular white matter hyperintensities and miR-92a-3p are associated with early-onset post-stroke depression. Front Aging Neurosci. (2017) 9:328. doi: 10.3389/fnagi.2017.00328
7. Geng L-Y, Qian F-Y, Qian J-F, and Zhang Z-J. The combination of plasma glutamate and physical impairment after acute stroke as a potential indicator for the early-onset post-stroke depression. J Psychosomatic Res. (2017) 96:35–41. doi: 10.1016/j.jpsychores.2017.01.006
8. Wang W, Kiik M, Peek N, Curcin V, Marshall IJ, Rudd AG, et al. A systematic review of machine learning models for predicting outcomes of stroke with structured data. PLoS One. (2020) 15:e0234722.
9. Fast L, Temuulen U, Villringer K, Kufner A, Ali HF, Siebert E, et al. Machine learning-based prediction of clinical outcomes after first-ever ischemic stroke. Front Neurol. (2023) 14:1114360. doi: 10.3389/fneur.2023.1114360
10. Monteiro M, Fonseca AC, Freitas AT, e Melo TP, Francisco AP, Ferro JM, et al. Using machine learning to improve the prediction of functional outcome in ischemic stroke patients. IEEE/ACM Trans Comput Biol Bioinf. (2018) 15:1953–9. doi: 10.1109/TCBB.2018.2811471
11. Heo J, Yoon JG, Park H, Kim YD, Nam HS, and Heo JH. Machine learning–based model for prediction of outcomes in acute stroke. Stroke. (2019) 50:1263–5. doi: 10.1161/STROKEAHA.118.024293
12. Ladwig S, Ziegler M, Sudmeyer M, and Werheid K. The post-stroke depression risk scale (PoStDeRiS): development of an acute-phase prediction model for depression 6 months after stroke. J Acad Consult Liaison Psychiatry. (2022) 63:144–52. doi: 10.1016/j.jaclp.2021.08.003
13. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. (2021) 372:n71.
14. Moons KGM, Damen JAA, Kaul T, Hooft L, Andaur Navarro C, Dhiman P, et al. PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ. (2025) 388:e082505. doi: 10.1136/bmj-2024-082505
15. Gong J, Zhang Y, Zhong X, Zhang Y, Chen Y, and Wang H. Liver function test indices-based prediction model for post-stroke depression: a multicenter, retrospective study. BMC Med Inf Decision Making. (2023) 23:127. doi: 10.1186/s12911-023-02241-0
16. Ryu YH, Kim SY, Kim TU, Lee SJ, Park SJ, Jung HY, et al. Prediction of poststroke depression based on the outcomes of machine learning algorithms. J Clin Med. (2022) 11. doi: 10.3390/jcm11082264
17. Oei CW, Ng EYK, Ng MHS, Tan RS, Chan YM, Chan LG, et al. Explainable risk prediction of post-stroke adverse mental outcomes using machine learning techniques in a population of 1780 patients. Sensors (Basel Switzerland). (2023) 23. doi: 10.3390/s23187946
18. Song SI, Hong HT, Lee C, and Lee SB. A machine learning approach for predicting suicidal ideation in post stroke patients. Sci Rep. (2022) 12:15906. doi: 10.1038/s41598-022-19828-8
19. Chen YM, Chen PC, Lin WC, Hung KC, Chen YB, Hung CF, et al. Predicting new-onset post-stroke depression from real-world data using machine learning algorithm. Front Psychiatry. (2023) 14:1195586. doi: 10.3389/fpsyt.2023.1195586
20. Zhang X, Wang X, Wang S, Zhang Y, Wang Z, Yang Q, et al. Zheng Y et al: Machine learning algorithms assisted identification of post-stroke depression associated biological features. Front Neurosci. (2023) 17:1146620. doi: 10.3389/fnins.2023.1146620
21. Qiu X, Miao J, Lan Y, Sun W, Li G, Pan C, et al. Artificial neural network and decision tree models of post-stroke depression at 3 months after stroke in patients with BMI >/= 24. J Psychosom Res. (2021) 150:110632. doi: 10.1016/j.jpsychores.2021.110632
22. Hama S, Yoshimura K, Yanagawa A, Shimonaga K, Furui A, Soh Z, et al. Relationships between motor and cognitive functions and subsequent post-stroke mood disorders revealed by machine learning analysis. Sci Rep. (2020) 10:19571. doi: 10.1038/s41598-020-76429-z
23. de Man-van Ginkel JM, Hafsteinsdottir TB, Lindeman E, Ettema RG, Grobbee DE, and Schuurmans MJ. In-hospital risk prediction for post-stroke depression: development and validation of the Post-stroke Depression Prediction Scale. Stroke. (2013) 44:2441–5. doi: 10.1161/STROKEAHA.111.000304
24. Yue Y, Jiang H, Liu R, Yin Y, Zhang Y, Liang J, et al. Geng D et al: Towards a multi protein and mRNA expression of biological predictive and distinguish model for post stroke depression. Oncotarget. (2016) 7:54329–38. doi: 10.18632/oncotarget.11105
25. Cheng L-S, Tu W-J, Shen Y, Zhang L-J, and Ji K. Combination of high-sensitivity C-reactive protein and homocysteine predicts the post-stroke depression in patients with ischemic stroke. Mol Neurobiol. (2018) 55:2952–8. doi: 10.1007/s12035-017-0549-8
26. Liegey JS, Sagnier S, Debruxelles S, Poli M, Olindo S, Renou P, et al. Influence of inflammatory status in the acute phase of stroke on post-stroke depression. Rev Neurol (Paris). (2021) 177:941–6. doi: 10.1016/j.neurol.2020.11.005
27. Li G, Jing P, Chen G, Mei J, Miao J, Sun W, et al. Development and Validation of 3-Month Major Post-Stroke Depression Prediction Nomogram After Acute Ischemic Stroke Onset. Clin Interv Aging. (2021) 16:1439–47. doi: 10.2147/CIA.S318857
28. Piechota M, Hoinkis D, Korostynski M, Golda S, Pera J, and Dziedzic T. Gene expression profiling in whole blood stimulated ex vivo with lipopolysaccharide as a tool to predict post-stroke depressive symptoms: Proof-of-concept study. J Neurochem. (2023) 166:623–32. doi: 10.1111/jnc.15902
29. Luo S, Zhang W, Mao R, Huang X, Liu F, Liao Q, et al. Establishment and verification of a nomogram model for predicting the risk of post-stroke depression. PeerJ. (2023) 11:e14822. doi: 10.7717/peerj.14822
30. Luo S, Liu F, Liao Q, Chen H, Zhang T, and Mao R. Nomogram model for predicting the risk of post-stroke depression based on clinical characteristics and DNA methylation. PeerJ. (2023) 11:e16240. doi: 10.7717/peerj.16240
31. Qiu HC, Liu HZ, Li X, Zeng X, and Zhao JZ. Insulin resistance as estimated by homeostasis model assessment predicts incident post-stroke depression in Chinese subjects from ischemic stroke. J Affect Disord. (2018) 231:1–7. doi: 10.1016/j.jad.2018.01.023
32. Li Z, Wu WX, Ji ZX, Zhao KP, Zhou RJ, Jiang S, et al. The factors influencing the occurrence of post-ischemic stroke depression and its relationship with the burden score of cerebral small vessel disease. Eur Rev Med Pharmacol Sci. (2024) 28:2677–85.
33. Tu WJ, Qiu HC, Liu Q, Li X, Zhao JZ, and Zeng X. Decreased level of irisin, a skeletal muscle cell-derived myokine, is associated with post-stroke depression in the ischemic stroke population. J Neuroinflamm. (2018) 15:133. doi: 10.1186/s12974-018-1177-6
Keywords: post-stroke depression, machine learning models, PROBAST+AI, artificial intelligence, depression prediction, meta-analysis, systematic review
Citation: Wu H, Xiao Y, Hu X, You C, Zheng N and Ma L (2025) Clinical prediction models for post-stroke depression: a systematic review and meta-analysis. Front. Psychiatry 16:1629023. doi: 10.3389/fpsyt.2025.1629023
Received: 22 May 2025; Accepted: 01 December 2025; Revised: 16 November 2025;
Published: 17 December 2025.
Edited by:
Massimiliano Toscano, Sapienza University of Rome, ItalyReviewed by:
Tommaso B. Jannini, University of Rome Tor Vergata, ItalyVenkateswarlu Nalluri, Chaoyang University of Technology, Taiwan
Copyright © 2025 Wu, Xiao, Hu, You, Zheng and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chao You, eW91Y2hhb0B2aXAuMTI2LmNvbQ==; Niandong Zheng, bmV1cm9zdXJnZXJ5bnhnQHZpcC4xMjYuY29t
†These authors have contributed equally to this work
Honggang Wu1†