Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 22 August 2025

Sec. Medicine and Public Health

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1618492

Enhancing one-year mortality prediction in STEMI patients post-PCI: an interpretable machine learning model with risk stratification

Wenqiang Li,Wenqiang Li1,2Dongdong Yan,,,Dongdong Yan1,3,4,5Wei Hu,,,Wei Hu1,3,4,5Xiaoling SuXiaoling Su2Zheng Zhang,,,
Zheng Zhang1,3,4,5*
  • 1The First Clinical Medical School, Lanzhou University, Lanzhou, China
  • 2Department of Cardiology, Qinghai Provincial People’s Hospital, Xining, China
  • 3Heart Center, The First Hospital of Lanzhou University, Lanzhou, China
  • 4Key Laboratory for Cardiovascular Diseases of Gansu Province, Lanzhou, China
  • 5Cardiovascular Clinical Research Center of Gansu Province, Lanzhou, China

Background: ST-elevation myocardial infarction (STEMI) poses a significant threat to global mortality and disability. Advances in percutaneous coronary intervention (PCI) have reduced in-hospital mortality, highlighting the importance of post-discharge management. Machine learning (ML) models have shown promise in predicting adverse clinical outcomes. However, a systematic approach that combines high predictive accuracy with model simplicity is still lacking.

Methods: This retrospective study applied three data processing and ML algorithms to address class imbalance and support model development. ML models were trained to predict one-year mortality in STEMI patients post-PCI, with performance evaluated using accuracy, sensitivity, precision, F1-score, area under the receiver operating characteristic curve (AUROC), and the area under the precision-recall curve (AUPRC).

Results: We analyzed data from 1,274 patients, incorporating 46 clinical and laboratory features. Using the Random Forest (RF) algorithm, we achieved an AUROC of 0.94 (95% confidence interval (CI): 0.90–0.98), an AUPRC of 0.44 (95% CI:0.15–0.76) in the internal validation set, identifying five key predictors: cardiogenic shock, creatinine, NT-proBNP, diastolic blood pressure, and left ventricular ejection fraction. By integrating risk stratification, the model’s performance improved, achieving an AUROC of 0.97 (95% CI: 0.96–0.99) and an AUPRC of 0.74 (95% CI: 0.60–0.84).

Conclusion: This study highlights the feasibility of constructing accurate and interpretable ML models using a minimal set of predictors, supplemented by risk stratification, to improve long-term outcome prediction in STEMI patients.

Introduction

Acute myocardial infarction (AMI) remains a leading cause of morbidity and mortality worldwide, with ST-elevation myocardial infarction (STEMI) representing the most severe form, accounting for 50–60% of AMI cases in contemporary registries (Gaudino et al., 2023; Miller, 2020; Cederström et al., 2024). Advances in emergency care systems, the establishment of chest pain centers, and the widespread implementation of percutaneous coronary intervention (PCI) have significantly reduced in-hospital mortality, now reported to be between 4 and 8% (Khera et al., 2021; Liu et al., 2021). Consequently, optimizing post-discharge management for STEMI survivors has become increasingly important, with a strong emphasis on the early identification of high-risk individuals to enhance long-term outcomes.

Machine learning (ML) has shown substantial promise in risk stratification and outcome prediction across various clinical domains (Wang et al., 2021; Oliveira et al., 2023; Khera et al., 2024). Compared to traditional statistical approaches, ML models offer enhanced predictive accuracy and individualized risk assessment (Mohd Faizal et al., 2021; Stephan et al., 2025). However, despite the increasing availability of predictive models, standardized approaches for selecting the optimal model remain limited (Ogunpola et al., 2024; Radwa et al., 2024). The area under the receiver operating characteristic curve (AUROC) is commonly used to evaluate model performance (Lee et al., 2020; Payrovnaziri et al., 2019; Fukumoto et al., 2021). However, when multiple models demonstrate similarly high AUROC values, it becomes challenging to determine a clear winner. Furthermore, as the complexity of medical data continues to grow—along with the number of candidate models and input variables—developing models that balance predictive accuracy with clinical simplicity has become a key challenge.

This study aimed to construct a predictive model tailored to the characteristics of the dataset by selecting appropriate data processing methods and ML algorithms. Our objective was to maximize predictive performance while minimizing model complexity. The key contributions of this work include: (1) demonstrating that different data processing strategies yield no statistically significant differences in model performance; (2) showing that in highly imbalanced datasets where AUROC lacks discriminative capacity, the area under the precision-recall curve (AUPRC) provides a more informative metric (Zhou et al., 2021; Zheng et al., 2024); (3) proposing bootstrap testing as a viable alternative to DeLong’s test for comparing area under curve values; and (4) constructing a high-performing predictive model using a minimal set of readily available clinical variables, with further gains achieved through risk stratification.

Additionally, this study aimed to develop predictive models for one-year mortality in patients with STEMI post-PCI, utilizing the Shapley Additive Explanations (SHAP) method for model interpretability. We also developed a web-based application that enables clinicians to predict individual patient outcomes by inputting the required model variables, thereby facilitating personalized risk assessment.

Materials and methods

Study population

This study involved a retrospective analysis of patients admitted to the First Hospital of Lanzhou University in Gansu Province, China, between January 1, 2019, and December 31, 2020. All consecutive hospitalized patients diagnosed with STEMI and treated with PCI during this period were screened for eligibility. STEMI was diagnosed based on criteria established by the European Society of Cardiology (ESC) Association (Huang et al., 2025; Razavi et al., 2025). Inclusion criteria were: (1) a confirmed diagnosis of STEMI, (2) age ≥18 years, (3) receipt of PCI, and (4) availability of complete clinical data. Details of the inclusion and exclusion process are provided in Supplementary Figure S1.

Follow-up information was obtained through structured telephone interviews conducted by a dedicated follow-up center, supplemented by outpatient clinical assessments.

Data collection

Clinical and laboratory data for patients with STEMI were extracted from electronic medical records, with one-year all-cause mortality defined as the primary outcome. Detailed clinical and laboratory variables are presented in Table 1. For each patient, the initial set of laboratory test results obtained upon hospital admission—including blood samples—was utilized for analysis. A two-step approach was employed to address missing data: variables with more than 20% missing values were excluded, while those with less than 20% missing values were imputed using the IterativeImputer method to minimize bias.

Table 1
www.frontiersin.org

Table 1. Comparison of baseline characteristics of the study population.

A total of 46 variables were included in the final dataset, comprising demographic information, cardiovascular history, and laboratory measurements. These features included: gender, age, total ischemic time (TIT), coronary artery disease (CAD), hypertension (HTN), type 2 diabetes mellitus (T2DM), hyperlipidemia (HLD), peripheral arterial disease (PAD), smoking history (SH), bleeding history (BH), body mass index (BMI), systolic and diastolic blood pressure (SBP, DBP), heart rate (HR), creatinine (CREA), uric acid (UA), random blood glucose (RBG), low-density lipoprotein cholesterol (LDL-C), estimated glomerular filtration rate (eGFR), hematocrit (HCT), neutrophil count (NEUT), lymphocyte count (LYMPH), neutrophil-to-lymphocyte ratio (NLR), hemoglobin (HGB), platelet count (PLT), C-reactive protein (CRP), myoglobin (MYO), creatine kinase-MB (CK-MB), troponin I (TNI), N-terminal pro B-type natriuretic peptide (NT-proBNP), hemoglobin A1c (HbA1c), left atrial diameter (LAD), left ventricular ejection fraction (LVEF), left ventricular end-diastolic and end-systolic volumes (LVEDV, LVESV), and medications including beta-blockers (BB), angiotensin-converting enzyme inhibitors (ACEI), angiotensin II receptor blockers (ARB), angiotensin receptor-neprilysin inhibitors (ARNI), sodium-glucose cotransporter-2 inhibitors (SGLT2i), statins (STAT), aspirin (ASA), ticagrelor (TICA), clopidogrel (CLOP), and proton pump inhibitors (PPI). Key clinical events such as ventricular fibrillation (VF), atrial fibrillation (AF), and cardiogenic shock (CS) were also documented.

Model development and explanation

Patients were randomly divided into training and testing cohorts in a 7:3 ratio. Prior to model construction, several data processing steps were implemented: the synthetic minority oversampling technique (SMOTE) was applied to address class imbalance (Shi and Fan, 2023), Boruta was used for feature selection (Kursa and Rudnicki, 2010), and grid search with cross-validation (GSCV) was employed for hyperparameter tuning. Six ML algorithms were evaluated: RF, light gradient boosting machine (LightGBM), extreme gradient boosting (XGBoost), logistic regression (LR), k-nearest neighbors (KNN), and deep neural networks (DNN).

Model development adhered to a structured three-step framework: (1) identifying the optimal data processing pipeline, (2) selecting the best-performing ML algorithm, and (3) constructing the final predictive model by integrating the chosen processing and algorithmic strategies (Figure 1). Model performance was primarily assessed using AUROC and AUPRC (Zhou et al., 2021). Additional evaluation metrics—accuracy, precision, sensitivity, specificity, and F1-score—were employed to support model selection.

Figure 1
Flowchart illustrating the process of model development and explanation. It includes two main sections:

Figure 1. The ML model development process. ML, machine learning; SMOTE, synthetic minority over-sampling technique; GSCV, grid search with cross-validation; SHAP, shapley additive explanation; LIME, Local Interpretable Model-agnostic Explanations.

To enhance interpretability, SHAP and local interpretable model-agnostic explanations (LIME) analyses were employed to evaluate feature importance and quantify the contribution of individual variables to model predictions (Qi et al., 2025). SHAP force plots and LIME were utilized to visualize the influence of each feature on individual predictions, thereby improving transparency and supporting personalized clinical interpretation.

Classes of risk

The pooled dataset, which includes 1,274 patients from both the derivation and external validation datasets, was categorized into low- and high-risk levels based on clinically meaningful thresholds. The stratification threshold was determined by analyzing the predicted probability distribution and calibration curves of the model in both the training and validation datasets. The model assesses individualized risk probabilities for specific patients based on input parameters and subsequently stratifies them into low-risk or high-risk groups using predefined thresholds. This risk stratification directly informs clinical decision-making regarding treatment intensity and follow-up frequency, thereby serving as a reference for personalized therapeutic strategies.

Statistical analysis

All statistical analyses were conducted using R Software (version 4.3.1; http://www.r-project.org). The development of predictive models was carried out with Python (version 3.11.9; https://www.python.org) and PyCharm (version 2024.1.4). For normally distributed continuous variables, data are presented as means with standard deviations, while comparisons between groups were performed using independent sample t-tests. For variables with abnormal distributions, data are expressed as median values (Q1, Q3), and the Mann–Whitney U test was utilized for comparisons between two groups. Differences in AUROCs between models were assessed using DeLong’s test, and differences in AUPRCs were evaluated using the bootstrap method. Categorical variables are reported as counts (percentages) and were compared using the Chi-squared test. If the expected frequency of any cell was less than 5, Fisher’s exact test was employed. A two-tailed p-value of less than 0.05 was considered statistically significant. The Pearson correlation matrix was used to quantify the linear relationships among features.

Results

Population characteristics

A total of 672 patients were included in the derivation cohort for developing the one-year mortality prediction model, among whom 38 (5.7%) died within one-year post-PCI. Regarding missing data, BMI and HR were absent in 1 case (0.1%), HbA1c in 69 cases (10.3%), LAD in 8 cases (1.2%), LVEF in 6 cases (0.9%), LVEDV and LVESV in 7 cases (1.0%). The external validation cohort comprised 602 patients, with 21 (3.5%) experiencing one-year mortality. A comparative analysis of demographic and clinical characteristics between survivors and non-survivors is presented in Table 1, while Supplementary Table S1 provides a detailed comparison between the training and testing cohorts. Our analysis of baseline characteristics revealed statistically significant differences between the modeling cohort and the external validation cohort across multiple clinical variables, including medical history, vital signs, laboratory values, medication use patterns, and clinical outcomes (p < 0.05). However, within the modeling cohort, the training and validation subsets exhibited well-balanced characteristics without significant differences, thereby supporting the robustness of our internal validation approach.

Model selection and performance comparison

To optimize predictive performance, data processing methods—including SMOTE, Boruta, and GSCV—were applied where appropriate. The effects of these processing strategies were evaluated using the AUROC and AUPRC metrics, as illustrated in Figures 2A,B, respectively.

Figure 2
Graph A shows the Receiver Operating Characteristic curves comparing models: RawData (AUROC = 0.88), GSCV (0.91), SMOTE+GSCV (0.88), Boruta+GSCV (0.95), SMOTE+Boruta+GSCV (0.88). Graph B illustrates the Precision-Recall curves for the same models with varying AUPRC values around 0.48 to 0.52. Both graphs include confidence intervals.

Figure 2. Model performance under five data preprocessing strategies. (A) AUROC values across different preprocessing levels. (B) AUPRC values across different preprocessing levels. SMOTE, synthetic minority over-sampling technique; GSCV, grid search with cross-validation; AUROC, area under the receiver operating characteristic curve; AUPRC, area under the precision-recall curve.

Although processing steps influenced model performance, the differences in AUROC were not statistically significant based on the DeLong test (Supplementary Figure S2A), nor were AUPRC differences significant according to bootstrap testing (Supplementary Figure S2B).

Key performance metrics—AUROC, AUPRC, accuracy, precision, sensitivity, specificity, and F1 score—were utilized to assess model efficacy. As shown in Supplementary Table S2, the model developed using Boruta and GSCV achieved an AUROC of 0.949, AUPRC of 0.486, accuracy of 0.950, precision of 0.500, sensitivity of 0.400, specificity of 0.979, and F1 score of 0.444.

Six ML algorithms were evaluated, with the results for AUROC and AUPRC presented in Figures 3A,B. Among these algorithms, RF, LightGBM, and XGBoost exhibited the best predictive performance. The RF model achieving an AUROC of 0.91 (95% confidence interval (CI): 0.81–0.98) and an AUPRC of 0.53 (95% CI: 0.17–0.82).

Figure 3
Two panels depicting model evaluation metrics. Panel A shows a Receiver Operating Characteristic curve comparing models: RF, LightGBM, XGBoost, LR, KNN, and DNN with AUROC scores ranging from 0.74 to 0.91. Panel B presents Precision-Recall Curves with AUPRC scores, ranging from 0.35 to 0.53. Each model is represented by a different colored line.

Figure 3. Model performance across six ML algorithms. (A) AUROC values for models developed using six algorithms. (B) AUPRC values for models developed using six algorithms. ML, machine learning; RF, random forest; LightGBM, light gradient boosting machine; XGBoost, extreme gradient boosting; LR, logistic regression; KNN, k-nearest neighbor; DNN, deep neural network; AUROC, area under the receiver operating characteristic curve; AUPRC, area under the precision-recall curve.

Upon reevaluation of the impact of processing methods, it was determined that the differences in AUROC among the models were not statistically significant, as assessed by the DeLong test (Supplementary Figure S3A). However, the RF model displayed a significantly higher AUPRC compared to the LR model, while the differences in AUPRC among the other models remained non-significant based on bootstrap testing (Supplementary Figure S3B).

As summarized in Supplementary Table S3, the RF model exhibited superior overall performance, achieving an AUROC of 0.911, AUPRC of 0.534, accuracy of 0.960, precision of 0.667, sensitivity of 0.400, specificity of 0.990, and an F1 score of 0.500.

Final model identification

No statistically significant differences in model performance (AUROC/AUPRC) were observed across various data processing strategies. Consequently, the method that offered the best balance between predictive precision and simplicity was selected. Among the evaluated ML algorithms, the RF model—optimized using Boruta and GSCV—achieved the highest AUROC and AUPRC values, thereby being designed as the optimal predictive model. Internal validation demonstrated an AUROC of 0.94 (95% CI: 0.90–0.98) and an AUPRC of 0.44 (95% CI: 0.15–0.76), while external validation yielded an AUROC of 0.93 (95% CI: 0.86–0.99) and an AUPRC of 0.70 (95% CI: 0.51–0.88) (Supplementary Figure S4).

To enhance clinical utility, patients were stratified into high- and low-risk groups using a probability threshold of 0.6. Calibration analysis (Supplementary Figure S5) demonstrated a systematic overestimation of high-risk predictions, while showing excellent agreement within the low-to-medium risk ranges, thereby supporting clinical reliability. Probability distributions by outcome (Supplementary Figure S6) exhibited a clear separation between survivors and non-survivors. Kernel density plots indicated significantly higher predicted probabilities for deceased patients (Mann–Whitney U test, p < 0.001), with the current threshold (p = 0.6, represented by the black dashed line) and the suggested optimal range (indicated by green shading). Boxplots further illustrated these distinct distributions.

The post-stratification model performance is illustrated in Figure 4. Internal validation yielded an AUROC of 0.97 (95% CI: 0.96–0.99) and AUPRC of 0.74 (95% CI: 0.60–0.84), while external validation showed an AUROC of 0.93 (95% CI: 0.85–0.99) and AUPRC of 0.70 (95% CI: 0.46–0.88).

Figure 4
Panel A shows a Receiver Operating Characteristic curve comparing internal and external validation, with AUROC values of 0.97 and 0.93, respectively. Panel B presents Precision-Recall curves for the same validations, showing AUPRC values of 0.74 for internal and 0.70 for external validation.

Figure 4. Performance of the stratified RF model in predicting one-year mortality among STEMI patients post-PCI. (A) AUROC values of stratified-RF model. (B) AUPRC values of the stratified-RF model. STEMI, ST-segment elevation myocardial infarction; RF, random forest; PCI, percutaneous coronary intervention; AUROC, area under the receiver operating characteristic curve; AUPRC, area under the precision-recall curve.

Performance metrics before and after risk stratification are summarized in Table 2. After stratification, the final model achieved an accuracy of 0.985, a precision of 0.875, a sensitivity of 0.667, a specificity of 0.997, and an F1 score of 0.757 in the external validation set.

Table 2
www.frontiersin.org

Table 2. Performance of the final model after applying the risk stratification method.

Model development

The application of the Boruta algorithm for feature selection significantly enhanced model interpretability and mitigated overfitting by reducing the input set from 46 features to just 5. This reduction not only preserved the model’s performance but also simplified its implementation. Figure 5 illustrates the features selected by the Boruta algorithm. Furthermore, the optimized RF model was configured with the following hyperparameters: a maximum tree depth of 30, a minimum of samples per split, a total of 300 decision trees.

Figure 5
Box plot showing Boruta feature importances for various factors such as age, gender, and medical conditions. Features are categorized as rejected (light red) and confirmed (dark red). Confirmed features with higher importance include DBP and NTproBNP.

Figure 5. Features chosen in the training cohort using the Boruta algorithm. TIT, total ischemic time; CAD, coronary artery disease; HTN, hypertension; T2DM, type 2 diabetes mellitus; HLD, hyperlipidemia; PAD, peripheral arterial disease; SH, smoking history; BH, bleeding history; BMI, body mass index; SBP, systolic blood pressure; DBP, diastolic blood pressure; HR, heart rate; CREA, serum creatinine; UA, uric acid; RBG, random blood glucose; LDL-C, low-density lipoprotein cholesterol; eGFR, estimated glomerular filtration rate; HCT, hematocrit; NEUT, neutrophil count; LYMPH, lymphocyte count; NLR, neutrophil-to-lymphocyte ratio; HGB, hemoglobin; PLT, platelet count; CRP, c-reactive protein; MYO, myoglobin; CK-MB, creatine kinase-mb; TNI, troponin I; NT-proBNP, n-terminal pro b-type natriuretic peptide; HbA1c, hemoglobin a1c; LAD, left atrial diameter; LVEF, left ventricular ejection fraction; LVEDV, left ventricular end-diastolic volume; LVESV, left ventricular end-systolic volume; BB, beta-blockers; ACEI, angiotensin-converting enzyme inhibitors; ARB, angiotensin II receptor blockers; ARNI, angiotensin receptor-neprilysin inhibitors; SGLT2i, sodium-glucose cotransporter-2 inhibitors; STAT, statins; ASA, aspirin; TICA, ticagrelor; CLOP, clopidogrel; PPI, proton pump inhibitors; VF, ventricular fibrillation; AF, atrial fibrillation; CS, cardiogenic shock.

Model explanation

We employed the Pearson correlation matrix to assess multicollinearity among all modeling features (Supplementary Figure S7). The results indicated four pairs of features exhibiting high collinearity (|r| > 0.7): SBP and DBP (r = 0.75), HCT and HGB (r = 0.95), CKMB and TNI (r = 0.83), and LVEDV and LESV (r = 0.91). These strong correlations suggest potential redundancy, emphasizing the need for dimensionality reduction or variable selection to mitigate multicollinearity during model development. The final model did not include the afore mentioned features simultaneously. Feature importance in the final RF model is illustrated in a radar plot (Figure 6A) and a SHAP summary bar plot (Figure 6B), highlighting the five variables that influence one-year mortality. The model comprises five features, ranked by weight from highest to lowest: CS (0.433), CREA (0.200), DBP (0.191), NT-proBNP (0.104), and LVEF (0.072).

Figure 6
Radar chart (A) shows variables: CREA, CS, LVEF, DBP, NTproBNP with values ranging up to 0.2. Bar chart (B) depicts feature importance, with CS highest, followed by CREA, DBP, NTproBNP, and LVEF.

Figure 6. Feature importance in the stratified RF model for predicting one-year mortality in STEMI patients post-PCI. (A) Radar plot depicting feature importance for the stratified RF model. (B) SHAP summary bar plot illustrating the contributions of individual features to the model’s predictions. STEMI, ST-segment elevation myocardial infarction; RF, random forest; PCI, percutaneous coronary intervention; CS, cardiogenic shock; CREA, creatinine; NT-proBNP, N-terminal pro-B-type natriuretic peptide; DBP, diastolic blood pressure; LVEF, left ventricular ejection fraction.

To enhance transparency and clinical interpretability, SHAP was used to analyze overall feature importance, while LIME was employed to examine local feature weights and decision rules for individual samples. Figure 7 shows SHAP summary plots ranking feature importance based on mean SHAP values. The most influential predictors included CS, CREA, NT-proBNP, DBP, and LVEF. Figure 8 illustrates the contribution of key clinical features to individual risk prediction across various patient subgroups. Notably, across all subgroups, the absence of cardiogenic shock (CS ≤ 0) consistently emerged as the strongest negative predictor. Elevated levels of NTproBNP (>1677.5 pg./mL) and impaired renal function (CREA >86 μmol/L) were identified as significant risk enhancers. Additionally, subtle variations in diastolic blood pressure (DBP) and left ventricular ejection fraction (LVEF) contributed to risk modulation specific to each subgroup.

Figure 7
Dot plots labeled A and B show feature values for variables: CS, CREA, NTproBNP, DBP, and LVEF. Plot A ranges from -0.6 to 0.0, with high values in red and low in blue, emphasizing neutrality. Plot B ranges from 0.0 to 0.6, indicating similar value distinctions with reversed feature emphasis. Both plots compare the impact of variables on a specific outcome.

Figure 7. SHAP summary dot plot for the stratified RF model predicting one-year mortality in STEMI patients post-PCI. (A) SHAP summary plot for the model predicting death as a positive outcome. (B) SHAP summary plot for the model predicting survival as a positive outcome. The plot illustrates the contribution of each feature to the model’s predictions. STEMI, ST-segment elevation myocardial infarction; RF, random forest; PCI, percutaneous coronary intervention; CS, cardiogenic shock; CREA, creatinine; NT-proBNP, N-terminal pro–B-type natriuretic peptide; DBP, diastolic blood pressure; LVEF, left ventricular ejection fraction.

Figure 8
Four bar plots labeled A, B, C, and D display variable importance. Red bars indicate negative importance and green bars indicate positive importance. The variables are CS, NTproBNP, CREA, DBP, and LVEF, with distinct value ranges for each plot.

Figure 8. Contribution of key clinical features to individual risk prediction across subgroups. (A–D) Illustrate the directional impact of selected discretized features on model output across different patient strata. The horizontal bars represent both the magnitude and direction of contribution for each feature, with red indicating a negative (risk-reducing) effect and green indicating a positive (risk-enhancing) effect.

Clinical utility and application

To facilitate clinical application, the final model was deployed as a web-based tool (Figure 9). Clinicians can input patient values for the five selected features, and the tool will automatically calculate the predicted risk of one-year mortality for STEMI patients post-PCI. Additionally, it generates a personalized SHAP force plot that visually highlights the factors influencing each prediction. In these plots, the blue features on the right indicate variables associated with improved survival, while the red features on the left represent factors that contribute to an increased risk of mortality.

Figure 9
STEMI One-Year Mortality Predictor interface showing inputs for cardiogenic shock, creatinine levels, NT-proBNP levels, diastolic blood pressure, and LVEF percentage. Inputs include no cardiogenic shock, creatinine at 172 micromoles per liter, NT-proBNP at 400 picograms per milliliter, diastolic blood pressure at 70 millimeters of mercury, and LVEF at 35 percent. The prediction button indicates a low-risk stratification with a predicted probability of risk at 24 percent. A horizontal scale below visualizes risk factors.

Figure 9. Convenient application for clinical utility. The convenient application of the stratified RF model with 5 features is available for STEMI one-year mortality prediction. When entering actual values of the 5 features, this application automatically displays the probability of 24%. Meanwhile, the force plot for individual child indicates the features that contribute to the decision of mortality: the blue features on the right are the features pushing the prediction towards decease, while the red features on the left are pushing the prediction towards survival. RF, random forest; STEMI, st-segment elevation myocardial infarction.

Discussion

This study successfully developed a predictive model for estimating the 1-year mortality risk in STEMI patients following PCI and validated a systematic approach to model construction. Based on the characteristics of the raw data, appropriate processing strategies and ML algorithms were selected. Model performance was evaluated using AUROC and AUPRC metrics. The final streamlined model included only five clinical variables: CS, CREA, NT-proBNP, DBP, and LVEF. Among all candidate models, the one constructed using the Boruta, GSCV, and the RF algorithm demonstrated the best performance, with an AUROC of 0.94 and an AUPRC of 0.44 in the internal validation set. However, the application of risk stratification substantially improved model performance, achieving an AUROC of 0.97 and an AUPRC of 0.74. We propose categorizing patients into two risk classes (low and high) for each outcome. This stratification aims to underscore the clinical implications of each risk value computed by the model. By selecting a relatively high threshold, we corrected the overestimated event rates in the overall test set, thereby enhancing model performance. This approach ensured stability within specific populations and mitigated inflated metrics caused by risk overestimation. Following stratification, the sample characteristics within high and low-risk groups became more homogeneous, thereby improving the model’s predictive accuracy within each subgroup. These findings suggest that risk stratification is an effective method for enhancing model performance in imbalanced datasets.

Clinical databases often exhibit missing values, high dimensionality, and class imbalance, necessitating robust processing strategies (Mohammadi et al., 2023; Chan et al., 2023; Iacobescu et al., 2024; Hosseini et al., 2024; Öztekin and Özyılmaz, 2025). Prior studies have addressed these challenges using techniques such as SMOTE and GSCV (Oliveira et al., 2023). For instance, Iacobescu et al. (2024) utilized SMOTE and GSCV to enhance model performance in a highly imbalanced dataset (91.91% vs. 8.09%); however, they did not quantify the impact of processing. Hosseini et al. (2024) analyzed 9,073 AMI patients and reported an AUROC of 0.866 for the RF model following the application of SMOTE. Similarly, Öztekin and Özyılmaz (2025) employed SMOTE with 38 features to develop an in-hospital mortality model, achieving high predictive performance, despite a small sample size. Nonetheless, these studies did not evaluate the statistical significance of processing on model performance or delineate when SMOTE is necessary. In our study, we systematically assessed various processing strategies and confirmed, using DeLong’s test, that there were no significant differences in AUROC. Consequently, we selected the simplest effective approach to maximize model interpretability and clinical relevance.

Choosing the most suitable ML algorithm remains a significant challenge in predictive modeling (Yang et al., 2025; Liu et al., 2022; Jeong et al., 2024). Most studies recommend algorithm based on conventional metrics—such as AUROC, F1-score, precision, sensitivity, and accuracy—yet few assess the statistical significance between models (Zhang et al., 2023; Zheng et al., 2023). For instance, D'Ascenzo et al. (2021) developed the PRAISE score for one-year mortality in ACS patients using adaptive boosting (Adaboost), naive bayes, KNN, and RF, concluding that Adaboost performed the best. Khera et al. (2021) compared four models (XGBoost, neural networks, meta-classifier, and LR) for predicting in-hospital AMI mortality, finding no significant improvement of ML models over LR, with all AUPRC values below 0.4. Although these comparisons are informative, they lack statistical testing. Notably, Hu et al. (2024) employed the DeLong test to evaluate algorithmic differences, thereby adding methodological rigor. In our study, we compared RF, LightGBM, XGBoost, LR, KNN, and DNN. Ensemble methods like RF reduce variance through bagging, while LightGBM and XGBoost enhance accuracy via boosting and regularization. LR provides interpretability, KNN offers flexibility with structured data, and DNNs excel in handling complex, high-dimensional data. Our findings support the growing consensus that there is no universally best algorithm—only the most contextually appropriate one. Despite algorithmic advances, improvements are not always statistically significant, as confirmed by our results.

While increasing the number of input features may initially enhance model performance, the improvements often plateau (Chen et al., 2024; Hamilton et al., 2024; Li et al., 2024; Liu et al., 2024; Hu et al., 2024). Therefore, constructing models with a limited number of clinically accessible features becomes a primary objective. Some researchers employ LASSO or SHAP to rank feature importance and manually select the most significant variables. For example, Hu et al. (2024) successfully reduced 33 features to 8 using SHAP without a notable decline in performance. Similarly, Liu et al. (2024) developed a readmission model for NSTEMI patients, condensing 96 features to 7 through the use of LR, RF, and LASSO, ultimately finding the LR-based model to be the most effective. However, these studies generally did not evaluate the statistical significance of performance differences between models, which limits their robustness.

Recognizing the complexity of real-world clinical data, we emphasize that simplicity and usability are critical for the adoption of models. Consequently, we employed DeLong and bootstrap tests to compare the AUROC and AUPRC across various models, ultimately finding no statistically significant differences. Based on these results, we selected the simplest model capable of reliably identifying positive cases, which aligns with the priorities of clinical decision-making.

In addition to limitations related to sample size and data quality, we contend that the reliance on static variables to predict dynamic clinical outcomes is inherently restrictive. Future advancements in real-time, dynamic prediction models—incorporating time-series analysis and multimodal data integration—may offer a more precise and personalized approach to risk stratification and clinical decision-making (Ren et al., 2025).

Limitations

There are several limitations to our study. First, as a single-center, retrospective investigation with a limited sample size, it is essential to conduct multicenter, prospective studies for more robust findings. Second, although we performed external validation, all data were sourced from the same hospital, which may introduce selection bias and lacks validation across diverse institutional settings. Third, the clinical utility of this prognostic prediction model is constrained by its inability to incorporate time-series analysis and integrate multimodal data, which limits its applicability in real-world scenarios.

Conclusion and future work

In conclusion, we developed an RF-based predictive model utilizing Boruta for feature selection and GSCV for optimization, effectively reducing dimensionality and enhancing model performance in predicting one-year mortality in STEMI patients post-PCI. Additionally, integrating a risk stratification approach significantly improved the model’s clinical applicability.

This work highlights several gaps that must be addressed in future research before implementing this type of model construction. First, while SMOTE is an effective method for addressing class imbalance, particularly when the imbalance ratio exceeds 10:1, it may overlook intra-class distributions. Additionally, its interpolation of minority-class samples can lead to the generation of noisy instances, potentially increasing model complexity and degrading performance. SMOTE is most effective when the data is imbalanced and the minority class exhibits clear distribution patterns. Second, although SHAP provides valuable insights into feature contributions, its reliability may be compromised in the presence of strong feature multicollinearity. Future work should incorporate correlation analysis and feature selection or reduction prior to interpretation to ensure more coherent SHAP results.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by the Ethics Committee of Lanzhou University First Hospital, Lanzhou, Gansu, China. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

WL: Writing – original draft, Conceptualization, Methodology, Validation, Writing – review & editing, Software, Formal analysis. DY: Investigation, Resources, Data curation, Writing – review & editing, Methodology. WH: Formal analysis, Visualization, Data curation, Validation, Writing – review & editing. XS: Supervision, Methodology, Validation, Writing – review & editing. ZZ: Formal analysis, Visualization, Project administration, Conceptualization, Resources, Funding acquisition, Methodology, Data curation, Investigation, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by the National Key Research and Development Program of China (Lanzhou, China) grant no. 2,018YFC131,1505 and the Cardiovascular Clinical research center of Gansu Province (Lanzhou, China) grant no. 18JR2FA005.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Cederström, S., Jernberg, T., Samnegård, A., Johansson, F., Silveira, A., Tornvall, P., et al. (2024). Inflammatory biomarkers and long-term outcome in young patients three months after a first myocardial infarction. Cytokine 182:156696. doi: 10.1016/j.cyto.2024.156696

PubMed Abstract | Crossref Full Text | Google Scholar

Chan, P. Z., Ramli, M., and Chew, H. S. J. (2023). Diagnostic test accuracy of artificial intelligence-assisted detection of acute coronary syndrome: a systematic review and meta-analysis. Comput. Biol. Med. 167:107636. doi: 10.1016/j.compbiomed.2023.107636

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, A.-T., Zhang, Y., and Zhang, J. (2024). Explainable machine learning and online calculators to predict heart failure mortality in intensive care units. ESC Heart Fail. 12, 353–368. doi: 10.1002/ehf2.15062

PubMed Abstract | Crossref Full Text | Google Scholar

D'Ascenzo, F., De Filippo, O., Gallone, G., Mittone, G., Deriu, M. A., Iannaccone, M., et al. (2021). Machine learning-based prediction of adverse events following an acute coronary syndrome (PRAISE): a modelling study of pooled datasets. Lancet 397, 199–207. doi: 10.1016/S0140-6736(20)32519-8

PubMed Abstract | Crossref Full Text | Google Scholar

Fukumoto, Y., Aziz, F., Malek, S., Ibrahim, K. S., Raja Shariff, R. E., Wan Ahmad, W. A., et al. (2021). Short- and long-term mortality prediction after an acute ST-elevation myocardial infarction (STEMI) in Asians: a machine learning approach. PLoS One 16:e254894. doi: 10.1371/journal.pone.0254894

Crossref Full Text | Google Scholar

Gaudino, M., Andreotti, F., and Kimura, T. (2023). Current concepts in coronary artery revascularisation. Lancet 401, 1611–1628. doi: 10.1016/S0140-6736(23)00459-2

PubMed Abstract | Crossref Full Text | Google Scholar

Hamilton, D. E., Albright, J., Seth, M., Painter, I., Maynard, C., Hira, R. S., et al. (2024). Merging machine learning and patient preference: a novel tool for risk prediction of percutaneous coronary interventions. Eur. Heart J. 45, 601–609. doi: 10.1093/eurheartj/ehad836

PubMed Abstract | Crossref Full Text | Google Scholar

Hosseini, K., Behnoush, A. H., Khalaji, A., Etemadi, A., Soleimani, H., Pasebani, Y., et al. (2024). Machine learning prediction of one-year mortality after percutaneous coronary intervention in acute coronary syndrome patients. Int. J. Cardiol. 409:132191. doi: 10.1016/j.ijcard.2024.132191

PubMed Abstract | Crossref Full Text | Google Scholar

Hu, J., Xu, J., Li, M., Jiang, Z., Mao, J., Feng, L., et al. (2024). Identification and validation of an explainable prediction model of acute kidney injury with prognostic implications in critically ill children: a prospective multicenter cohort study. eClinicalMedicine 68:102409. doi: 10.1016/j.eclinm.2023.102409

PubMed Abstract | Crossref Full Text | Google Scholar

Huang, Q., Jiang, Z., Shi, B., Meng, J., Shu, L., Hu, F., et al. (2025). Characterisation of cardiovascular disease (CVD) incidence and machine learning risk prediction in middle-aged and elderly populations: data from the China health and retirement longitudinal study (CHARLS). BMC Public Health 25:518. doi: 10.1186/s12889-025-21609-7

PubMed Abstract | Crossref Full Text | Google Scholar

Iacobescu, P., Marina, V., Anghel, C., and Anghele, A.-D. (2024). Evaluating binary classifiers for cardiovascular disease prediction: enhancing early diagnostic capabilities. J. Cardiovasc. Dev. Dis. 11:396. doi: 10.3390/jcdd11120396

PubMed Abstract | Crossref Full Text | Google Scholar

Jeong, J. H., Lee, K. S., Park, S. M., Kim, S. R., Kim, M. N., Chae, S. C., et al. (2024). Prediction of longitudinal clinical outcomes after acute myocardial infarction using a dynamic machine learning algorithm. Front. Cardiovasc. Med. 11:1340022. doi: 10.3389/fcvm.2024.1340022

PubMed Abstract | Crossref Full Text | Google Scholar

Khera, R., Haimovich, J., Hurley, N. C., McNamara, R., Spertus, J. A., Desai, N., et al. (2021). Use of machine learning models to predict death after acute myocardial infarction. JAMA Cardiol. 6, 633–641. doi: 10.1001/jamacardio.2021.0122

PubMed Abstract | Crossref Full Text | Google Scholar

Khera, R., Oikonomou, E. K., Nadkarni, G. N., Morley, J. R., Wiens, J., Butte, A. J., et al. (2024). Transforming cardiovascular care with artificial intelligence: from discovery to practice. J. Am. Coll. Cardiol. 84, 97–114. doi: 10.1016/j.jacc.2024.05.003

PubMed Abstract | Crossref Full Text | Google Scholar

Kursa, M. B., and Rudnicki, W. R. (2010). Feature selection with the Boruta package. J. Stat. Softw. 36, 1–13. doi: 10.18637/jss.v036.i11

Crossref Full Text | Google Scholar

Lee, H. C., Park, J. S., Choe, J. C., Ahn, J. H., Lee, H. W., Oh, J.-H., et al. (2020). Prediction of 1-year mortality from acute myocardial infarction using machine learning. Am. J. Cardiol. 133, 23–31. doi: 10.1016/j.amjcard.2020.07.048

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Q., Lv, H., Chen, Y., Shen, J., shi, J., and Zhou, C. (2024). Development and validation of a machine learning predictive model for perioperative myocardial injury in cardiac surgery with cardiopulmonary bypass. J. Cardiothorac. Surg. 19:384. doi: 10.1186/s13019-024-02856-y

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, Y., Du, L., Li, L., Xiong, L., Luo, H., Kwaku, E., et al. (2024). Development and validation of a machine learning-based readmission risk prediction model for non-ST elevation myocardial infarction patients after percutaneous coronary intervention. Sci. Rep. 14:13393. doi: 10.1038/s41598-024-64048-x

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, W.-C., Lin, C.-S., Tsai, C.-S., Tsao, T.-P., Cheng-Chung, C., Liou, J.-T., et al. (2021). A deep learning algorithm for detecting acute myocardial infarction. EuroIntervention 17, 765–773. doi: 10.4244/EIJ-D-20-01155

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, L., Qiao, C., Zha, J. R., Qin, H., Wang, X. R., Zhang, X. Y., et al. (2022). Early prediction of clinical scores for left ventricular reverse remodeling using extreme gradient random forest, boosting, and logistic regression algorithm representations. Front. Cardiovasc. Med. 9:864312. doi: 10.3389/fcvm.2022.864312

PubMed Abstract | Crossref Full Text | Google Scholar

Miller, D. D. (2020). Machine intelligence in cardiovascular medicine. Cardiol. Rev. 28, 53–64. doi: 10.1097/CRD.0000000000000294

PubMed Abstract | Crossref Full Text | Google Scholar

Mohammadi, T., D'Ascenzo, F., Pepe, M., Zanghì, S. B., Bernardi, M., Spadafora, L., et al. (2023). Unsupervised machine learning with cluster analysis in patients discharged after an acute coronary syndrome: insights from a 23,270-patient study. Am. J. Cardiol. 193, 44–51. doi: 10.1016/j.amjcard.2023.01.048

PubMed Abstract | Crossref Full Text | Google Scholar

Mohd Faizal,, Aizatul Shafiqah,, Thevarajah, T. M., Khor, S. M., and Chang, S.-W. (2021). A review of risk prediction models in cardiovascular disease: conventional approach vs. artificial intelligent approach. Comput. Methods Prog. Biomed. 207:6190. doi: 10.1016/j.cmpb.2021.106190

Crossref Full Text | Google Scholar

Ogunpola, A., Saeed, F., Basurra, S., Albarrak, A. M., and Qasem, S. N. (2024). Machine learning-based predictive models for detection of cardiovascular diseases. Diagnostics 14:144. doi: 10.3390/diagnostics14020144

PubMed Abstract | Crossref Full Text | Google Scholar

Oliveira, M., Seringa, J., Pinto, F. J., Henriques, R., and Magalhães, T. (2023). Machine learning prediction of mortality in acute myocardial infarction. BMC Med. Inform. Decis. Mak. 23:70. doi: 10.1186/s12911-023-02168-6

PubMed Abstract | Crossref Full Text | Google Scholar

Öztekin, A., and Özyılmaz, B. (2025). A machine learning based death risk analysis and prediction of ST-segment elevation myocardial infarction (STEMI) patients. Comput. Biol. Med. 188:109839. doi: 10.1016/j.compbiomed.2025.109839

PubMed Abstract | Crossref Full Text | Google Scholar

Payrovnaziri, S. N., Barrett, L. A., Bis, D., Bian, J., and He, Z. (2019). Enhancing prediction models for one-year mortality in patients with acute myocardial infarction and post myocardial infarction syndrome. Stud. Health Technol. Inform. 264, 273–277. doi: 10.3233/SHTI190226

PubMed Abstract | Crossref Full Text | Google Scholar

Qi, X., Wang, S., Fang, C., Jia, J., Lin, L., and Yuan, T. (2025). Machine learning and SHAP value interpretation for predicting comorbidity of cardiovascular disease and cancer with dietary antioxidants. Redox Biol. 79:103470. doi: 10.1016/j.redox.2024.103470

PubMed Abstract | Crossref Full Text | Google Scholar

Radwa, E., Ridha, H., and Faycal, B. (2024). Deep learning-based approaches for myocardial infarction detection: a comprehensive review recent advances and emerging challenges. Med. Novel Technol. Devices 23:100322. doi: 10.1016/j.medntd.2024.100322

Crossref Full Text | Google Scholar

Razavi, S. R., Zaremba, A. C., Szun, T., Cheung, S., Shah, A. H., and Moussavi, Z. (2025). Comprehensive prediction of outcomes in patients with ST elevation myocardial infarction (STEMI) using tree-based machine learning algorithms. Comput. Biol. Med. 184:109439. doi: 10.1016/j.compbiomed.2024.109439

PubMed Abstract | Crossref Full Text | Google Scholar

Ren, H., Kwok, Q., Sun, M., Huang, X., Zhu, J., and Li, H. (2025). Toward artificial general intelligence in health care. Vis. Comput. 41, 7341–7350. doi: 10.1007/s00371-025-03808-w

Crossref Full Text | Google Scholar

Shi, Fangrui, and Fan, Guochao (2023). “An unbalanced data classification method based on improved SMOTE.” in 2023 4th International Conference on Big Data & Artificial Intelligence & Software Engineering (ICBASE).

Google Scholar

Stephan, A.-J., Hanselmann, M., Bajramovic, M., Schosser, S., and Laxy, M. (2025). Development and validation of prediction models for stroke and myocardial infarction in type 2 diabetes based on health insurance claims: does machine learning outperform traditional regression approaches? Cardiovasc. Diabetol. 24:80. doi: 10.1186/s12933-025-02640-9

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, H., Zu, Q., Chen, J., Yang, Z., and Ahmed, M. A. (2021). Application of artificial intelligence in acute coronary syndrome: a brief literature review. Adv. Ther. 38, 5078–5086. doi: 10.1007/s12325-021-01908-2

PubMed Abstract | Crossref Full Text | Google Scholar

Yang, Y., Yan, Y., Zhou, Z., Zhang, J., Han, H., Zhang, W., et al. (2025). Accurate prediction of bleeding risk after coronary artery bypass grafting with dual antiplatelet therapy: a machine learning model vs. the PRECISE-DAPT score. Int. J. Cardiol. 421:132925. doi: 10.1016/j.ijcard.2024.132925

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, P., Wu, L., Zou, T. T., Zou, Z. X., Tu, J. X., Gong, R., et al. (2023). Machine learning for early prediction of major adverse cardiovascular events after first percutaneous coronary intervention in patients with acute myocardial infarction: retrospective cohort study. JMIR Form. Res. 8:48487. doi: 10.2196/48487

Crossref Full Text | Google Scholar

Zheng, B., Liang, T., Mei, J., Shi, X., Liu, X., Li, S., et al. (2024). Prediction of 90 day readmission in heart failure with preserved ejection fraction by interpretable machine learning. ESC Heart Fail. 11, 4267–4276. doi: 10.1002/ehf2.15033

PubMed Abstract | Crossref Full Text | Google Scholar

Zheng, Z. C., Yuan, W., Wang, N., Jiang, B., Ma, C. P., Ai, H., et al. (2023). Exploring the feasibility of machine learning to predict risk stratification within 3 months in chest pain patients with suspected NSTE-ACS. Biomed. Environ. Sci. 36, 625–634. doi: 10.3967/bes2023.089

PubMed Abstract | Crossref Full Text | Google Scholar

Zhou, Q. M., Zhe, L., Brooke, R. J., Hudson, M. M., and Yuan, Y. (2021). A relationship between the incremental values of area under the ROC curve and of area under the precision-recall curve. Diagn. Progn. Res. 5:13. doi: 10.1186/s41512-021-00102-w

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: one-year mortality, ST-segment elevation myocardial infarction, machine learning, prediction model, data imbalance, data process method

Citation: Li W, Yan D, Hu W, Su X and Zhang Z (2025) Enhancing one-year mortality prediction in STEMI patients post-PCI: an interpretable machine learning model with risk stratification. Front. Artif. Intell. 8:1618492. doi: 10.3389/frai.2025.1618492

Received: 26 April 2025; Accepted: 08 August 2025;
Published: 22 August 2025.

Edited by:

Jerzy Beltowski, Medical University of Lublin, Poland

Reviewed by:

Imen Boudali, Higher National Engineering School of Tunis, Tunisia
Bing Xiao, Second Hospital of Hebei Medical University, China

Copyright © 2025 Li, Yan, Hu, Su and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zheng Zhang, emhhbmdjY3VAMTYzLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.