ORIGINAL RESEARCH article
Front. Pharmacol.
Sec. Cardiovascular and Smooth Muscle Pharmacology
This article is part of the Research TopicInnovative Approaches and Molecular Mechanisms in Cardiovascular PharmacologyView all 24 articles
Ensemble Learning Approach with Explainable AI for Improved Heart Disease Prediction
Provisionally accepted- 1Birmingham City University, Birmingham, United Kingdom
- 2University of Tabuk, Tabuk, Saudi Arabia
- 3Imam Muhammad Ibn Saud Islamic University, Riyadh, Saudi Arabia
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Heart disease remains a leading cause of global morbidity and mortality, motivating the development of predictive models that are both accurate and clinically interpretable. We introduce the Interpretable Ensemble Learning Framework (IELF), which integrates Explainable Boosting Machines (EBM) with XGBoost, SHAP-based explanations, and LIME for enhanced local interpretability. IELF was evaluated on two benchmark datasets: Cleveland (n=303) and Framingham (n=4,240). Across 5-fold cross-validation and held-out test sets, IELF achieved robust discrimination (AUC 0.899, accuracy 88.5% on Cleveland; AUC 0.696, accuracy 82.6% on Framingham) while maintaining balanced precision–recall trade-offs. Compared with EBM, IELF delivered significant improvements in recall, F1, and AUC on the Framingham dataset (p < 0.05), with differences versus XGBoost being less consistent. Beyond predictive performance, IELF provided transparent feature rankings aligned with established cardiovascular risk factors (e.g., age, blood pressure, cholesterol, diabetes, smoking), stable explanations across folds, and subgroup fairness analyses. To our knowledge, IELF is the first framework to combine EBM and XGBoost with SHAP and LIME, providing quantitative validation of explanation stability (Kendall’s τ, Overlap@10) under strict nested cross-validation with calibration and subgroup analyses. IELF further incorporates local feature explainability through LIME, offering complementary patient-level insights beyond SHAP. While headline accuracies were lower than recent state-of-the-art models reporting >97%, IELF was developed under stricter methodological controls (nested CV, calibration, subgroup validation), ensuring reproducibility, interpretability, and clinical reliability. These findings position IELF as a trustworthy benchmark for translational AI in cardiovascular risk prediction, complementing high accuracy but less transparent models.
Keywords: cardiovascular disease, Explainable AI, Heart disease, ensemble learning, XGBoost
Received: 26 Jun 2025; Accepted: 18 Nov 2025.
Copyright: © 2025 Adekoya, Saeed, Ghaban and Qasem. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Sultan Noman Qasem, snmohammed@imamu.edu.sa
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.