Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Pharmacol.

Sec. Cardiovascular and Smooth Muscle Pharmacology

This article is part of the Research TopicInnovative Approaches and Molecular Mechanisms in Cardiovascular PharmacologyView all 24 articles

Ensemble Learning Approach with Explainable AI for Improved Heart Disease Prediction

Provisionally accepted
  • 1Birmingham City University, Birmingham, United Kingdom
  • 2University of Tabuk, Tabuk, Saudi Arabia
  • 3Imam Muhammad Ibn Saud Islamic University, Riyadh, Saudi Arabia

The final, formatted version of the article will be published soon.

Heart disease remains a leading cause of global morbidity and mortality, motivating the development of predictive models that are both accurate and clinically interpretable. We introduce the Interpretable Ensemble Learning Framework (IELF), which integrates Explainable Boosting Machines (EBM) with XGBoost, SHAP-based explanations, and LIME for enhanced local interpretability. IELF was evaluated on two benchmark datasets: Cleveland (n=303) and Framingham (n=4,240). Across 5-fold cross-validation and held-out test sets, IELF achieved robust discrimination (AUC 0.899, accuracy 88.5% on Cleveland; AUC 0.696, accuracy 82.6% on Framingham) while maintaining balanced precision–recall trade-offs. Compared with EBM, IELF delivered significant improvements in recall, F1, and AUC on the Framingham dataset (p < 0.05), with differences versus XGBoost being less consistent. Beyond predictive performance, IELF provided transparent feature rankings aligned with established cardiovascular risk factors (e.g., age, blood pressure, cholesterol, diabetes, smoking), stable explanations across folds, and subgroup fairness analyses. To our knowledge, IELF is the first framework to combine EBM and XGBoost with SHAP and LIME, providing quantitative validation of explanation stability (Kendall’s τ, Overlap@10) under strict nested cross-validation with calibration and subgroup analyses. IELF further incorporates local feature explainability through LIME, offering complementary patient-level insights beyond SHAP. While headline accuracies were lower than recent state-of-the-art models reporting >97%, IELF was developed under stricter methodological controls (nested CV, calibration, subgroup validation), ensuring reproducibility, interpretability, and clinical reliability. These findings position IELF as a trustworthy benchmark for translational AI in cardiovascular risk prediction, complementing high accuracy but less transparent models.

Keywords: cardiovascular disease, Explainable AI, Heart disease, ensemble learning, XGBoost

Received: 26 Jun 2025; Accepted: 18 Nov 2025.

Copyright: © 2025 Adekoya, Saeed, Ghaban and Qasem. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Sultan Noman Qasem, snmohammed@imamu.edu.sa

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.