Home / Journals / CMC / Online First / doi:10.32604/cmc.2026.076400
Special Issues
Table of Content

Open Access

ARTICLE

Enhancing SHAP Explainability for Diagnostic and Prognostic ML Models in Alzheimer’s Disease

Pablo Guillén1, Enrique Frias-Martinez2,*
1 MIU City University, Miami, FL, USA
2 Research Institute UNIR iTED, Universidad Internacional de La Rioja (UNIR), La Rioja, Spain
* Corresponding Author: Enrique Frias-Martinez. Email: email
(This article belongs to the Special Issue: Bridging the Gap: AutoML and Explainable AI for Industrial and Healthcare Innovations)

Computers, Materials & Continua https://doi.org/10.32604/cmc.2026.076400

Received 19 November 2025; Accepted 05 February 2026; Published online 14 February 2026

Abstract

Alzheimer’s disease (AD) diagnosis and prognosis increasingly rely on machine learning (ML) models. Although these models provide good results, clinical adoption is limited by the need for technical expertise and the lack of trustworthy and consistent model explanations. SHAP (SHapley Additive exPlanations) is commonly used to interpret AD models, but existing studies tend to focus on explanations for isolated tasks, providing little evidence about their robustness across disease stages, model architectures, or prediction objectives. This paper proposes a multi-level explainability framework that measures the coherence, stability and consistency of explanations by integrating: (1) within-model coherence metrics between feature importance and SHAP, (2) SHAP stability across AD boundaries, and (3) SHAP cross-task consistency between diagnosis and prognosis. Using AutoML to optimize classifiers on the NACC dataset, we trained four diagnostic and four prognostic models covering the standard AD progression stages: normal-control (NC), mild-cognitive impairment (NCI) and AD. For each model, we generated SHAP and feature importance (FI) plots. Stability was then evaluated using correlation metrics (Spearman, Kendall), top-k feature overlap (Jaccard@10/20), SHAP sign consistency, and domain-level contribution ratios. Results show that cognitive and functional markers (e.g., MEMORY, JUDGMENT, ORIENT, PAYATTN) dominate SHAP explanations in both diagnosis and prognosis. SHAP-SHAP consistency between diagnostic and prognostic models was high across all classifiers (ρ = 0.61–0.94), with 100% sign stability and minimal shifts in explanatory magnitude (mean Δ|SHAP| < 0.03). Domain-level contributions also remained stable, with only minimal increases in genetic features for prognosis. These results demonstrate that SHAP explanations can be quantitatively validated for robustness and transferability, providing clinicians with more reliable interpretations of ML predictions. The proposed framework provides a reproducible methodology for evaluating explainability stability and coherence, supporting the deployment of trustworthy ML systems in AD clinical settings.

Keywords

Alzheimer’s disease (AD); automated machine learning (AutoML); Pycaret; SHAP
  • 283

    View

  • 49

    Download

  • 0

    Like

Share Link