Applied Sciences (Sep 2024)

Alzheimer’s Multiclassification Using Explainable AI Techniques

  • Kamese Jordan Junior,
  • Kouayep Sonia Carole,
  • Tagne Poupi Theodore Armand,
  • Hee-Cheol Kim,
  • The Alzheimer’s Disease Neuroimaging Initiative

DOI
https://doi.org/10.3390/app14188287
Journal volume & issue
Vol. 14, no. 18
p. 8287

Abstract

Read online

In this study, we address the early detection challenges of Alzheimer’s disease (AD) using explainable artificial intelligence (XAI) techniques. AD, characterized by amyloid plaques and tau tangles, leads to cognitive decline and remains hard to diagnose due to genetic and environmental factors. Utilizing deep learning models, we analyzed brain MRI scans from the ADNI database, categorizing them into normal cognition (NC), mild cognitive impairment (MCI), and AD. The ResNet-50 architecture was employed, enhanced by a channel-wise attention mechanism to improve feature extraction. To ensure model transparency, we integrated local interpretable model-agnostic explanations (LIMEs) and gradient-weighted class activation mapping (Grad-CAM), highlighting significant image regions contributing to predictions. Our model achieved 85% accuracy, effectively distinguishing between the classes. The LIME and Grad-CAM visualizations provided insights into the model’s decision-making process, particularly emphasizing changes near the hippocampus for MCI. These XAI methods enhance the interpretability of AI-driven AD diagnosis, fostering trust and aiding clinical decision-making. Our approach demonstrates the potential of combining deep learning with XAI for reliable and transparent medical applications.

Keywords