Home / Journals / CMES / Online First / doi:10.32604/cmes.2026.080335
Special Issues
Table of Content

Open Access

ARTICLE

Multimodal Graph-Enhanced Vision Transformer for Interpretable Skin Lesion Classification

Faten S. Alamri1, Noor Ayesha2, Afia Zafar3, Adil Ali Saleem4,*, Amjad R. Khan5
1 Department of Mathematical Sciences, College of Science, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
2 Center of Excellence in Cyber Security (CYBEX), Prince Sultan University, Riyadh, Saudi Arabia
3 Computer Science Department, The National University of Computer and Emerging Sciences (NUCES-FAST), Islamabad, Pakistan
4 Institute of Computer Science, Khwaja Fareed University of Engineering and Information Technology, Abu Dhabi Road, Rahim Yar Khan, Punjab, Pakistan
5 Artificial Intelligence & Data Analytics Lab, College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
* Corresponding Author: Adil Ali Saleem. Email: email
(This article belongs to the Special Issue: Advances in Deep Learning and Computer Vision for Intelligent Systems: Methods, Applications, and Future Directions)

Computer Modeling in Engineering & Sciences https://doi.org/10.32604/cmes.2026.080335

Received 07 February 2026; Accepted 18 March 2026; Published online 01 April 2026

Abstract

The use of automated skin lesion classification is still a disadvantage, since there is a great visual similarity between benign and malignant lesions. The majority of deep learning methods utilize dermoscopic images only, without taking into account clinical metadata employed by dermatologists on a regular basis. The following paper proposes a vision-graph multimodal framework that links Image encoding to graph neural networks based on metadata representation through the fusion of learnable attention. The framework focuses on three limitations, which are underutilization of clinical context, absence of interpretability, and suboptimal incorporation of modalities. Gradient-weighted Class Activation Mapping++ (Grad-CAM++) is used to obtain dual explainability of visual attention, and SHapley Additive exPlanations (SHAP) to obtain feature importance. Examining the HAM10000 and Derm7pt datasets, statistically significant advances (p < 0.001) of 89.3% and 92.1% accuracy are obtained, which is 4.1% and 2.7% higher than baselines that can only use images. Focusing on weight analysis will provide metadata with 37.7% averaged variance with an error of 8.4%, which confirms the clinical importance of multimodal modeling. The study of ablation shows that graph-based metadata encoding is 1.4% better than standard multilayer perceptron encoding (p = 0.003).

Graphical Abstract

Multimodal Graph-Enhanced Vision Transformer for Interpretable Skin Lesion Classification

Keywords

Skin lesion classification; vision transformer; graph neural network; multimodal learning; explainable AI; medical image analysis
  • 189

    View

  • 46

    Download

  • 0

    Like

Share Link