TY - EJOU AU - Alzamil, Deema AU - Alkhamees, Bader AU - Hassan, Mohammad Mehedi TI - A Systematic Review of Multimodal Fusion and Explainable AI Applications in Breast Cancer Diagnosis T2 - Computer Modeling in Engineering \& Sciences PY - 2025 VL - 145 IS - 3 SN - 1526-1506 AB - Breast cancer diagnosis relies heavily on many kinds of information from diverse sources—like mammogram images, ultrasound scans, patient records, and genetic tests—but most AI tools look at only one of these at a time, which limits their ability to produce accurate and comprehensive decisions. In recent years, multimodal learning has emerged, enabling the integration of heterogeneous data to improve performance and diagnostic accuracy. However, doctors cannot always see how or why these AI tools make their choices, which is a significant bottleneck in their reliability, along with adoption in clinical settings. Hence, people are adding explainable AI techniques that show the steps the model takes. This review investigates previous work that has employed multimodal learning and XAI for the diagnosis of breast cancer. It discusses the types of data, fusion techniques, and XAI models employed. It was done following the PRISMA guidelines and included studies from 2021 to April 2025. The literature search was performed systematically and resulted in 61 studies. The review highlights a gradual increase in current studies focusing on multimodal fusion and XAI, particularly in the years 2023–2024. It found that studies using multi-modal data fusion achieved the highest accuracy by 5%–10% on average compared to other studies that used single-modality data, an intermediate fusion strategy, and modern fusion techniques, such as cross attention, achieved the highest accuracy and best performance. The review also showed that SHAP, Grad-CAM, and LIME techniques are the most used in explaining breast cancer diagnostic models. There is a clear research shift toward integrating multimodal learning and XAI techniques into the breast cancer diagnostics field. However, several gaps were identified, including the scarcity of public multimodal datasets. Lack of a unified explainable framework in multimodal fusion systems, and lack of standardization in evaluating explanations. These limitations call for future research focused on building more shared datasets and integrating multimodal data and explainable AI techniques to improve decision-making and enhance transparency. KW - Breast cancer; classification; explainable artificial intelligence; XAI; deep learning; multi-modal data; explainability; data fusion DO - 10.32604/cmes.2025.070867