Open Access iconOpen Access

ARTICLE

SparseMoE-MFN: A Sparse Attention and Mixture-of-Experts Framework for Multimodal Fake News Detection on Social Media

Yuechuan Zhang1,2, Mingshu Zhang1,2,*, Bin Wei1,2, Hongyu Jin1,2, Yaxuan Wang1,2

1 College of Cryptography Engineering, Engineering University of People’s Armed Police, Xi’an, China
2 Key Laboratory of Network and Information Security, Engineering University of People’s Armed Police, Xi’an, China

* Corresponding Author: Mingshu Zhang. Email: email

Computers, Materials & Continua 2026, 87(2), 70 https://doi.org/10.32604/cmc.2026.073996

Abstract

Detecting fake news in multimodal and multilingual social media environments is challenging due to inherent noise, inter-modal imbalance, computational bottlenecks, and semantic ambiguity. To address these issues, we propose SparseMoE-MFN, a novel unified framework that integrates sparse attention with a sparse-activated Mixture-of-Experts (MoE) architecture. This framework aims to enhance the efficiency, inferential depth, and interpretability of multimodal fake news detection. SparseMoE-MFN leverages LLaVA-v1.6-Mistral-7B-HF for efficient visual encoding and Qwen/Qwen2-7B for text processing. The sparse attention module adaptively filters irrelevant tokens and focuses on key regions, reducing computational costs and noise. The sparse MoE module dynamically routes inputs to specialized experts (visual, language, cross-modal alignment) based on content heterogeneity. This expert specialization design boosts computational efficiency and semantic adaptability, enabling precise processing of complex content and improving performance on ambiguous categories. Evaluated on the large-scale, multilingual MR2 dataset, SparseMoE-MFN achieves state-of-the-art performance. It obtains an accuracy of 86.7% and a macro-averaged F1 score of 0.859, outperforming strong baselines like MiniGPT-4 by 3.4% and 3.2%, respectively. Notably, it shows significant advantages in the “unverified” category. Furthermore, SparseMoE-MFN demonstrates superior computational efficiency, with an average inference latency of 89.1 ms and 95.4 GFLOPs, substantially lower than existing models. Ablation studies and visualization analyses confirm the effectiveness of both sparse attention and sparse MoE components in improving accuracy, generalization, and efficiency.

Keywords

Fake news detection; multimodal; sparse attention; mixture-of-experts; interpretability; computational efficiency

Cite This Article

APA Style
Zhang, Y., Zhang, M., Wei, B., Jin, H., Wang, Y. (2026). SparseMoE-MFN: A Sparse Attention and Mixture-of-Experts Framework for Multimodal Fake News Detection on Social Media. Computers, Materials & Continua, 87(2), 70. https://doi.org/10.32604/cmc.2026.073996
Vancouver Style
Zhang Y, Zhang M, Wei B, Jin H, Wang Y. SparseMoE-MFN: A Sparse Attention and Mixture-of-Experts Framework for Multimodal Fake News Detection on Social Media. Comput Mater Contin. 2026;87(2):70. https://doi.org/10.32604/cmc.2026.073996
IEEE Style
Y. Zhang, M. Zhang, B. Wei, H. Jin, and Y. Wang, “SparseMoE-MFN: A Sparse Attention and Mixture-of-Experts Framework for Multimodal Fake News Detection on Social Media,” Comput. Mater. Contin., vol. 87, no. 2, pp. 70, 2026. https://doi.org/10.32604/cmc.2026.073996



cc Copyright © 2026 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 559

    View

  • 250

    Download

  • 0

    Like

Share Link