Home / Journals / CMC / Online First / doi:10.32604/cmc.2026.073996
Special Issues
Table of Content

Open Access

ARTICLE

SparseMoE-MFN: A Sparse Attention and Mixture-of-Experts Framework for Multimodal Fake News Detection on Social Media

Yuechuan Zhang1,2, Mingshu Zhang1,2,*, Bin Wei1,2, Hongyu Jin1,2, Yaxuan Wang1,2
1 College of Cryptography Engineering, Engineering University of People’s Armed Police, Xi’an, China
2 Key Laboratory of Network and Information Security, Engineering University of People’s Armed Police, Xi’an, China
* Corresponding Author: Mingshu Zhang. Email: email

Computers, Materials & Continua https://doi.org/10.32604/cmc.2026.073996

Received 29 September 2025; Accepted 11 December 2025; Published online 18 February 2026

Abstract

Detecting fake news in multimodal and multilingual social media environments is challenging due to inherent noise, inter-modal imbalance, computational bottlenecks, and semantic ambiguity. To address these issues, we propose SparseMoE-MFN, a novel unified framework that integrates sparse attention with a sparse-activated Mixture-of-Experts (MoE) architecture. This framework aims to enhance the efficiency, inferential depth, and interpretability of multimodal fake news detection. SparseMoE-MFN leverages LLaVA-v1.6-Mistral-7B-HF for efficient visual encoding and Qwen/Qwen2-7B for text processing. The sparse attention module adaptively filters irrelevant tokens and focuses on key regions, reducing computational costs and noise. The sparse MoE module dynamically routes inputs to specialized experts (visual, language, cross-modal alignment) based on content heterogeneity. This expert specialization design boosts computational efficiency and semantic adaptability, enabling precise processing of complex content and improving performance on ambiguous categories. Evaluated on the large-scale, multilingual MR2 dataset, SparseMoE-MFN achieves state-of-the-art performance. It obtains an accuracy of 86.7% and a macro-averaged F1 score of 0.859, outperforming strong baselines like MiniGPT-4 by 3.4% and 3.2%, respectively. Notably, it shows significant advantages in the “unverified” category. Furthermore, SparseMoE-MFN demonstrates superior computational efficiency, with an average inference latency of 89.1 ms and 95.4 GFLOPs, substantially lower than existing models. Ablation studies and visualization analyses confirm the effectiveness of both sparse attention and sparse MoE components in improving accuracy, generalization, and efficiency.

Keywords

Fake news detection; multimodal; sparse attention; mixture-of-experts; interpretability; computational efficiency
  • 93

    View

  • 16

    Download

  • 0

    Like

Share Link