Seunghan Kim1,#, Changhoon Lim2,#, Gwonsang Ryu3, Hyunil Kim2,*
CMES-Computer Modeling in Engineering & Sciences, Vol.145, No.2, pp. 2617-2630, 2025, DOI:10.32604/cmes.2025.071190
- 26 November 2025
Abstract Federated Learning enables privacy-preserving training of Transformer-based language models, but remains vulnerable to backdoor attacks that compromise model reliability. This paper presents a comparative analysis of defense strategies against both classical and advanced backdoor attacks, evaluated across autoencoding and autoregressive models. Unlike prior studies, this work provides the first systematic comparison of perturbation-based, screening-based, and hybrid defenses in Transformer-based FL environments. Our results show that screening-based defenses consistently outperform perturbation-based ones, effectively neutralizing most attacks across architectures. However, this robustness comes with significant computational overhead, revealing a clear trade-off between security and efficiency. By explicitly More >