Qian Zhu1, Qian Kang1, Tao Xu2, Dengxiu Yu3,*, Zhen Wang1
CMC-Computers, Materials & Continua, Vol.83, No.2, pp. 1855-1879, 2025, DOI:10.32604/cmc.2025.061913
- 16 April 2025
Abstract In this study, we present a deterministic convergence analysis of Gated Recurrent Unit (GRU) networks enhanced by a smoothing regularization technique. While GRU architectures effectively mitigate gradient vanishing/exploding issues in sequential modeling, they remain prone to overfitting, particularly under noisy or limited training data. Traditional regularization, despite enforcing sparsity and accelerating optimization, introduces non-differentiable points in the error function, leading to oscillations during training. To address this, we propose a novel smoothing regularization framework that replaces the non-differentiable absolute function with a quadratic approximation, ensuring gradient continuity and stabilizing the optimization landscape. Theoretically, we rigorously… More >