IM-BERT: Enhancing Robustness of BERT through the Implicit Euler Method
IM-BERT: Enhancing Robustness of BERT through the Implicit Euler Method
Pre-trained Language Models (PLMs) have achieved remarkable performance on diverse NLP tasks through pre-training and fine-tuning. However, fine-tuning the model with a large number of parameters on limited downstream datasets often leads to vulnerability to adversarial attacks, causing overfitting of the model on standard datasets. To address these issues, we propose IM-BERT from the perspective of a dynamic system by conceptualizing a layer of BERT as a solution of Ordinary Differential Equations (ODEs). Under the situation of initial value perturbation, we analyze the numerical stability of two main numerical ODE solvers: the explicit and implicit Euler approaches. Based on these analyses, we introduce a numerically robust IM-connection incorporating BERT's layers. This strategy enhances the robustness of PLMs against adversarial attacks, even in low-resource scenarios, without introducing additional parameters or adversarial training strategies. Experimental results on the adversarial GLUE (AdvGLUE) dataset validate the robustness of IM-BERT under various conditions. Compared to the original BERT, IM-BERT exhibits a performance improvement of approximately 8.3\%p on the AdvGLUE dataset. Furthermore, in low-resource scenarios, IM-BERT outperforms BERT by achieving 5.9\%p higher accuracy.
Mihyeon Kim、Juhyoung Park、Youngbin Kim
计算技术、计算机技术
Mihyeon Kim,Juhyoung Park,Youngbin Kim.IM-BERT: Enhancing Robustness of BERT through the Implicit Euler Method[EB/OL].(2025-05-11)[2025-07-17].https://arxiv.org/abs/2505.06889.点此复制
评论