Enhancing Adversarial Robustness through Multi-Objective Representation Learning
Enhancing Adversarial Robustness through Multi-Objective Representation Learning
Deep neural networks (DNNs) are vulnerable to small adversarial perturbations, which are tiny changes to the input data that appear insignificant but cause the model to produce drastically different outputs. Many defense methods require modifying model architectures during evaluation or performing test-time data purification. This not only introduces additional complexity but is often architecture-dependent. We show, however, that robust feature learning during training can significantly enhance DNN robustness. We propose MOREL, a multi-objective approach that aligns natural and adversarial features using cosine similarity and multi-positive contrastive losses to encourage similar features for same-class inputs. Extensive experiments demonstrate that MOREL significantly improves robustness against both white-box and black-box attacks. Our code is available at https://github.com/salomonhotegni/MOREL
Sebastian Peitz、Sedjro Salomon Hotegni
计算技术、计算机技术
Sebastian Peitz,Sedjro Salomon Hotegni.Enhancing Adversarial Robustness through Multi-Objective Representation Learning[EB/OL].(2025-06-29)[2025-07-16].https://arxiv.org/abs/2410.01697.点此复制
评论