Consistency Regularization for Adversarial RobustnessDownload PDF

Published: 23 Jun 2021, Last Modified: 22 Oct 2023ICML 2021 Workshop AML OralReaders: Everyone
Keywords: adversarial example, adversarial robustness, consistency regularization
TL;DR: We propose an effective consistency regularization technique that prevents robust overfitting
Abstract: Adversarial training (AT) is currently one of the most successful methods to obtain the adversarial robustness of deep neural networks. However, the phenomenon of robust overfitting, i.e., the robustness starts to decrease significantly during AT, has been problematic, not only making practitioners consider a bag of tricks for a successful training, e.g., early stopping, but also incurring a significant generalization gap in the robustness. In this paper, we propose an effective regularization technique that prevents robust overfitting by optimizing an auxiliary 'consistency' regularization loss during AT. Specifically, it forces the predictive distributions after attacking from two different augmentations of the same instance to be similar with each other. Our experimental results demonstrate that such a simple regularization technique brings significant improvements in the test robust accuracy of a wide range of AT methods. More remarkably, we also show that our method could significantly help the model to generalize its robustness against unseen adversaries, e.g., other types or larger perturbations compared to those used during training.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 3 code implementations](
2 Replies