Improved robustness to adversarial examples using Lipschitz regularization of the lossDownload PDF

27 Sept 2018 (modified: 14 Oct 2024)ICLR 2019 Conference Blind SubmissionReaders: Everyone
Abstract: We augment adversarial training (AT) with worst case adversarial training (WCAT) which improves adversarial robustness by 11% over the current state- of-the-art result in the `2-norm on CIFAR-10. We interpret adversarial training as Total Variation Regularization, which is a fundamental tool in mathematical im- age processing, and WCAT as Lipschitz regularization, which appears in Image Inpainting. We obtain verifiable worst and average case robustness guarantees, based on the expected and maximum values of the norm of the gradient of the loss.
Keywords: Adversarial training, adversarial examples, deep neural networks, regularization, Lipschitz constant
TL;DR: Improvements to adversarial robustness, as well as provable robustness guarantees, are obtained by augmenting adversarial training with a tractable Lipschitz regularization
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/improved-robustness-to-adversarial-examples/code)
15 Replies

Loading