Shift Invariance Can Reduce Adversarial RobustnessDownload PDF

21 May 2021, 20:44 (modified: 21 Jan 2022, 23:44)NeurIPS 2021 PosterReaders: Everyone
Keywords: adversarial robustness, adversarial examples, adversarial machine learning, shift invariance
TL;DR: We provide theoretical and empirical evidence that the property of shift invariance in convolutional neural networks can decrease adversarial robustness.
Abstract: Shift invariance is a critical property of CNNs that improves performance on classification. However, we show that invariance to circular shifts can also lead to greater sensitivity to adversarial attacks. We first characterize the margin between classes when a shift-invariant {\em linear} classifier is used. We show that the margin can only depend on the DC component of the signals. Then, using results about infinitely wide networks, we show that in some simple cases, fully connected and shift-invariant neural networks produce linear decision boundaries. Using this, we prove that shift invariance in neural networks produces adversarial examples for the simple case of two classes, each consisting of a single image with a black or white dot on a gray background. This is more than a curiosity; we show empirically that with real datasets and realistic architectures, shift invariance reduces adversarial robustness. Finally, we describe initial experiments using synthetic data to probe the source of this connection.
Supplementary Material: zip
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/SongweiGe/shift-invariance-adv-robustness
14 Replies

Loading