A Closer Look at the Adversarial Robustness of Information Bottleneck ModelsDownload PDF

Published: 21 Jun 2021, Last Modified: 03 Jul 2024ICML 2021 Workshop AML PosterReaders: Everyone
Keywords: Information Bottlenecks, Adversarial Robustness
TL;DR: Information bottleneck models are less robust to adversarial attacks than previously thought
Abstract: We study the adversarial robustness of information bottleneck models for classification. Previous works showed that the robustness of models trained with information bottlenecks can improve upon adversarial training. Our evaluation under a diverse range of white-box $l_{\infty}$ attacks suggests that information bottlenecks alone are not a strong defense strategy, and that previous results were likely influenced by gradient obfuscation.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 2 code implementations](https://www.catalyzex.com/paper/a-closer-look-at-the-adversarial-robustness/code)
2 Replies

Loading