Attention for Adversarial Attacks: Learning from your MistakesDownload PDF

22 Nov 2021, 06:35 (modified: 08 Dec 2021, 16:50)AAAI-22 AdvML Workshop OralReaders: Everyone
Keywords: Adversarial Examples, Attention Mechanism, Graph Neural Networks, PGD Attack, Adversarial Attacks
TL;DR: We boost PGD, replacing random restarts with an attention mechanism that is based on Graph Neural Network and learns from previous unsuccessful attacks.
Abstract: In order to apply Neural Networks in safety-critical settings, such as healthcare or autonomous driving, we need to be able to analyse their robustness against adversarial attacks. As complete verification is often computationally prohibitive, we rely on cheap and effective adversarial attacks to estimate their robustness. However, state-of-the-art adversarial attacks, such as the frequently used PGD attack, often require many random restarts to generate adversarial examples. Each time we perform a restart we ignore all previous unsuccessful runs. In order to alleviate this inefficiency, we propose a method that learns from its mistakes. Specifically, our method uses Graph Neural Networks (GNNs) as an attention mechanism, to greatly reduce the search space for the attacks. The architecture of the GNN is based on the neural network we are attacking, and we perform forward and backward passes though the GNN mimicking the back-propagation algorithm of PGD attacks. The GNN outputs a smaller subspace for the PGD attack to focus on. Using our method, we manage to boost the attacks' performance: the GNN increases the success rate of PGD by over 35\% on a recent published dataset used for comparing adversarial attacks, while simultaneously reducing its average computation time.
3 Replies