TtBA: Two-third Bridge Approach for Decision-Based Adversarial Attack

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: We propose a novel Two-third Bridge Attack (TtBA) for decision-based black-box attack.
Abstract: A key challenge in black-box adversarial attacks is the high query complexity in hard-label settings, where only the top-1 predicted label from the target deep model is accessible. In this paper, we propose a novel normal-vector-based method called Two-third Bridge Attack (TtBA). A innovative bridge direction is introduced which is a weighted combination of the current unit perturbation direction and its unit normal vector, controlled by a weight parameter $k$. We further use binary search to identify $k=k_\text{bridge}$, which has identical decision boundary as the current direction. Notably, we observe that $k=2/3 k_\text{bridge}$ yields a near-optimal perturbation direction, ensuring the stealthiness of the attack. In addition, we investigate the critical importance of local optima during the perturbation direction optimization process and propose a simple and effective approach to detect and escape such local optima. Experimental results on MNIST, FASHION-MNIST, CIFAR10, CIFAR100, and ImageNet datasets demonstrate the strong performance and scalability of our approach. Compared to state-of-the-art non-targeted and targeted attack methods, TtBA consistently delivers superior performance across most experimented datasets and deep learning models. Code is available at https://anonymous.4open.science/r/TtBA-6ECF.
Lay Summary: Deep learning systems, such as those used in image recognition, can be easily fooled by tiny, carefully crafted changes called adversarial attacks. These changes are often invisible to the human eye but can cause models to make incorrect decisions. In decision-based black box attack settings, where only the final label (such as “cat” or “dog”) output by the model is visible, generating such attacks becomes especially difficult and requires many attempts. Our research introduces a method called the *Two-third Bridge Attack* (TtBA), which significantly reduces the number of attempts needed to successfully fool a model. We propose a novel metric, $k_\text{bridge}$, to capture the shape of a model’s decision boundary and discover that using $2/3k_\text{bridge}$ leads to an effective attack. This metric also helps detect when the attack is stuck in a suboptimal region and guides it toward better attacks. By uncovering these vulnerabilities, our work contributes to developing more robust and trustworthy AI systems that are safer for real-world use.
Link To Code: https://github.com/BUPTAIOC/TtBA
Primary Area: Deep Learning->Robustness
Keywords: Adversarial Attacks, Black Box Adversarial Attacks, Hard Label Attacks, Decision-Based Attacks, Machine Learning Security, Artificial Intelligence Security, Robustness
Submission Number: 1583
Loading