Examining the Human Perceptibility of Black-Box Adversarial Attacks on Face RecognitionDownload PDF

Jun 18, 2021 (edited Jul 19, 2021)ICML 2021 Workshop AML PosterReaders: Everyone
  • Keywords: adversarial attack, black-box attack, face recognition, human perceptibility
  • TL;DR: Traditional bounding metrics are insufficient for approximating human perception, and human surveys show leading attacks become perceptible as success rates increase.
  • Abstract: The modern open internet contains billions of public images of human faces across the web, especially on social media websites used by half the world's population. In this context, Face Recognition (FR) systems have the potential to match faces to specific names and identities, creating glaring privacy concerns. Adversarial attacks are a promising way to grant users privacy from FR systems by disrupting their capability to recognize faces. Yet, such attacks can be perceptible to human observers, especially under the more challenging black-box threat model. In the literature, the justification for the imperceptibility of such attacks hinges on bounding metrics such as $\ell_p$ norms. However, there is not much research on how these norms match up with human perception. Through examining and measuring both the effectiveness of recent black-box attacks in the face recognition setting and their corresponding human perceptibility through survey data, we demonstrate the trade-offs in perceptibility that occur as attacks become more aggressive. We also show how the $\ell_2$ norm and other metrics do not correlate with human perceptibility in a linear fashion, thus making these norms suboptimal at measuring adversarial attack perceptibility.
2 Replies