Breaking Machine Learning Models with Adversarial Attacks and its Variants

Published: 2025, Last Modified: 04 Sept 2025FLAIRS 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Machine learning models can be by adversarial attacks, subtle, imperceptible perturbations to inputs that cause the model to produce erroneous outputs. This tutorial introduces adversarial examples and its variants, explaining why even stateof-the-art models are vulnerable and how this impacts security in AI. It provides an overview of key concepts (such as black-box vs. white-box attack scenarios) and survey common attack techniques and defensive strategies. A hands-on component using Google Colab and the open-source Adversarial Lab toolkit allows attendees to craft adversarial examples and test model robustness in real time. Throughout, we emphasize both the practical skills and the ethical considerations needed to apply adversarial machine learning in a responsiblemanner. Attendees will gain a comprehensive foundationin adversarial attacks and insights into building morerobust, secure machine learning models.
Loading