Adversarial Attacks on Robotic Vision Language Action Models

Published: 20 Jun 2025, Last Modified: 20 Jun 2025RSS 2025 Workshop ReliableRoboticsEveryoneRevisionsBibTeXCC BY 4.0
Keywords: adversarial attacks, robotics, foundation models
TL;DR: We propose adversarial attacks for robotic vision-language-action models
Abstract: The emergence of vision-language-action models (VLAs) for end-to-end control is reshaping the field of robotics by enabling the fusion of multimodal sensory inputs at the billion-parameter scale. The capabilities of VLAs stem primarily from their architectures, which are often based on frontier large language models (LLMs). However, LLMs are known to be susceptible to adversarial misuse, and given the significant physical risks inherent to robotics, questions remain regarding the extent to which VLAs inherit these vulnerabilities. Motivated by these concerns, in this work we initiate the study of adversarial attacks on VLA-controlled robots.
Submission Number: 10
Loading