Challenger: Affordable Adversarial Driving Video Generation

Published: 22 Nov 2025, Last Modified: 22 Nov 2025SAFE-ROL OralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Autonomous Driving, Adversarial Driving Environments, Diffusion Models, Photorealistic Video Generation
Abstract: Generating photorealistic driving videos has seen significant progress recently, but current methods largely focus on natural scenarios. Meanwhile, efforts to generate adversarial driving scenarios often operate on abstract trajectory or BEV representations, falling short of delivering realistic sensor data that can truly stress-test autonomous driving (AD) systems. In this work, we introduce Challenger, a framework that produces physically plausible yet photorealistic adversarial driving videos. Generating such videos poses a fundamental challenge: it requires jointly optimizing over the space of traffic interactions and high-fidelity sensor observations. Challenger makes this affordable through: (1) a physics-aware multi-round trajectory refinement process that narrows down candidate adversarial maneuvers, and (2) a tailored trajectory scoring function that encourages realistic yet adversarial behavior while maintaining compatibility with downstream video synthesis. As tested on the nuScenes dataset, Challenger generates a diverse range of aggressive driving scenarios—including cut-ins, sudden lane changes, tailgating, and blind spot intrusions—and renders them into multiview photorealistic videos. Extensive evaluations show that these scenarios significantly increase the collision rate of state-of-the-art end-to-end AD models (UniAD, VAD, SparseDrive, and DiffusionDrive), and importantly, adversarial behaviors discovered for one model often transfer to others. Our code, models, and dataset can be found at https://pixtella.github.io/Challenger/ .
Supplementary Zip: zip
Submission Number: 7
Loading