Abstract: As large language models become increasingly deployed, understanding the complexity and evolution of jailbreaking strategies is critical for AI safety.
External IDs:dblp:conf/sgai/CreoFC25
Loading
OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2026 OpenReview