Expert Survey: Technical AI Safety & Security Research Priorities

Published: 05 Jun 2025, Last Modified: 15 Jul 2025ICML 2025 Workshop TAIG PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: AI safety, AI security, Technical AI governance, Risk mitigation, Expert survey, Deceptive AI behavior, AI oversight, Model evaluations, Cyber-physical security, Mechanistic interpretability, Frontier model access, Scalable alignment, Policy-relevant AI research
TL;DR: We surveyed 53 experts to identify which technical AI safety and security research directions are most likely to reduce risk and are tractable with moderate funding.
Abstract: As AI systems approach broadly human-level performance, safety and security research is urgently needed to prevent severe harms from AI and ensure its benefits can be safely and reliably realized. To inform strategic investment, we surveyed 53 experts on the importance and tractability of 105 technical AI safety and security research areas. Each expert was asked to rate a subset of areas within their expertise. Our survey revealed several highly promising research directions centered around robust early warning and monitoring of AI risks. Some of the most promising sub-areas included specific capability evaluations (e.g., CBRN, cyber, and deception), understanding emergence and scaling laws, and advancing agent oversight. This study is the first to quantify expert priorities across a comprehensive taxonomy of AI safety and security research directions and to produce a data-driven ranking of their potential impact. These rankings may support evidence-based decisions about how to effectively deploy resources toward AI safety and security research.
Submission Number: 44
Loading