Interactive Agents to Overcome Underspecificity in Software Engineering

ICLR 2026 Conference Submission20744 Authors

19 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Ambiguity, Underspecificity, SWE Agent, Software Engineering, Clarification, Evaluation, Interaction
TL;DR: We study how LLM agents handle underspecified instructions in interactive code generation, focusing on (a) using interactivity to improve performance, (b) detecting underspecificity, and (c) asking targeted questions.
Abstract: AI agents are increasingly being deployed to automate tasks, often based on underspecified user instructions. Making unwarranted assumptions to compensate for the missing information and failing to ask clarifying questions can lead to suboptimal outcomes, safety risks due to tool misuse, and wasted computational resources. In this work, we study the ability of LLM agents to handle underspecified instructions in interactive code generation settings by evaluating proprietary and open-weight models on their performance across three key steps: (a) detecting underspecificity, (b) asking targeted clarification questions, and (c) leveraging the interaction to improve performance in underspecified scenarios. Our findings reveal that models struggle to distinguish between well-specified and underspecified instructions. However, when models interact for underspecified inputs, they effectively obtain vital information from the user leading to significant improvements in performance, up to 74\% over the non-interactive settings, underscoring the value of effective interaction. Our study highlights critical gaps in how current state-of-the-art models handle missing information in complex software engineering tasks and structures the evaluation into distinct steps to enable targeted improvements.
Supplementary Material: zip
Primary Area: datasets and benchmarks
Submission Number: 20744
Loading