Bootstrapping LLM Agents via VerificationDownload PDF

Anonymous

16 Feb 2024ACL ARR 2024 February Blind SubmissionReaders: Everyone
Abstract: We present a self-training method that allows language model-based agents to improve performance without distilling proprietary models. Existing self-verification methods struggle to validate function signatures defined in agent prompts. A common failure is the verifier hallucinating non-existent constraints on function calls due to interference between model knowledge and examples in prompts. To address this, we devise a neural-symbolic verification system that prioritizes language models for validating solution completeness and pertinence while delegating fact checks to a symbolic system. We propose bootstrap-by-verification learning which combines massive agent trajectory sampling with our verification for self-training. Experiments on spreadsheet and web browsing benchmarks show the method's effectiveness.
Paper Type: long
Research Area: NLP Applications
Languages Studied: English
0 Replies

Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview