Keywords: Constraint Learning, Constrained Decoding, Large Language Models, Context-Sensitive Grammars
TL;DR: We present a two-phase framework for automatically learning context-sensitive constraints from LLM generations and using them to guarantee correctness at inference time.
Abstract: Controlling the output of Large Language Models (LLMs) through context-sensitive constraints has emerged as a promising approach to overcome the limitations of Context-Free Grammars (CFGs) in guaranteeing generation validity. However, such constraints typically require manual specification---a significant barrier demanding specialized expertise. We introduce a framework that automatically learns context-sensitive constraints from LLM interactions through a two-phase process: syntactic exploration to gather diverse outputs for constraint learning, followed by constraint exploitation to enforce these learned rules during generation. Experiments demonstrate that our method enables even small LLMs (1B parameters) to learn and generate with perfect constraint adherence, outperforming larger counterparts and state-of-the-art reasoning models. This work represents the first integration of context-sensitive grammar learning with LLM generation, eliminating manual specification while maintaining generation validity.
Archival Status: Archival
Paper Length: Short Paper (up to 4 pages of content)
Submission Number: 199
Loading