Track: long paper (up to 10 pages)
Keywords: belief revision, AGM theory, large language models, logical reasoning, benchmark, rationality postulates, belief dynamics, iterated revision
TL;DR: AGM-Bench is the first benchmark grounded in AGM belief revision theory; we find that LLMs frequently violate Inclusion and Preservation and show belief inertia and collateral damage under iterated revision.
Abstract: Large language models (LLMs) are increasingly deployed in settings that require updating conclusions as new information arrives, from multi-turn dialogue to agentic workflows with evolving evidence. Yet virtually all evaluations of LLM logical reasoning focus on static problems: given fixed premises, derive a conclusion. We introduce AGM-Bench, the first benchmark grounded in the AGM theory of belief revision, which tests whether LLMs update their beliefs rationally when confronted with new, potentially contradictory information. AGM-Bench operationalizes six classical rationality postulates, namely Success, Consistency, Inclusion, Vacuity, Extensionality, and Preservation, as well as the Darwiche–Pearl postulates for iterated revision, across 2,400 synthetic reasoning scenarios of controlled logical complexity. We evaluate seven frontier LLMs and find that: (1) all models satisfy Success and Consistency at high rates, but systematically violate Inclusion (minimal change) and Preservation (stability of unrelated beliefs); (2) under iterated revision, models exhibit severe belief inertia (retaining retracted information) and collateral damage (retracting beliefs not logically affected by the new evidence); and (3) reasoning-trained models (o3-mini, DeepSeek-R1) show improved single-step revision but degrade faster under iteration than standard chat models. Our results reveal a fundamental gap between LLM reasoning and rational belief dynamics.
Anonymization: This submission has been anonymized for double-blind review via the removal of identifying information such as names, affiliations, and identifying URLs.
Funding: Yes, the presenting author of this submission falls under ICLR’s funding aims, and funding would significantly impact their ability to attend the workshop in person.
Submission Number: 37
Loading