Proactive Interference Reveals Working Memory Limits in LLMs Beyond Context Length

Published: 10 Jan 2026, Last Modified: 10 Jan 2026LaMAS 2026 OralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: agent tracking; agent state maintenance; model evaluation, long-context language models, working memory limitations, contextual interference, In-context learning, proactive interference, robustness & reliability, top‑down control, cognitive‑science–inspired evaluation
TL;DR: LLMs fail to retrieve recent updates when earlier input interferes, revealing working-memory-like limits beyond context length and undermining Agent state tracking.
Abstract: Information retrieval in Large Language Models (LLMs) is increasingly recognized as intertwined with generation capabilities rather than mere lookup. While longer contexts are often assumed to improve retrieval, the effects of intra-context interference remain understudied. To address this, we adapt the proactive interference (PI) paradigm from cognitive science, where earlier information disrupts recall of newer updates. In humans, susceptibility to such interference is inversely linked to working memory capacity. We introduce PI-LLM, an evaluation that sequentially streams co-referenced key–value updates and queries only the final values. Although these final values are clearly positioned just before the query, LLM retrieval accuracy declines log-linearly toward zero as co-referenced interference accumulates; errors arise from retrieving previously overwritten values. Attempts to mitigate interference via prompt engineering (e.g., instructing models to ignore earlier input) yield limited success. These findings reveal a fundamental constraint on LLMs’ ability to disentangle interference and flexibly manipulate information, suggesting a working memory bottleneck beyond mere context access. For Agentic systems, reliable operation hinges on reconciling past and present states. Proactive interference corrupts long-horizon state maintenance; dependable agent tracking therefore needs explicit memory control and interference-aware context management. We expose a “know, cannot do” failure. Coreference-only needles expose a plan–execute disjunction: LLMs form the correct last-value retrieval plan but fail to carry it out, with execution reliability declining systematically with task complexity. Code and data will be publicly available.
Submission Number: 9
Loading