Track: tiny / short paper (up to 4 pages)
Keywords: agentic AI, LLM reasoning, context architecture, planning consistency, logical reasoning, multi-agent systems, tool use, Model Context Protocol
Abstract: Large Language Models (LLMs) serving as autonomous agents often conflate reasoning (planning) with action (tool execution) within a single context window. We investigate how context architecture impacts logical consistency and planning effectiveness through six ablation studies using MCPBench with GPT-4o, evaluating monolithic, partially partitioned, and fully role-separated architectures under 6-25 concurrent distractors. We find that monolithic architectures maximize raw efficiency and fine-grained accuracy (Dependency Awareness 5.95 vs. 5.67; Parameter Accuracy 8.08 vs. 7.12), while role-separated architectures offer superior scaling stability (1.38x vs. 3.24x latency degradation). These findings reveal a "context-reasoning trade-off" that challenges the assumption that modular agents are universally superior.
Anonymization: This submission has been anonymized for double-blind review via the removal of identifying information such as names, affiliations, and identifying URLs.
Funding: Yes, the presenting author of this submission falls under ICLR’s funding aims, and funding would significantly impact their ability to attend the workshop in person.
Submission Number: 43
Loading