Collaborative Beam Search: Enhancing LLM Reasoning via Collective Consensus

ACL ARR 2025 May Submission1771 Authors

18 May 2025 (modified: 03 Jul 2025)ACL ARR 2025 May SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Complex multi-step reasoning remains challenging for large language models (LLMs). While parallel inference-time scaling methods, such as step-level beam search, offer a promising solution, existing approaches typically depend on either domain-specific external verifiers, or self-evaluation which is brittle and prompt-sensitive. To address these issues, we propose Collaborative Beam Search (CBS), an iterative framework that harnesses the collective intelligence of multiple LLMs across both generation and verification stages. For generation, CBS leverages multiple LLMs to explore a broader search space, resulting in more diverse candidate steps. For verifications, CBS employs a perplexity-based collective consensus among these models, eliminating reliance on an external verifier or complex prompts. Between iterations, CBS leverages a dynamic quota allocation strategy that reassigns generation budget based on each model’s past consensus performance, striking a balance between candidate diversity and quality. Experimental results on six tasks across arithmetic, logical, and commonsense reasoning show that CBS outperforms single‑model scaling and multi-model ensemble baselines by over 4 percentage points in average accuracy, demonstrating its effectiveness and broad applicability.
Paper Type: Long
Research Area: NLP Applications
Research Area Keywords: inference methods
Contribution Types: NLP engineering experiment
Languages Studied: English
Keywords: inference methods
Submission Number: 1771
Loading