Incorporating Discarded Candidate Tokens of Large Language Models for Efficient Query Expansion

ACL ARR 2025 February Submission7461 Authors

16 Feb 2025 (modified: 09 May 2025)ACL ARR 2025 February SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract:

In the era of Large Language Models (LLMs), efficient retrieval is crucial for integration with modern retrieval-augmented LLM systems, making sparse retrieval modules a popular choice due to their efficiency and robustness in low-resource settings. To enhance sparse retrieval performance, LLM-based Query Expansion (QE) has emerged as a solution to bridge the lexical gap between queries and documents. However, existing QE methods face a fundamental trade-off between efficiency and effectiveness, driven by the length of generated tokens. To address this, we propose Discarded candidate Tokens Query Expansion (DTQE), a novel query expansion method that leverages conventionally unselected tokens from the LLM’s decoding process by indexing them separately. Experimental results demonstrate that DTQE maintains high efficiency compared to more resource-intensive baselines while significantly outperforming keyword-based expansion ones.

Paper Type: Short
Research Area: Information Retrieval and Text Mining
Research Area Keywords: Information retrieval, Query expansion, Large Language Model
Contribution Types: NLP engineering experiment
Languages Studied: English
Submission Number: 7461
Loading