OpenRAG: Optimizing RAG End-to-End via In-Context Retrieval Learning

Published: 06 Mar 2025, Last Modified: 22 Mar 2025ICLR 2025 FM-Wild WorkshopEveryoneRevisionsBibTeXCC BY 4.0
Keywords: retrieval-augmented generation, RAG, LLM
TL;DR: A RAG framework that tunes a retriever to capture end-to-end relevance for wild tasks.
Abstract: In this paper, we analyze and empirically demonstrate that the relevance learned for traditional information retrieval scenarios may not consistently apply to retrieval-augmented generation (RAG) in wild environments. To bridge this gap, we introduce OpenRAG, a RAG framework that is optimized end-to-end by tuning the retriever to capture in-context, open-ended relevance, enabling adaptation to the diverse and evolving needs. Extensive experiments across a wide range of tasks demonstrate that OpenRAG, by tuning a retriever end-to-end, leads to a consistent improvement of 4.0% over the original retriever, consistently outperforming existing state-of-the-art retrievers by 2.1%. Additionally, our results show that for certain tasks, a 0.2B retriever tuned end-to-end can achieve improvements surpassing those of RAG-oriented or instruction-tuned 8B LLMs, underscoring the cost-effectiveness of our approach for improving RAG systems.
Submission Number: 12
Loading