Boosted Dense RetrieverDownload PDF

Anonymous

08 Mar 2022 (modified: 05 May 2023)NAACL 2022 Conference Blind SubmissionReaders: Everyone
Paper Link: https://openreview.net/forum?id=tuGA1H84mfp
Paper Type: Long paper (up to eight pages of content + unlimited references and appendices)
Abstract: We propose DrBoost, a dense retrieval ensemble inspired by boosting. DrBoost is trained in stages: each component model is learned sequentially and specialized by focusing only on retrieval mistakes made by the current ensemble. The final representation is the concatenation of the output vectors of all the component models, making it a drop-in replacement for standard dense retrievers at test time. DrBoost enjoys several advantages compared to standard dense retrieval models. It produces representations which are 4x more compact, while delivering comparable retrieval results. It also performs surprisingly well under approximate search with coarse quantization, reducing latency and bandwidth needs by another 4x. In practice, this can make the difference between serving indices from disk versus from memory, paving the way for much cheaper deployments.
Presentation Mode: This paper will be presented in person in Seattle
Virtual Presentation Timezone: UTC-8
Copyright Consent Signature (type Name Or NA If Not Transferrable): Barlas Oguz
Copyright Consent Name And Address: Meta, 1 Facebook Way, Menlo Park CA, USA
0 Replies

Loading