Routing-Guided Learned Product Quantization for Graph-Based Approximate Nearest Neighbor Search

Published: 2024, Last Modified: 15 Jan 2026ICDE 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Given a vector dataset $\mathcal{X}$, a query vector $\vec{x}_{q}$, graph-based Approximate Nearest Neighbor Search (ANNS) aims to build a proximity graph (PG) as an index of $\mathcal{X}$ and approximately return vectors with minimum distances to $\vec{x}_{q}$ by searching over the PG index. It has been widely recognized that graph-based ANNS is effective and efficient, however, it suffers from the large-scale $\mathcal{X}$ because of the entire PG is too large to fit into the memory. To solve this, Product Quantization (PQ) integrated graph-based ANNS is proposed to reduce the memory usage, by replacing a large PG with original vectors by the one with smaller compact codes of quantized vectors. Existing PQ methods do not consider the important routing features of PG, thus resulting in low-quality quantized vectors that significantly affect the ANNS's effectiveness. In this paper, we present an end-to-end Routing-guided learned Product Quantization (RPQ) for graph-based ANNS, which easily can be adaptive to existing popular PGs. Specifically, RPQ consists of (1) a differentiable quantizer used to make the standard discrete PQ differentiable to suit for back-propagation of end-to-end learning, (2) a sampling-based feature extractor used to extract neighborhood and routing features of a PG by using the quantized vectors, and (3) a multi-feature joint training module with two types of feature-aware losses to continuously optimize the differentiable quantizer. As a result, the inherent features of a specific PG would be embedded into the learned PQ, thus generating high-quality quantized vectors that facilitate the graph-based ANNS's effectiveness and efficiency. Moreover, we integrate our RPQ with the state-of-the-art DiskANN and existing PGs to improve their performance. Comprehensive experimental studies on real-world large-scale datasets (scale from 1M to 1B) demonstrate RPQ's superiority.
Loading