🤖 AI Summary
This work addresses the high cost of tuning the α-reachability parameter in large-scale graph-based vector retrieval, which conventionally requires rebuilding the entire index. The authors propose RP-Tuning, the first method enabling dynamic adjustment of α-reachable graphs without reconstruction. By repurposing DiskANN’s pruning mechanism as a post-processing step, RP-Tuning efficiently modifies graph connectivity while preserving theoretical reachability guarantees. The approach is applicable to both general metric spaces and Euclidean space, demonstrating significant practical benefits: across four public datasets, it accelerates DiskANN’s tuning process by up to 43× with negligible additional overhead.
📝 Abstract
Vector similarity search is an essential primitive in modern AI and ML applications. Most vector databases adopt graph-based approximate nearest neighbor (ANN) search algorithms, such as DiskANN (Subramanya et al., 2019), which have demonstrated state-of-the-art empirical performance. DiskANN's graph construction is governed by a reachability parameter $\alpha$, which gives a trade-off between construction time, query time, and accuracy. However, adaptively tuning this trade-off typically requires rebuilding the index for different $\alpha$ values, which is prohibitive at scale. In this work, we propose RP-Tuning, an efficient post-hoc routine, based on DiskANN's pruning step, to adjust the $\alpha$ parameter without reconstructing the full index. Within the $\alpha$-reachability framework of prior theoretical works (Indyk and Xu, 2023; Gollapudi et al., 2025), we prove that pruning an initially $\alpha$-reachable graph with RP-Tuning preserves worst-case reachability guarantees in general metrics and improved guarantees in Euclidean metrics. Empirically, we show that RP-Tuning accelerates DiskANN tuning on four public datasets by up to $43\times$ with negligible overhead.