Scalable Distributed Vector Search via Accuracy Preserving Index Construction

📅 2025-12-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Scalable distributed approximate nearest neighbor (ANN) search over billion-scale vector datasets demands careful trade-offs among accuracy, latency, and throughput. This paper proposes a scalable multi-level distributed indexing framework: first, a load-aware fine-grained partitioning strategy is designed to minimize cross-node read overhead; second, a precision-preserving recursive hierarchical index structure is introduced to ensure predictable search cost and stable recall. Evaluated on a 46-node cluster, the system supports 8 billion vectors with throughput up to 9.64× higher than the current state-of-the-art—without any accuracy degradation—and demonstrates superior linear scalability. The core innovation lies in a synergistic co-design mechanism that jointly optimizes partitioning efficiency and indexing accuracy, enabling both high throughput and guaranteed retrieval quality at billion-scale.

Technology Category

Application Category

📝 Abstract
Scaling Approximate Nearest Neighbor Search (ANNS) to billions of vectors requires distributed indexes that balance accuracy, latency, and throughput. Yet existing index designs struggle with this tradeoff. This paper presents SPIRE, a scalable vector index based on two design decisions. First, it identifies a balanced partition granularity that avoids read-cost explosion. Second, it introduces an accuracy-preserving recursive construction that builds a multi-level index with predictable search cost and stable accuracy. In experiments with up to 8 billion vectors across 46 nodes, SPIRE achieves high scalability and up to 9.64X higher throughput than state-of-the-art systems.
Problem

Research questions and friction points this paper is trying to address.

Designs scalable distributed vector indexes for billion-scale ANNS
Balances accuracy, latency, and throughput in distributed search
Avoids read-cost explosion with balanced partition granularity
Innovation

Methods, ideas, or system contributions that make the work stand out.

Balanced partition granularity prevents read-cost explosion
Accuracy-preserving recursive construction ensures stable accuracy
Multi-level index design enables predictable search cost
🔎 Similar Papers
No similar papers found.
Y
Yuming Xu
University of Science and Technology of China
Qianxi Zhang
Qianxi Zhang
MSRA
database
Q
Qi Chen
Microsoft Research
Baotong Lu
Baotong Lu
Microsoft Research
Database SystemsMachine Learning Systems
M
Menghao Li
Microsoft
P
Philip Adams
Microsoft
M
Mingqin Li
Shopify
Z
Zengzhong Li
Microsoft AI and Research
J
Jing Liu
Microsoft Research
C
Cheng Li
University of Science and Technology of China
F
Fan Yang
Microsoft Research