🤖 AI Summary
Scalable distributed approximate nearest neighbor (ANN) search over billion-scale vector datasets demands careful trade-offs among accuracy, latency, and throughput. This paper proposes a scalable multi-level distributed indexing framework: first, a load-aware fine-grained partitioning strategy is designed to minimize cross-node read overhead; second, a precision-preserving recursive hierarchical index structure is introduced to ensure predictable search cost and stable recall. Evaluated on a 46-node cluster, the system supports 8 billion vectors with throughput up to 9.64× higher than the current state-of-the-art—without any accuracy degradation—and demonstrates superior linear scalability. The core innovation lies in a synergistic co-design mechanism that jointly optimizes partitioning efficiency and indexing accuracy, enabling both high throughput and guaranteed retrieval quality at billion-scale.
📝 Abstract
Scaling Approximate Nearest Neighbor Search (ANNS) to billions of vectors requires distributed indexes that balance accuracy, latency, and throughput. Yet existing index designs struggle with this tradeoff. This paper presents SPIRE, a scalable vector index based on two design decisions. First, it identifies a balanced partition granularity that avoids read-cost explosion. Second, it introduces an accuracy-preserving recursive construction that builds a multi-level index with predictable search cost and stable accuracy. In experiments with up to 8 billion vectors across 46 nodes, SPIRE achieves high scalability and up to 9.64X higher throughput than state-of-the-art systems.