VSAG: An Optimized Search Framework for Graph-based Approximate Nearest Neighbor Search

📅 2025-03-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing graph-based approximate nearest neighbor search (ANNS) algorithms suffer in production environments from inefficient random memory access, high computational overhead in distance computation, and sensitivity to hyperparameter tuning—requiring frequent index rebuilds. To address these limitations, we propose three key innovations: (1) a novel index-rebuild-free automated hyperparameter search mechanism; (2) a cache-friendly vector organization scheme coupled with an L3-cache-aware prefetching strategy; and (3) hardware-aware dynamic low-precision distance computation, integrating scalar quantization with AVX-512 acceleration. Evaluated on real-world datasets, our approach achieves up to 4× higher query throughput than HNSWlib while preserving full recall accuracy. The design significantly improves both system throughput and deployment efficiency—eliminating costly index reconstruction cycles and reducing memory bandwidth pressure without compromising retrieval quality.

Technology Category

Application Category

📝 Abstract
Approximate nearest neighbor search (ANNS) is a fundamental problem in vector databases and AI infrastructures. Recent graph-based ANNS algorithms have achieved high search accuracy with practical efficiency. Despite the advancements, these algorithms still face performance bottlenecks in production, due to the random memory access patterns of graph-based search and the high computational overheads of vector distance. In addition, the performance of a graph-based ANNS algorithm is highly sensitive to parameters, while selecting the optimal parameters is cost-prohibitive, e.g., manual tuning requires repeatedly re-building the index. This paper introduces VSAG, an open-source framework that aims to enhance the in production performance of graph-based ANNS algorithms. VSAG has been deployed at scale in the services of Ant Group, and it incorporates three key optimizations: (i) efficient memory access: it reduces L3 cache misses with pre-fetching and cache-friendly vector organization; (ii) automated parameter tuning: it automatically selects performance-optimal parameters without requiring index rebuilding; (iii) efficient distance computation: it leverages modern hardware, scalar quantization, and smartly switches to low-precision representation to dramatically reduce the distance computation costs. We evaluate VSAG on real-world datasets. The experimental results show that VSAG achieves the state-of-the-art performance and provides up to 4x speedup over HNSWlib (an industry-standard library) while ensuring the same accuracy.
Problem

Research questions and friction points this paper is trying to address.

Optimizes memory access to reduce L3 cache misses
Automates parameter tuning without index rebuilding
Reduces distance computation costs with hardware optimizations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reduces L3 cache misses with prefetching
Automates parameter tuning without rebuilding
Uses quantization for faster distance computation
🔎 Similar Papers
No similar papers found.
X
Xiaoyao Zhong
Ant Group, Shanghai, China
H
Haotian Li
Ant Group, Shanghai, China
Jiabao Jin
Jiabao Jin
Ant Group
Vector DataBase
M
Mingyu Yang
Ant Group, Shanghai, China
Deming Chu
Deming Chu
University of New South Wales
Graph Data Management
Xiangyu Wang
Xiangyu Wang
Professor, Curtin University
Civil EngineeringBuilding Information ModelingSmart CityAutomation and RoboticsSmart
Zhitao Shen
Zhitao Shen
Ant Group
databasedata storage
W
Wei Jia
Ant Group, Shanghai, China
G
George Gu
Intel Corporation, Shanghai, China
Y
Yi Xie
Intel Corporation, Shanghai, China
X
Xuemin Lin
Shanghai Jiaotong University, Shanghai, China
H
Heng Tao Shen
Tongji University, Shanghai, China
J
Jingkuan Song
Tongji University, Shanghai, China
P
Peng Cheng
Tongji University, Shanghai, China