Dimensionality-Reduction Techniques for Approximate Nearest Neighbor Search: A Survey and Evaluation

📅 2024-03-20
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
High-dimensional vector approximate nearest neighbor search (ANNS) suffers from efficiency bottlenecks due to linear growth of distance computation cost with dimensionality—especially acute for LLM-derived semantic vectors. This work systematically evaluates six dimensionality reduction (DR) techniques—PCA, product quantization, autoencoders, contrastive learning-based DR, LSH variants, and random projection—quantifying their acceleration effects on mainstream ANNS engines (e.g., FAISS, Annoy) under a unified experimental framework. We propose two analyzable DR–search co-design architectures and theoretically derive critical pruning gain thresholds, characterizing the fundamental trade-off between dimensionality compression and retrieval accuracy degradation. Experiments across six public benchmarks show that deep DR methods achieve 1.8–3.5× speedup while maintaining >90% recall. Furthermore, we provide a data-aware guideline for optimal DR technique selection based on intrinsic data properties.

Technology Category

Application Category

📝 Abstract
Approximate Nearest Neighbor Search (ANNS) on high-dimensional vectors has become a fundamental and essential component in various machine learning tasks. Recently, with the rapid development of deep learning models and the applications of Large Language Models (LLMs), the dimensionality of the vectors keeps growing in order to accommodate a richer semantic representation. This poses a major challenge to the ANNS solutions since distance calculation cost in ANNS grows linearly with the dimensionality of vectors. To overcome this challenge, dimensionality-reduction techniques can be leveraged to accelerate the distance calculation in the search process. In this paper, we investigate six dimensionality-reduction techniques that have the potential to improve ANNS solutions, including classical algorithms such as PCA and vector quantization, as well as algorithms based on deep learning approaches. We further describe two frameworks to apply these techniques in the ANNS workflow, and theoretically analyze the time and space costs, as well as the beneficial threshold for the pruning ratio of these techniques. The surveyed techniques are evaluated on six public datasets. The analysis of the results reveals the characteristics of the different families of techniques and provides insights into the promising future research directions.
Problem

Research questions and friction points this paper is trying to address.

Accelerate ANNS with dimensionality-reduction
Evaluate six dimensionality-reduction techniques
Analyze time, space costs, and pruning ratios
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dimensionality-reduction techniques
Deep learning approaches
Theoretical cost analysis
🔎 Similar Papers
No similar papers found.
Z
Zeyu Wang
Fudan University
H
Haoran Xiong
Fudan University
Z
Zhenying He
Fudan University
P
Peng Wang
Fudan University
W
Wei Wang
Fudan University