๐ค AI Summary
In Retrieval-Augmented Generation (RAG), disk-based vector retrieval suffers from non-uniform cluster access patterns induced by query semantic similarity, while existing approaches neglect cross-query locality effectsโleading to low cache efficiency and high tail latency. This paper proposes a context-aware dynamic query grouping mechanism: it clusters semantically similar queries based on shared cluster access patterns to enable fine-grained scheduling; and introduces opportunistic cluster-level asynchronous prefetching coupled with I/O-aware cache management. Its core innovation lies in the first joint modeling of query semantic clustering and cluster-access locality to co-optimize cache locality and I/O efficiency. Experiments show up to 51.55% reduction in 99th-percentile tail latency, consistently higher cache hit rates than baselines, and significant improvements in retrieval throughput and response stability.
๐ Abstract
Modern embedding models capture both semantic and syntactic structures of queries, often mapping different queries to similar regions in vector space. This results in non-uniform cluster access patterns in disk-based vector search systems, particularly in Retrieval Augmented Generation (RAG) framework. While existing approaches optimize individual queries, they overlook the impact of cluster access patterns, failing to account for the locality effects of queries that access similar clusters. This oversight reduces cache efficiency and increases search latency due to excessive disk I/O. To address this, we introduce CaGR-RAG, a context-aware query grouping mechanism that organizes queries based on shared cluster access patterns. Additionally, it incorporates opportunistic cluster prefetching to minimize cache misses during transitions between query groups, further optimizing retrieval performance. Experimental results show that CaGR-RAG reduces 99th percentile tail latency by up to 51.55% while consistently maintaining a higher cache hit ratio than the baseline.