🤖 AI Summary
To address the challenge of balancing cost and performance in large-scale deployment of large language models (LLMs), this paper proposes an adaptive multi-response routing framework. Departing from conventional single-response routing, the method dynamically selects models based on query difficulty and generates multiple candidate responses from small, low-cost models; a quality-aware filtering mechanism then selects the optimal response. A tunable quality threshold enables fine-grained trade-offs between accuracy and inference cost. The core innovation lies in tightly coupling multi-response sampling with model routing—reducing expensive LLM invocations while preserving response quality. Experiments on real-world datasets demonstrate that the framework achieves up to a 60% reduction in inference cost compared to baseline methods, with less than a 1% degradation in task performance.
📝 Abstract
Large language models (LLMs) are powerful tools but are often expensive to deploy at scale. LLM query routing mitigates this by dynamically assigning queries to models of varying cost and quality to obtain a desired trade-off. Prior query routing approaches generate only one response from the selected model and a single response from a small (inexpensive) model was often not good enough to beat a response from a large (expensive) model due to which they end up overusing the large model and missing out on potential cost savings. However, it is well known that for small models, generating multiple responses and selecting the best can enhance quality while remaining cheaper than a single large-model response. We leverage this idea to propose BEST-Route, a novel routing framework that chooses a model and the number of responses to sample from it based on query difficulty and the quality thresholds. Experiments on real-world datasets demonstrate that our method reduces costs by up to 60% with less than 1% performance drop.