🤖 AI Summary
Existing LLM routing systems rely solely on input-query classification, overlooking implicit user intent and contextual nuances that only emerge during response generation—leading to suboptimal routing decisions for complex or ambiguous queries. To address this, we propose Lookahead, a novel routing framework featuring a “lookahead” mechanism that predicts the latent-space representation of the target model’s output to anticipate response semantics, thereby enabling more accurate model selection. Lookahead instantiates two lightweight routers—one built upon causal language models and the other on masked language models—both capable of efficient routing without full-response generation. Evaluated across seven public benchmarks, Lookahead achieves an average 7.7% improvement over state-of-the-art methods, with substantial gains in instruction following, mathematical reasoning, and code generation tasks.
📝 Abstract
Large language model (LLM) routers improve the efficiency of multi-model systems by directing each query to the most appropriate model while leveraging the diverse strengths of heterogeneous LLMs. Most existing approaches frame routing as a classification problem based solely on the input query. While this reduces overhead by avoiding inference across all models, it overlooks valuable information that could be gleaned from potential outputs and fails to capture implicit intent or contextual nuances that often emerge only during response generation. These limitations can result in suboptimal routing decisions, particularly for complex or ambiguous queries that require deeper semantic understanding. To address this challenge, we propose Lookahead, a routing framework that "foresees" potential model outputs by predicting their latent representations and uses these predictions to guide model selection, thus enabling more informed routing without full inference. Within this framework, we implement two approaches based on causal and masked language models. Empirical evaluations across seven public benchmarks - spanning instruction following, mathematical reasoning, and code generation - show that Lookahead consistently outperforms existing routing baselines, achieving an average performance gain of 7.7% over the state-of-the-art. Our code is available at https://github.com/huangcb01/lookahead-routing.