RankSteer: Activation Steering for Pointwise LLM Ranking

📅 2026-02-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the sensitivity of large language models (LLMs) to prompt phrasing in zero-shot pointwise ranking, particularly their reliance on role-playing instructions, which leads to unstable ranking behavior. The authors propose a post-hoc activation manipulation framework that requires no modification of model weights and, for the first time, disentangles ranking behavior into three interpretable and intervenable representation directions: decision, evidence, and role. By applying geometric projection to calibrate LLM outputs along these directions, the method significantly enhances ranking quality using only a few anchor queries. Experiments on TREC DL 2020 and multiple BEIR benchmarks demonstrate substantial improvements, revealing latent yet underutilized ranking capabilities inherent in current pointwise LLM rankers.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have recently shown strong performance as zero-shot rankers, yet their effectiveness is highly sensitive to prompt formulation, particularly role-play instructions. Prior analyses suggest that role-related signals are encoded along activation channels that are largely separate from query-document representations, raising the possibility of steering ranking behavior directly at the activation level rather than through brittle prompt engineering. In this work, we propose RankSteer, a post-hoc activation steering framework for zero-shot pointwise LLM ranking. We characterize ranking behavior through three disentangled and steerable directions in representation space: a \textbf{decision direction} that maps hidden states to relevance scores, an \textbf{evidence direction} that captures relevance signals not directly exploited by the decision head, and a \textbf{role direction} that modulates model behavior without injecting relevance information. Using projection-based interventions at inference time, RankSteer jointly controls these directions to calibrate ranking behavior without modifying model weights or introducing explicit cross-document comparisons. Experiments on TREC DL 20 and multiple BEIR benchmarks show that RankSteer consistently improves ranking quality using only a small number of anchor queries, demonstrating that substantial ranking capacity remains under-utilized in pointwise LLM rankers. We further provide a geometric analysis revealing that steering improves ranking by stabilizing ranking geometry and reducing dispersion, offering new insight into how LLMs internally represent and calibrate relevance judgments.
Problem

Research questions and friction points this paper is trying to address.

LLM ranking
prompt sensitivity
activation steering
zero-shot ranking
role-play instructions
Innovation

Methods, ideas, or system contributions that make the work stand out.

activation steering
zero-shot ranking
representation disentanglement
LLM ranking
post-hoc intervention
🔎 Similar Papers
No similar papers found.