🤖 AI Summary
In e-commerce search, queries containing superlative terms (e.g., “best”, “most popular”) imply implicit multi-attribute comparisons, demanding both deep semantic understanding and low-latency inference. To address this, we propose a prompt-enhanced two-stage re-ranking framework. First, a large language model (LLM) parses the user’s latent intent and decomposes the original query into structured attribute-value prompts. Second, these prompts are injected into a lightweight ranking model, enabling effective transfer of high-fidelity semantic knowledge while preserving real-time performance. This design circumvents the prohibitive latency of direct LLM-based re-ranking. Experiments demonstrate substantial improvements: +10.9 MAP and +5.9 MRR over strong baselines, with no degradation in deployment efficiency. Our key contributions are (i) an intent-driven structured prompt generation mechanism and (ii) a co-optimization strategy that tightly couples prompt engineering with lightweight model adaptation.
📝 Abstract
Search queries with superlatives (e.g., best, most popular) require comparing candidates across multiple dimensions, demanding linguistic understanding and domain knowledge. We show that LLMs can uncover latent intent behind these expressions in e-commerce queries through a framework that extracts structured interpretations or hints. Our approach decomposes queries into attribute-value hints generated concurrently with retrieval, enabling efficient integration into the ranking pipeline. Our method improves search performanc eby 10.9 points in MAP and ranking by 5.9 points in MRR over baselines. Since direct LLM-based reranking faces prohibitive latency, we develop an efficient approach transferring superlative interpretations to lightweight models. Our findings provide insights into how superlative semantics can be represented and transferred between models, advancing linguistic interpretation in retrieval systems while addressing practical deployment constraints.