Don't Reinvent the Wheel: Efficient Instruction-Following Text Embedding based on Guided Space Transformation

πŸ“… 2025-05-30
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the high computational cost of re-encoding entire corpora for each new instruction in instruction-aware text embedding, this paper proposes GSTransformβ€”a lightweight Guided Space Transformation framework. Its core innovation lies in the first explicit disentanglement of instruction-invariant representations from instruction-relevant information within pre-trained embeddings. Leveraging learnable linear or nonlinear spatial transformation modules, GSTransform dynamically redirects existing embeddings to align with user-specified instructions using only a few instruction-labeled examples. By jointly optimizing instruction semantic alignment loss and few-shot supervision, it eliminates the need for full re-encoding. Evaluated across nine real-world datasets and three downstream task categories, GSTransform consistently outperforms state-of-the-art methods, achieving 6×–300Γ— faster inference while maintaining or improving embedding quality.

Technology Category

Application Category

πŸ“ Abstract
In this work, we investigate an important task named instruction-following text embedding, which generates dynamic text embeddings that adapt to user instructions, highlighting specific attributes of text. Despite recent advancements, existing approaches suffer from significant computational overhead, as they require re-encoding the entire corpus for each new instruction. To address this challenge, we propose GSTransform, a novel instruction-following text embedding framework based on Guided Space Transformation. Our key observation is that instruction-relevant information is inherently encoded in generic embeddings but remains underutilized. Instead of repeatedly encoding the corpus for each instruction, GSTransform is a lightweight transformation mechanism that adapts pre-computed embeddings in real time to align with user instructions, guided by a small amount of text data with instruction-focused label annotation. We conduct extensive experiments on three instruction-awareness downstream tasks across nine real-world datasets, demonstrating that GSTransform improves instruction-following text embedding quality over state-of-the-art methods while achieving dramatic speedups of 6~300x in real-time processing on large-scale datasets. The source code is available at https://github.com/YingchaojieFeng/GSTransform.
Problem

Research questions and friction points this paper is trying to address.

Dynamic text embeddings adapting to user instructions
Reducing computational overhead in instruction-following tasks
Real-time transformation of pre-computed embeddings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Lightweight transformation for pre-computed embeddings
Real-time adaptation to user instructions
Guided by instruction-focused label annotations
πŸ”Ž Similar Papers
No similar papers found.
Yingchaojie Feng
Yingchaojie Feng
Zhejiang University
Visual AnalyticsNatural Language ProcessingHuman Computer Interaction
Y
Yiqun Sun
School of Computing, National University of Singapore
Y
Yandong Sun
School of Computing, National University of Singapore
Minfeng Zhu
Minfeng Zhu
Zhejiang University
VisualisationMath
Q
Qiang Huang
School of Intelligence Science and Engineering, Harbin Institute of Technology (Shenzhen)
A
Anthony K. H. Tung
School of Computing, National University of Singapore
W
Wei Chen
State Key Lab of CAD&CG, Zhejiang University