Instruction-aware User Embedding via Synergistic Language and Representation Modeling

📅 2025-10-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing user representation methods suffer from poor cross-domain generalization and vulnerability to behavioral noise. To address these limitations, we propose the Instruction-aware User Embedding Model (IUEM), a foundation model for user embeddings that leverages large language models (LLMs) to generate representations with both strong generalizability and instruction-following capability. Methodologically, IUEM employs a multi-encoder architecture augmented with lightweight adapters to jointly encode linguistic and embedding spaces; it further introduces a contrastive–autoregressive joint training framework, enabling instruction-guided behavioral denoising and robust cross-domain modeling. The model is co-optimized on the UserQA dataset. Extensive experiments demonstrate that IUEM significantly outperforms state-of-the-art methods across real-world tasks—including user prediction, marketing, and recommendation—validating its superior cross-domain adaptability and robustness to behavioral noise.

Technology Category

Application Category

📝 Abstract
User representation modeling has become increasingly crucial for personalized applications, yet existing approaches struggle with generalizability across domains and sensitivity to noisy behavioral signals. We present InstructUE, an instruction-aware user embedding foundation model that leverages large language models (LLMs) to generate general and instruction-aware user representations. InstructUE introduces a multi-encoder architecture with a lightweight adapter that efficiently processes heterogeneous data from six different sources while preserving their structural characteristics. Additionally, it proposes a novel contrastive-autoregressive training framework that bridges language and representation spaces through a curated UserQA dataset. The contrastive-autoregressive training framework simultaneously leverages autoregressive learning to capture domain knowledge in language space and contrastive learning to align user-text embeddings in representation space, thereby enhancing the instruction-awareness and noise-robustness of user embeddings. Through extensive experiments on real-world applications, we demonstrate that InstructUE significantly outperforms existing methods across multiple domains including user prediction, marketing, and recommendation scenarios. Our results show that instruction-aware user modeling can effectively achieve instruction-guided denoising of user information in specific scenarios, paving the way for more generalizable and robust user representation learning.
Problem

Research questions and friction points this paper is trying to address.

Creating generalizable user embeddings across domains with noisy behavioral data
Aligning language and representation spaces for instruction-aware modeling
Developing noise-robust user representations through multi-source heterogeneous data processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages LLMs for instruction-aware user embeddings
Uses multi-encoder architecture with lightweight adapter
Proposes contrastive-autoregressive training framework
🔎 Similar Papers
No similar papers found.
Z
Ziyi Gao
Ant Group
Y
Yike Xu
Ant Group
J
Jiahao Yuan
Ant Group
B
Baokun Wang
Ant Group
J
Jinyong Wen
Ant Group
Xiaotong Lin
Xiaotong Lin
Sun Yat-sen University
computer vision
Y
Yun Liu
Ant Group
Xing Fu
Xing Fu
Ant Group
Y
Yu Cheng
Ant Group
Y
Yongchao Liu
Ant Group
W
Weiqiang Wang
Ant Group
Zhongle Xie
Zhongle Xie
Zhejiang University
AI4DBML SystemsDB SystemsOLAP