Prompt Tuning as User Inherent Profile Inference Machine

📅 2024-08-13
🏛️ arXiv.org
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
To address text noise in large language models (LLMs) for recommendation—stemming from unstable instruction following, modality gaps between textual and collaborative signals, and high inference latency—this paper proposes UserIP-Tuning. It reformulates prompt tuning as a causal inference task over implicit user profiles, leveraging an Expectation-Maximization (EM) algorithm to disentangle users’ intrinsic profiles from behavioral sequences. Crucially, it introduces a profile quantization codebook that maps continuous profile embeddings into discrete collaborative IDs, thereby bridging the modality gap and enabling lightweight online deployment. Evaluated on four public benchmarks, UserIP-Tuning achieves significant improvements over state-of-the-art methods: 52% reduction in inference latency, 37% lower memory footprint, enhanced robustness, cross-domain transferability, and millisecond-scale response time.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have exhibited significant promise in recommender systems by empowering user profiles with their extensive world knowledge and superior reasoning capabilities. However, LLMs face challenges like unstable instruction compliance, modality gaps, and high inference latency, leading to textual noise and limiting their effectiveness in recommender systems. To address these challenges, we propose UserIP-Tuning, which uses prompt-tuning to infer user profiles. It integrates the causal relationship between user profiles and behavior sequences into LLMs' prompts. And employs expectation maximization to infer the embedded latent profile, minimizing textual noise by fixing the prompt template. Furthermore, A profile quantization codebook bridges the modality gap by categorizing profile embeddings into collaborative IDs, which are pre-stored for online deployment. This improves time efficiency and reduces memory usage. Experiments on four public datasets show that UserIP-Tuning outperforms state-of-the-art recommendation algorithms. Additional tests and case studies confirm its effectiveness, robustness, and transferability.
Problem

Research questions and friction points this paper is trying to address.

Inferring user profiles from behavior sequences using prompt-tuning
Reducing textual noise and modality gaps in LLM-based recommendations
Improving inference efficiency for real-world recommendation systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses prompt-tuning to infer user profiles
Employs Expectation Maximization to minimize textual noise
Applies profile quantization codebook to bridge modality gaps
🔎 Similar Papers
No similar papers found.
Y
Yusheng Lu
Tongji University, Shanghai, China; City University of Hong Kong, Hong Kong, China
Zhaocheng Du
Zhaocheng Du
Huawei Noah Ark's Lab
Machine LearningRecommendation System
X
Xiangyang Li
Huawei Noah’s Ark Lab, Shenzhen, China
X
Xiangyu Zhao
City University of Hong Kong, Hong Kong, China
Weiwen Liu
Weiwen Liu
Associate Professor, Shanghai Jiao Tong University
large language modelsAI agentsrecommender systems
Y
Yichao Wang
Huawei Noah’s Ark Lab, Shenzhen, China
Huifeng Guo
Huifeng Guo
Huawei, Harbin Institute of Technology
Recommender SystemDeep LearningData Mining.
R
Ruiming Tang
Huawei Noah’s Ark Lab, Shenzhen, China
Zhenhua Dong
Zhenhua Dong
Noah's ark lab, Huawei Technologies Co., Ltd.
Recommender systemcausal inferencecountrfactual learningtrustworthy AImachine learning
Y
Yongrui Duan
Tongji University, Shanghai, China