Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models

📅 2024-12-05
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address severe cold-start issues and over-reliance on collaborative signals in sequential recommendation (SR), as well as high inference latency, insufficient distribution alignment, and catastrophic forgetting in large language model (LLM) deployment, this paper proposes PAD—a Pretraining–Alignment–Decoupling framework. PAD employs dual-path (collaborative/textual) pretraining to enhance semantic understanding; introduces a recommendation-task-anchored multi-kernel maximum mean discrepancy (MMD) alignment loss for fine-grained distribution matching; and designs a frequency-aware three-expert decoupled fine-tuning architecture—comprising an alignment expert and two modality-specific experts—to jointly preserve cross-modal consistency and modality specificity. Evaluated on three public benchmarks, PAD significantly improves overall recommendation performance, especially for cold items. It is compatible with diverse SR backbone models, demonstrating strong generalizability and reproducibility.

Technology Category

Application Category

📝 Abstract
Sequential Recommendation (SR) aims to leverage the sequential patterns in users' historical interactions to accurately track their preferences. However, the primary reliance of existing SR methods on collaborative data results in challenges such as the cold-start problem and sub-optimal performance. Concurrently, despite the proven effectiveness of large language models (LLMs), their integration into commercial recommender systems is impeded by issues such as high inference latency, incomplete capture of all distribution statistics, and catastrophic forgetting. To address these issues, we introduce a novel Pre-train, Align, and Disentangle (PAD) framework to enhance SR models with LLMs. In particular, we initially pre-train both the SR and LLM models to obtain collaborative and textual embeddings. Subsequently, we propose a characteristic recommendation-anchored alignment loss using multi-kernel maximum mean discrepancy with Gaussian kernels. Lastly, a triple-experts architecture, comprising aligned and modality-specific experts with disentangled embeddings, is fine-tuned in a frequency-aware manner. Experimental results on three public datasets validate the efficacy of PAD, indicating substantial enhancements and compatibility with various SR backbone models, particularly for cold items. The code and datasets are accessible for reproduction at https://github.com/Applied-Machine-Learning-Lab/PAD.
Problem

Research questions and friction points this paper is trying to address.

Address cold-start and performance issues in Sequential Recommendation
Mitigate LLM integration challenges like latency and data capture
Enhance SR models using pre-trained collaborative and textual embeddings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Pre-train SR and LLM models for embeddings
Align embeddings using multi-kernel discrepancy
Fine-tune triple-experts with disentangled embeddings
🔎 Similar Papers
2024-02-02ACM Transactions on Recommender SystemsCitations: 1
Y
Yuhao Wang
City University of Hong Kong, Hong Kong, China
Junwei Pan
Junwei Pan
Tencent, Yahoo Research
Computational AdvertisingRecommendation SystemDeep Learning
X
Xiangyu Zhao
City University of Hong Kong, Hong Kong, China
Pengyue Jia
Pengyue Jia
PhD candidate of Data Science, City University of Hong Kong
Information RetrievalLarge Language ModelsGeoAI
W
Wanyu Wang
City University of Hong Kong, Hong Kong, China
Y
Yuan Wang
Tencent Inc., Shenzhen, China
Y
Yue Liu
Tencent Inc., Shenzhen, China
D
Dapeng Liu
Tencent Inc., Shenzhen, China
J
Jie Jiang
Tencent Inc., Shenzhen, China