Efficient Sequential Recommendation for Long Term User Interest Via Personalization

📅 2026-01-07
🏛️ arXiv.org
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the high computational complexity and inefficiency of Transformer-based sequential recommendation models when handling long user interaction histories. To this end, the authors propose a general-purpose personalized compression mechanism that leverages learnable personalized tokens to efficiently compress extensive historical interactions, which are then fused with recent user behaviors to generate recommendations. The approach is compatible with mainstream Transformer architectures such as HSTU and HLLM. Extensive experiments across multiple benchmark models demonstrate that the proposed method significantly reduces computational overhead while maintaining or even improving recommendation accuracy, effectively balancing efficiency and performance.

Technology Category

Application Category

📝 Abstract
Recent years have witnessed success of sequential modeling, generative recommender, and large language model for recommendation. Though the scaling law has been validated for sequential models, it showed inefficiency in computational capacity when considering real-world applications like recommendation, due to the non-linear(quadratic) increasing nature of the transformer model. To improve the efficiency of the sequential model, we introduced a novel approach to sequential recommendation that leverages personalization techniques to enhance efficiency and performance. Our method compresses long user interaction histories into learnable tokens, which are then combined with recent interactions to generate recommendations. This approach significantly reduces computational costs while maintaining high recommendation accuracy. Our method could be applied to existing transformer based recommendation models, e.g., HSTU and HLLM. Extensive experiments on multiple sequential models demonstrate its versatility and effectiveness. Source code is available at \href{https://github.com/facebookresearch/PerSRec}{https://github.com/facebookresearch/PerSRec}.
Problem

Research questions and friction points this paper is trying to address.

sequential recommendation
long-term user interest
computational efficiency
transformer model
personalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

personalized token compression
efficient sequential recommendation
long-term user interest modeling
transformer-based recommendation
computational efficiency
🔎 Similar Papers
No similar papers found.
Q
Qiang Zhang
Meta Recommendation Systems (MRS), Menlo Park, USA
Hanchao Yu
Hanchao Yu
AI at Meta
Multimodal UnderstandingComputer VisionDeep LearningMedical Image Analysis
I
Ivan Ji
Meta Recommendation Systems (MRS), Menlo Park, USA
Chen Yuan
Chen Yuan
Shanghai Jiao Tong University
coding theorysecure multiparty computation
Y
Yi Zhang
Meta Recommendation Systems (MRS), Menlo Park, USA
C
Chihuang Liu
Meta Recommendation Systems (MRS), Menlo Park, USA
X
Xiaolong Wang
Meta Recommendation Systems (MRS), Menlo Park, USA
C
Christopher E. Lambert
Meta Recommendation Systems (MRS), Menlo Park, USA
R
Ren Chen
Meta Recommendation Systems (MRS), Menlo Park, USA
C
Chen Kovacs
Meta Recommendation Systems (MRS), Menlo Park, USA
X
Xinzhu Bei
Meta Recommendation Systems (MRS), Menlo Park, USA
Renqin Cai
Renqin Cai
University of Virginia
Fairness Transparency RobustnessInformation RetrievalWeb SearchRecommender
Rui Li
Rui Li
Meta
searchdata miningmachine learningdatabaserecommendation systems
Lizhu Zhang
Lizhu Zhang
Meta
Recommendation SystemLLM
X
Xiangjun Fan
Meta Recommendation Systems (MRS), Menlo Park, USA
Q
Qunshu Zhang
Meta Recommendation Systems (MRS), Menlo Park, USA
Benyu Zhang
Benyu Zhang
Meta
Artificial IntelligencePrivacy Preserving Machine LearningCloud ComputingComputational