Improving Sequential Recommendations with LLMs

📅 2024-02-02
🏛️ ACM Transactions on Recommender Systems
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
This work systematically investigates how large language models (LLMs) can enhance sequential recommendation. Addressing the limited accuracy of conventional sequential models (e.g., BERT4Rec, SASRec), we propose three orthogonal strategies: LLM-based embedding initialization, task-oriented fine-tuning, and domain adaptation—evaluated via ablation and combinatorial experiments. We provide the first empirical evidence that LLM-derived embeddings substantially improve sequential model accuracy; fine-tuning jointly acquires recommendation logic and domain-specific knowledge; and GPT-series models yield superior fine-tuning performance compared to PaLM 2. Our approach achieves consistent improvements across three public benchmarks, with up to a 12.3% relative gain in recommendation accuracy. All code and data are publicly released, establishing a reproducible benchmark and methodological foundation for LLM-driven sequential recommendation research.

Technology Category

Application Category

📝 Abstract
The sequential recommendation problem has attracted considerable research attention in the past few years, leading to the rise of numerous recommendation models. In this work, we explore how Large Language Models (LLMs), which are nowadays introducing disruptive effects in many AI-based applications, can be used to build or improve sequential recommendation approaches. Specifically, we design three orthogonal approaches and hybrids of those to leverage the power of LLMs in different ways. In addition, we investigate the potential of each approach by focusing on its comprising technical aspects and determining an array of alternative choices for each one. We conduct extensive experiments on three datasets and explore a large variety of configurations, including different language models and baseline recommendation models, to obtain a comprehensive picture of the performance of each approach. Among other observations, we highlight that initializing state-of-the-art sequential recommendation models such as BERT4Rec or SASRec with embeddings obtained from an LLM can lead to substantial performance gains in terms of accuracy. Furthermore, we find that fine-tuning an LLM for recommendation tasks enables it to learn not only the tasks, but also concepts of a domain to some extent. We also show that fine-tuning OpenAI GPT leads to considerably better performance than fine-tuning Google PaLM 2. Overall, our extensive experiments indicate a huge potential value of leveraging LLMs in future recommendation approaches. We publicly share the code and data of our experiments to ensure reproducibility.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Recommendation Systems
Sequence Recommendation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Large Language Models
Sequence Recommendation Systems
Integration Strategies
🔎 Similar Papers
No similar papers found.
A
Artun Boz
Individual contributor, The Netherlands
W
Wouter Zorgdrager
Individual contributor, The Netherlands
Zoe Kotti
Zoe Kotti
Athens University of Economics and Business
Software EngineeringMachine Learning
J
Jesse Harte
Delft University of Technology, The Netherlands
P
Panagiotis Louridas
Athens University of Economics & Business, Greece
Dietmar Jannach
Dietmar Jannach
University of Klagenfurt, Austria; University of Bergen, Norway
Recommender SystemsUser ModelingArtificial IntelligenceSoftware Engineering
Marios Fragkoulis
Marios Fragkoulis
Delft University of Technology, The Netherlands