Let It Go? Not Quite: Addressing Item Cold Start in Sequential Recommendations with Content-Based Initialization

📅 2025-07-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the cold-start problem for new items in sequential recommendation—where items lack interaction history, rendering their embeddings unlearnable—this paper proposes a content-aware embedding initialization and lightweight fine-tuning method. The approach freezes pre-trained content encoders (e.g., text or audio feature extractors) and introduces only small, learnable delta offsets for task-specific adaptation, thereby preserving semantic stability while enabling recommendation-aware customization. Integrated into standard sequential recommendation frameworks, the method is trained end-to-end. Experiments on multiple e-commerce and music datasets demonstrate significant improvements over both frozen-embedding and full-fine-tuning baselines. It effectively mitigates the new-item cold-start issue without inducing representation drift, achieving a superior trade-off between recommendation accuracy and generalization.

Technology Category

Application Category

📝 Abstract
Many sequential recommender systems suffer from the cold start problem, where items with few or no interactions cannot be effectively used by the model due to the absence of a trained embedding. Content-based approaches, which leverage item metadata, are commonly used in such scenarios. One possible way is to use embeddings derived from content features such as textual descriptions as initialization for the model embeddings. However, directly using frozen content embeddings often results in suboptimal performance, as they may not fully adapt to the recommendation task. On the other hand, fine-tuning these embeddings can degrade performance for cold-start items, as item representations may drift far from their original structure after training. We propose a novel approach to address this limitation. Instead of entirely freezing the content embeddings or fine-tuning them extensively, we introduce a small trainable delta to frozen embeddings that enables the model to adapt item representations without letting them go too far from their original semantic structure. This approach demonstrates consistent improvements across multiple datasets and modalities, including e-commerce datasets with textual descriptions and a music dataset with audio-based representation.
Problem

Research questions and friction points this paper is trying to address.

Addressing item cold start in sequential recommendations
Improving content-based initialization for item embeddings
Balancing frozen and fine-tuned embeddings for optimal performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Content-based initialization for cold-start items
Trainable delta to frozen embeddings
Balances adaptation and semantic structure preservation
🔎 Similar Papers
No similar papers found.
A
Anton Pembek
Sber AI Lab, Lomonosov Moscow State University (MSU)
A
Artem Fatkulin
Sber AI Lab, HSE University
Anton Klenitskiy
Anton Klenitskiy
Sber AI Lab
Machine learningDeep learning
Alexey Vasilev
Alexey Vasilev
Sber AI Lab; HSE University; MSU
Machine LearningData scienceRecommender SystemNLP