STORE: Streamlining Semantic Tokenization and Generative Recommendation with A Single LLM

πŸ“… 2024-09-11
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 7
✨ Influential: 1
πŸ“„ PDF
πŸ€– AI Summary
Traditional recommender systems rely on item IDs, limiting semantic modeling and generalization to cold-start and long-tail scenarios. To address this, we propose STOREβ€”the first framework unifying semantic tokenization (text-to-semantic-token) and generative recommendation (token-to-token) within a single large language model (LLM), eliminating fragile quantization modules such as RQ-VAE. STORE employs multi-stage generative modeling: jointly optimizing token-to-token prediction, token-to-text reconstruction, and text-to-token auxiliary tasks for end-to-end semantic understanding and recommendation. By removing modular coupling, STORE achieves superior semantic expressiveness, significantly improving coverage of long-tail items and cold-start performance. It outperforms state-of-the-art methods across multiple benchmarks. The code and configurations are publicly available.

Technology Category

Application Category

πŸ“ Abstract
Traditional recommendation models often rely on unique item identifiers (IDs) to distinguish between items, which can hinder their ability to effectively leverage item content information and generalize to long-tail or cold-start items. Recently, semantic tokenization has been proposed as a promising solution that aims to tokenize each item's semantic representation into a sequence of discrete tokens. In this way, it preserves the item's semantics within these tokens and ensures that semantically similar items are represented by similar tokens. These semantic tokens have become fundamental in training generative recommendation models. However, existing generative recommendation methods typically involve multiple sub-models for embedding, quantization, and recommendation, leading to an overly complex system. In this paper, we propose to streamline the semantic tokenization and generative recommendation process with a unified framework, dubbed STORE, which leverages a single large language model (LLM) for both tasks. Specifically, we formulate semantic tokenization as a text-to-token task and generative recommendation as a token-to-token task, supplemented by a token-to-text reconstruction task and a text-to-token auxiliary task. All these tasks are framed in a generative manner and trained using a single LLM backbone. Extensive experiments have been conducted to validate the effectiveness of our STORE framework across various recommendation tasks and datasets. We will release the source code and configurations for reproducible research.
Problem

Research questions and friction points this paper is trying to address.

Overcoming limitations of unique item IDs in recommendation models
Improving semantic tokenization for generative recommendation systems
Addressing challenges in embedding extraction and training stability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-aspect semantic tokenization via item palette
Domain-specific tuning for semantic encoders
Improved embedding extraction and training stability
πŸ”Ž Similar Papers
2024-05-12International Conference on Information and Knowledge ManagementCitations: 60
Q
Qijiong Liu
The HK PolyU, Hong Kong, China
J
Jieming Zhu
Huawei Noah’s Ark Lab, Shenzhen, China
Lu Fan
Lu Fan
PolyU
graph mininglow-resource language understanding
Zhou Zhao
Zhou Zhao
Zhejiang University
Machine LearningData MiningMultimedia Computing
X
Xiao-Ming Wu
The HK PolyU, Hong Kong, China