Training Large Recommendation Models via Graph-Language Token Alignment

📅 2025-02-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional recommender systems struggle to jointly model graph-structured interaction signals and textual semantics, while direct invocation of large language models (LLMs) suffers from generation ambiguity and low inference efficiency. To address these challenges, we propose the Graph–Language Token Alignment (GLTA) and Graph–Language Logical Matching (GLLM) framework, which maps user/item node embeddings into the LLM’s semantic space, enabling end-to-end differentiable alignment between graph structure and text semantics. Our approach integrates graph neural networks, LLM embedding alignment, contrastive learning, and logits-level supervised matching—bypassing autoregressive text generation entirely. This design significantly improves recommendation accuracy and generalization. Extensive experiments on three benchmark datasets demonstrate consistent superiority over state-of-the-art methods. Ablation studies validate the effectiveness and necessity of each component.

Technology Category

Application Category

📝 Abstract
Recommender systems (RS) have become essential tools for helping users efficiently navigate the overwhelming amount of information on e-commerce and social platforms. However, traditional RS relying on Collaborative Filtering (CF) struggles to integrate the rich semantic information from textual data. Meanwhile, large language models (LLMs) have shown promising results in natural language processing, but directly using LLMs for recommendation introduces challenges, such as ambiguity in generating item predictions and inefficiencies in scalability. In this paper, we propose a novel framework to train Large Recommendation models via Graph-Language Token Alignment. By aligning item and user nodes from the interaction graph with pretrained LLM tokens, GLTA effectively leverages the reasoning abilities of LLMs. Furthermore, we introduce Graph-Language Logits Matching (GLLM) to optimize token alignment for end-to-end item prediction, eliminating ambiguity in the free-form text as recommendation results. Extensive experiments on three benchmark datasets demonstrate the effectiveness of GLTA, with ablation studies validating each component.
Problem

Research questions and friction points this paper is trying to address.

Integrate textual semantic in recommendation
Overcome ambiguity in LLMs for item prediction
Enhance scalability in recommendation systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Graph-Language Token Alignment
Graph-Language Logits Matching
Pretrained LLM tokens integration
🔎 Similar Papers
No similar papers found.
Mingdai Yang
Mingdai Yang
University of Illinois at Chicago
Recommender SystemsDeep Learning
Z
Zhiwei Liu
Salesforce AI Research, Palo Alto, USA
Liangwei Yang
Liangwei Yang
Salesforce Research
Network ScienceRecommender SystemEfficient Modeling
X
Xiaolong Liu
University of Illinois at Chicago, Chicago, USA
C
Chen Wang
University of Illinois at Chicago, Chicago, USA
H
Hao Peng
Beihang University, & Hangzhou Innovation Institute of BUAA, Beijing & Hangzhou, China
Philip S. Yu
Philip S. Yu
Professor of Computer Science, University of Illinons at Chicago
Data miningDatabasePrivacy