ReaSeq: Unleashing World Knowledge via Reasoning for Sequential Modeling

πŸ“… 2025-12-24
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Industrial recommendation systems face two key challenges: (1) impoverished ID-based representations, leading to fragile interest modeling in sparse scenarios, and (2) log boundary limitations, preventing models from capturing off-platform user behavior. To address these, we propose a world-knowledge-enhanced sequential recommendation paradigm featuring a novel explicit-implicit dual-path reasoning mechanism: explicitly distilling structured item semantics via multi-agent Chain-of-Thought reasoning, and implicitly inferring cross-domain latent behaviors using a diffusion-based large language model (Diffusion LLM). Our method integrates semantically aligned embeddings with industrial real-time ranking systems. Deployed on Taobao’s primary search engine, it achieves >6.0% improvements in IPV and CTR, and 2.9% and 2.5% gains in order volume and GMV, respectively. This work marks the first systematic integration and large-scale industrial deployment of world knowledge in sequential recommendation.

Technology Category

Application Category

πŸ“ Abstract
Industrial recommender systems face two fundamental limitations under the log-driven paradigm: (1) knowledge poverty in ID-based item representations that causes brittle interest modeling under data sparsity, and (2) systemic blindness to beyond-log user interests that constrains model performance within platform boundaries. These limitations stem from an over-reliance on shallow interaction statistics and close-looped feedback while neglecting the rich world knowledge about product semantics and cross-domain behavioral patterns that Large Language Models have learned from vast corpora. To address these challenges, we introduce ReaSeq, a reasoning-enhanced framework that leverages world knowledge in Large Language Models to address both limitations through explicit and implicit reasoning. Specifically, ReaSeq employs explicit Chain-of-Thought reasoning via multi-agent collaboration to distill structured product knowledge into semantically enriched item representations, and latent reasoning via Diffusion Large Language Models to infer plausible beyond-log behaviors. Deployed on Taobao's ranking system serving hundreds of millions of users, ReaSeq achieves substantial gains: >6.0% in IPV and CTR, >2.9% in Orders, and >2.5% in GMV, validating the effectiveness of world-knowledge-enhanced reasoning over purely log-driven approaches.
Problem

Research questions and friction points this paper is trying to address.

Enhances item representations with world knowledge to combat data sparsity
Infers beyond-log user interests using reasoning to break platform boundaries
Leverages LLM reasoning to overcome limitations of shallow interaction statistics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages world knowledge from Large Language Models for reasoning
Uses explicit Chain-of-Thought reasoning to enrich item representations
Employs latent reasoning with Diffusion LLMs to infer beyond-log behaviors
πŸ”Ž Similar Papers
No similar papers found.
C
Chuan Wang
G
Gaoming Yang
H
Han Wu
Jiakai Tang
Jiakai Tang
Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China
Recommender SystemsMulti-Agent Systems
J
Jiahao Yu
J
Jian Wu
J
Jianwu Hu
J
Junjun Zheng
S
Shuwen Xiao
Y
Yeqiu Yang
Y
Yuning Jiang
A
Ahjol Nurlanbek
B
Binbin Cao
B
Bo Zheng
F
Fangmei Zhu
G
Gaoming Zhou
H
Huimin Yi
H
Huiping Chu
J
Jin Huang
J
Jinzhe Shan
K
Kenan Cui
L
Longbin Li
S
Silu Zhou
W
Wen Chen
X
Xia Ming