FuXi-$γ$: Efficient Sequential Recommendation with Exponential-Power Temporal Encoder and Diagonal-Sparse Positional Mechanism

📅 2025-12-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high computational overhead in Transformer-based sequential recommendation—caused by discontinuous memory access in temporal encoding and redundant attention computation over long sequences—this paper proposes an efficient framework for modeling long user behavior sequences. The method introduces three key innovations: (1) a tunable exponential decay time encoder, inspired by the Ebbinghaus forgetting curve, to explicitly model the dynamic decay of user preferences; (2) a diagonal sliding sparse attention mechanism leveraging Toeplitz matrix symmetry, enabling hardware-friendly local-global temporal modeling; and (3) a decoder-only architecture optimized exclusively via matrix operations. Evaluated on four real-world datasets, the approach achieves state-of-the-art accuracy while accelerating training and inference by 4.74× and 6.18×, respectively—significantly improving practicality and scalability for long-sequence recommendation.

Technology Category

Application Category

📝 Abstract
Sequential recommendation aims to model users' evolving preferences based on their historical interactions. Recent advances leverage Transformer-based architectures to capture global dependencies, but existing methods often suffer from high computational overhead, primarily due to discontinuous memory access in temporal encoding and dense attention over long sequences. To address these limitations, we propose FuXi-$γ$, a novel sequential recommendation framework that improves both effectiveness and efficiency through principled architectural design. FuXi-$γ$ adopts a decoder-only Transformer structure and introduces two key innovations: (1) An exponential-power temporal encoder that encodes relative temporal intervals using a tunable exponential decay function inspired by the Ebbinghaus forgetting curve. This encoder enables flexible modeling of both short-term and long-term preferences while maintaining high efficiency through continuous memory access and pure matrix operations. (2) A diagonal-sparse positional mechanism that prunes low-contribution attention blocks using a diagonal-sliding strategy guided by the persymmetry of Toeplitz matrix. Extensive experiments on four real-world datasets demonstrate that FuXi-$γ$ achieves state-of-the-art performance in recommendation quality, while accelerating training by up to 4.74$ imes$ and inference by up to 6.18$ imes$, making it a practical and scalable solution for long-sequence recommendation. Our code is available at https://github.com/Yeedzhi/FuXi-gamma.
Problem

Research questions and friction points this paper is trying to address.

Improves efficiency in sequential recommendation by reducing computational overhead
Models both short-term and long-term user preferences with flexible temporal encoding
Accelerates training and inference for long-sequence recommendation tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Exponential-power temporal encoder models preferences efficiently
Diagonal-sparse positional mechanism prunes low-contribution attention blocks
Decoder-only Transformer structure enhances efficiency and scalability
🔎 Similar Papers
No similar papers found.
D
Dezhi Yi
College of Computer Science, DISSec, Nankai University, Tianjin, China
W
Wei Guo
Huawei Technologies, Shanghai, China
W
Wenyang Cui
College of Computer Science, DISSec, Nankai University, Tianjin, China
W
Wenxuan He
College of Computer Science, DISSec, Nankai University, Tianjin, China
Huifeng Guo
Huifeng Guo
Huawei, Harbin Institute of Technology
Recommender SystemDeep LearningData Mining.
Y
Yong Liu
Huawei Technologies, Shanghai, China
Zhenhua Dong
Zhenhua Dong
Noah's ark lab, Huawei Technologies Co., Ltd.
Recommender systemcausal inferencecountrfactual learningtrustworthy AImachine learning
Y
Ye Lu
College of Computer Science, DISSec, Nankai University, Tianjin, China