MTFM: A Scalable and Alignment-free Foundation Model for Industrial Recommendation in Meituan

📅 2026-02-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges in industrial recommendation systems where multi-scenario data are often misaligned and existing cross-domain methods suffer from high computational costs and poor scalability. To overcome these limitations, we propose MTFM, a Transformer-based foundational model for industrial recommendation that introduces a novel paradigm for multi-scenario modeling without requiring input alignment. MTFM unifies heterogeneous cross-domain data through a shared token representation and enhances efficiency via user-level sample aggregation and tailored attention mechanisms, including Grouped-Query Attention and Hybrid Target Attention. Experimental results demonstrate that MTFM consistently improves recommendation performance as model size and scenario data scale up, while achieving substantially higher training and inference throughput compared to existing approaches.

Technology Category

Application Category

📝 Abstract
Industrial recommendation systems typically involve multiple scenarios, yet existing cross-domain (CDR) and multi-scenario (MSR) methods often require prohibitive resources and strict input alignment, limiting their extensibility. We propose MTFM (Meituan Foundation Model for Recommendation), a transformer-based framework that addresses these challenges. Instead of pre-aligning inputs, MTFM transforms cross-domain data into heterogeneous tokens, capturing multi-scenario knowledge in an alignment-free manner. To enhance efficiency, we first introduce a multi-scenario user-level sample aggregation that significantly enhances training throughput by reducing the total number of instances. We further integrate Grouped-Query Attention and a customized Hybrid Target Attention to minimize memory usage and computational complexity. Furthermore, we implement various system-level optimizations, such as kernel fusion and the elimination of CPU-GPU blocking, to further enhance both training and inference throughput. Offline and online experiments validate the effectiveness of MTFM, demonstrating that significant performance gains are achieved by scaling both model capacity and multi-scenario training data.
Problem

Research questions and friction points this paper is trying to address.

industrial recommendation
cross-domain recommendation
multi-scenario recommendation
scalability
input alignment
Innovation

Methods, ideas, or system contributions that make the work stand out.

alignment-free
foundation model
multi-scenario recommendation
heterogeneous tokens
system-level optimization
X
Xin Song
Meituan, Beijing, China
Z
Zhilin Guan
Meituan, Beijing, China
Ruidong Han
Ruidong Han
Meituan
recommender systemgenerative model
B
Binghao Tang
Meituan, Beijing, China
Tianwen Chen
Tianwen Chen
Applied AI/ML Researcher
Machine Learning
B
Bing Li
Meituan, Beijing, China
Z
Zihao Li
Meituan, Beijing, China
H
Han Zhang
Meituan, Beijing, China
F
Fei Jiang
Meituan, Beijing, China
Qing Wang
Qing Wang
IBM Research China
computer visionstatistical signal processingmobile communication
F
Fengyi Li
Meituan, Beijing, China
C
Chunzhen Jing
Meituan, Beijing, China
L
Lei Yu
Meituan, Beijing, China
W
Wei Lin
Meituan, Beijing, China