🤖 AI Summary
Deploying trillion-parameter foundation models for online advertising recommendation faces critical challenges in meeting strict low-latency inference requirements and adapting to dynamic data distributions in industrial settings.
Method: This paper proposes ExFM, an external distillation framework featuring: (1) a teacher foundation model reuse mechanism enabling cross-distribution knowledge transfer; (2) co-designed auxiliary heads and student adapters for computational cost sharing; and (3) a collaborative architecture integrating a Data Augmentation System (DAS) with streaming self-adaptive modeling across base and vertical models.
Results: Evaluated on both industrial and public benchmarks, ExFM significantly improves recommendation accuracy while maintaining controllable inference latency and reducing training overhead by over 40%. It establishes a novel paradigm for efficient deployment of ultra-large-scale models in real-time recommendation systems.
📝 Abstract
Ads recommendation is a prominent service of online advertising systems and has been actively studied. Recent studies indicate that scaling-up and advanced design of the recommendation model can bring significant performance improvement. However, with a larger model scale, such prior studies have a significantly increasing gap from industry as they often neglect two fundamental challenges in industrial-scale applications. First, training and inference budgets are restricted for the model to be served, exceeding which may incur latency and impair user experience. Second, large-volume data arrive in a streaming mode with data distributions dynamically shifting, as new users/ads join and existing users/ads leave the system. We propose the External Large Foundation Model (ExFM) framework to address the overlooked challenges. Specifically, we develop external distillation and a data augmentation system (DAS) to control the computational cost of training/inference while maintaining high performance. We design the teacher in a way like a foundation model (FM) that can serve multiple students as vertical models (VMs) to amortize its building cost. We propose Auxiliary Head and Student Adapter to mitigate the data distribution gap between FM and VMs caused by the streaming data issue. Comprehensive experiments on internal industrial-scale applications and public datasets demonstrate significant performance gain by ExFM.