Scalable Pretraining of Large Mixture of Experts Language Models on Aurora Super Computer

📅 2026-04-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the computational scalability and training stability challenges inherent in pretraining extremely large-scale Mixture-of-Experts (MoE) language models. Leveraging the Aurora supercomputer and a self-developed Optimus training framework, the authors efficiently pretrained the Mula series of MoE models from scratch, spanning from billions to 220 billion parameters. Key innovations include an EP-Aware sharded optimizer, custom GPU expert computation kernels, and a highly reliable distributed fault-tolerance mechanism. The system achieves approximately 90% strong scaling efficiency across 12,288 GPUs, yielding up to a 1.71× speedup in training throughput while successfully completing trillion-token-scale pretraining with high stability.
📝 Abstract
Pretraining Large Language Models (LLMs) from scratch requires massive amount of compute. Aurora super computer is an ExaScale machine with 127,488 Intel PVC (Ponte Vechio) GPU tiles. In this work, we showcase LLM pretraining on Aurora at the scale of 1000s of GPU tiles. Towards this effort, we developed Optimus, an inhouse training library with support for standard large model training techniques. Using Optimus, we first pretrained Mula-1B, a 1 Billion dense model and Mula-7B-A1B, a 7 Billion Mixture of Experts (MoE) model from scratch on 3072 GPU tiles for the full 4 trillion tokens of the OLMoE-mix-0924 dataset. We then demonstrated model scaling by pretraining three large MoE models Mula-20B-A2B, Mula-100B-A7B, and Mula-220B-A10B till 100 Billion tokens on the same dataset. On our largest model Mula-220B-A10B, we pushed the compute scaling from 384 to 12288 GPU tiles and observed scaling efficiency of around 90% at 12288 GPU tiles. We significantly improved the runtime performance of MoE models using custom GPU kernels for expert computation, and a novel EP-Aware sharded optimizer resulting in training speedups up to 1.71x. As part of the Optimus library, we also developed a robust set of reliability and fault tolerant features to improve training stability and continuity at scale.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Mixture of Experts
Pretraining
Scalability
Fault Tolerance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Mixture of Experts
ExaScale Training
Custom GPU Kernels
EP-Aware Optimizer
Fault Tolerance
🔎 Similar Papers
No similar papers found.