🤖 AI Summary
Addressing the challenge of efficiently and accurately modeling multiscale features in high-resolution turbulence simulations—from large-scale eddies down to viscous dissipation scales—this work introduces a multiscale hierarchical Turbulence Transformer architecture coupled with RingX sequence parallelism. Our method enables pixel-level long-context learning, employs hierarchical attention mechanisms, and compresses sequence length to achieve, for the first time, full-resolution reconstruction of dissipative-scale eddies by an AI model—overcoming computational bottlenecks inherent in conventional long-context modeling. We trained the model at trillion-grid scale on the Frontier supercomputer using 32,768 AMD GPUs, attaining a peak performance of 1.1 EFLOPS and 94% strong scaling efficiency. This establishes a scalable, high-fidelity AI modeling paradigm for extreme-scale, multiphysics-coupled predictions.
📝 Abstract
Turbulence plays a crucial role in multiphysics applications, including aerodynamics, fusion, and combustion. Accurately capturing turbulence's multiscale characteristics is essential for reliable predictions of multiphysics interactions, but remains a grand challenge even for exascale supercomputers and advanced deep learning models. The extreme-resolution data required to represent turbulence, ranging from billions to trillions of grid points, pose prohibitive computational costs for models based on architectures like vision transformers. To address this challenge, we introduce a multiscale hierarchical Turbulence Transformer that reduces sequence length from billions to a few millions and a novel RingX sequence parallelism approach that enables scalable long-context learning. We perform scaling and science runs on the Frontier supercomputer. Our approach demonstrates excellent performance up to 1.1 EFLOPS on 32,768 AMD GPUs, with a scaling efficiency of 94%. To our knowledge, this is the first AI model for turbulence that can capture small-scale eddies down to the dissipative range.