Ensemble Knowledge Distillation for Machine Learning Interatomic Potentials

๐Ÿ“… 2025-03-18
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
High-accuracy coupled-cluster (CC) energy data for small molecules are scarce, and such datasets typically lack corresponding force labelsโ€”leading to poor force prediction accuracy and unstable molecular dynamics simulations when training machine-learned interatomic potentials (MLIPs) on energy-only data. Method: We propose an Ensemble Knowledge Distillation (EKD) framework that leverages multiple pre-trained teacher models to generate high-quality force labels, enabling a student MLIP to jointly fit both reference energies and distilled forces. Contribution/Results: This is the first application of ensemble knowledge distillation to MLIP training; it eliminates the need for ground-truth force labels while substantially improving force prediction accuracy and MD stability. Evaluated on the COMP6 benchmark, EKD achieves new state-of-the-art performance, demonstrating strong generalization across chemical, biological, and materials systems.

Technology Category

Application Category

๐Ÿ“ Abstract
Machine learning interatomic potentials (MLIPs) are a promising tool to accelerate atomistic simulations and molecular property prediction. The quality of MLIPs strongly depends on the quantity of available training data as well as the quantum chemistry (QC) level of theory used to generate that data. Datasets generated with high-fidelity QC methods, such as coupled cluster, are typically restricted to small molecules and may be missing energy gradients. With this limited quantity of data, it is often difficult to train good MLIP models. We present an ensemble knowledge distillation (EKD) method to improve MLIP accuracy when trained to energy-only datasets. In our EKD approach, first, multiple teacher models are trained to QC energies and then used to generate atomic forces for all configurations in the dataset. Next, a student MLIP is trained to both QC energies and to ensemble-averaged forces generated by the teacher models. We apply this workflow on the ANI-1ccx dataset which consists of organic molecules with configuration energies computed at the coupled cluster level of theory. The resulting student MLIPs achieve new state-of-the-art accuracy on the out-of-sample COMP6 benchmark and improved stability for molecular dynamics simulations. The EKD approach for MLIP is broadly applicable for chemical, biomolecular and materials science simulations.
Problem

Research questions and friction points this paper is trying to address.

Improves MLIP accuracy with limited energy-only datasets.
Generates atomic forces using ensemble teacher models.
Enhances molecular dynamics simulation stability and accuracy.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Ensemble knowledge distillation improves MLIP accuracy
Teacher models generate atomic forces for training
Student MLIP trained on QC energies and forces
๐Ÿ”Ž Similar Papers
No similar papers found.
Sakib Matin
Sakib Matin
Postdoc, Los Alamos National Laboratory
Machine LearningStatistical MechanicsBiophysicsNon-linear Dynamics
Emily Shinkle
Emily Shinkle
Scientist, Los Alamos National Laboratory
Y
Yulia Pimonova
Computer, Computational, and Statistical Sciences Division, Los Alamos National Laboratory, Los Alamos, NM 87545, USA
G
G. Craven
Theoretical Division, Los Alamos National Laboratory, Los Alamos, NM 87545, USA
Aleksandra Pachalieva
Aleksandra Pachalieva
Los Alamos National Laboratory
Computational Fluid MechanicsMachine Learning
Y
Ying Wai Li
Computer, Computational, and Statistical Sciences Division, Los Alamos National Laboratory, Los Alamos, NM 87545, USA
Kipton Barros
Kipton Barros
Los Alamos National Laboratory
Statistical physicsComputational physicsMachine learning
N
N. Lubbers
Computer, Computational, and Statistical Sciences Division, Los Alamos National Laboratory, Los Alamos, NM 87545, USA