Learning Inter-Atomic Potentials without Explicit Equivariance

πŸ“… 2025-09-25
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of implicitly enforcing rotational and translational symmetries (SO(3) Γ— ℝ³) in machine-learned interatomic potentials (MLIPs) without resorting to explicit equivariant architectures. We propose TransIPβ€”the first Transformer-based paradigm for implicitly equivariant potential energy modeling. By jointly optimizing embedding spaces and representation learning, TransIP enables non-equivariant models to spontaneously acquire physical symmetries during training, circumventing the architectural rigidity and computational overhead inherent in explicit equivariant models. Evaluated on the large-scale OMol25 dataset, TransIP achieves 40–60% higher force prediction accuracy than data-augmentation baselines and matches the performance of state-of-the-art explicit equivariant methods. Consequently, TransIP significantly enhances the flexibility, computational efficiency, and scalability of MLIPs while preserving physical consistency.

Technology Category

Application Category

πŸ“ Abstract
Accurate and scalable machine-learned inter-atomic potentials (MLIPs) are essential for molecular simulations ranging from drug discovery to new material design. Current state-of-the-art models enforce roto-translational symmetries through equivariant neural network architectures, a hard-wired inductive bias that can often lead to reduced flexibility, computational efficiency, and scalability. In this work, we introduce TransIP: Transformer-based Inter-Atomic Potentials, a novel training paradigm for interatomic potentials achieving symmetry compliance without explicit architectural constraints. Our approach guides a generic non-equivariant Transformer-based model to learn SO(3)-equivariance by optimizing its representations in the embedding space. Trained on the recent Open Molecules (OMol25) collection, a large and diverse molecular dataset built specifically for MLIPs and covering different types of molecules (including small organics, biomolecular fragments, and electrolyte-like species), TransIP attains comparable performance in machine-learning force fields versus state-of-the-art equivariant baselines. Further, compared to a data augmentation baseline, TransIP achieves 40% to 60% improvement in performance across varying OMol25 dataset sizes. More broadly, our work shows that learned equivariance can be a powerful and efficient alternative to equivariant or augmentation-based MLIP models.
Problem

Research questions and friction points this paper is trying to address.

Learning inter-atomic potentials without architectural equivariance constraints
Achieving symmetry compliance through embedding space optimization
Improving computational flexibility and scalability in molecular simulations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transformer-based model learns equivariance without architectural constraints
Optimizes representations in embedding space for SO(3)-equivariance
Achieves performance comparable to equivariant baselines without hard-wired symmetry
πŸ”Ž Similar Papers
No similar papers found.
A
Ahmed A. Elhag
Computer Science, University of Oxford
A
Arun Raja
Statistics, University of Oxford
Alex Morehead
Alex Morehead
Hopper Postdoctoral Fellow, Berkeley Lab
Machine LearningDeep LearningComputational BiologyGenerative Modeling
S
Samuel M. Blau
Energy Technologies Area, Lawrence Berkeley National Laboratory
Garrett M. Morris
Garrett M. Morris
University of Oxford
Computational ChemistryComputer-Aided Drug DesignVirtual ScreeningDockingMachine Learning & AI
M
Michael M. Bronstein
Computer Science, University of Oxford