Elign: Equivariant Diffusion Model Alignment from Foundational Machine Learning Force Fields

📅 2026-01-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing E(3)-equivariant diffusion models for 3D molecular conformation generation are prone to biases from low-accuracy training data and struggle to approximate the thermodynamic equilibrium distribution governed by high-fidelity Hamiltonians. This work proposes Elign, a framework that uniquely integrates physics-based guidance entirely into the training phase. By leveraging a pretrained machine learning force field to replace costly quantum chemical computations, Elign introduces the FED-GRPO algorithm, which enhances reinforcement learning reward signals through force–energy decoupling and group-normalized optimization. The approach achieves significantly lower DFT-computed energies and forces in generated conformations—improving structural stability and physical consistency—while maintaining inference speed comparable to unguided sampling.

Technology Category

Application Category

📝 Abstract
Generative models for 3D molecular conformations must respect Euclidean symmetries and concentrate probability mass on thermodynamically favorable, mechanically stable structures. However, E(3)-equivariant diffusion models often reproduce biases from semi-empirical training data rather than capturing the equilibrium distribution of a high-fidelity Hamiltonian. While physics-based guidance can correct this, it faces two computational bottlenecks: expensive quantum-chemical evaluations (e.g., DFT) and the need to repeat such queries at every sampling step. We present Elign, a post-training framework that amortizes both costs. First, we replace expensive DFT evaluations with a faster, pretrained foundational machine-learning force field (MLFF) to provide physical signals. Second, we eliminate repeated run-time queries by shifting physical steering to the training phase. To achieve the second amortization, we formulate reverse diffusion as a reinforcement learning problem and introduce Force--Energy Disentangled Group Relative Policy Optimization (FED-GRPO) to fine-tune the denoising policy. FED-GRPO includes a potential-based energy reward and a force-based stability reward, which are optimized and group-normalized independently. Experiments show that Elign generates conformations with lower gold-standard DFT energies and forces, while improving stability. Crucially, inference remains as fast as unguided sampling, since no energy evaluations are required during generation.
Problem

Research questions and friction points this paper is trying to address.

E(3)-equivariant diffusion models
molecular conformations
physics-based guidance
computational bottlenecks
equilibrium distribution
Innovation

Methods, ideas, or system contributions that make the work stand out.

Equivariant Diffusion
Machine Learning Force Field
Reinforcement Learning
Conformation Generation
FED-GRPO
🔎 Similar Papers
No similar papers found.
Y
Yunyang Li
Department of Computer Science, Yale University, New Haven, USA
Lin Huang
Lin Huang
Stanford University, The Chinese University of Hong Kong
computational genomicsfault-tolerant computingdesign automationand multi-core architecture
L
Luojia Xia
Program in Computational Biology and Biomedical Informatics, Yale University, New Haven, USA
W
Wenhe Zhang
Department of Computer Science, Yale University, New Haven, USA
M
Mark B. Gerstein
Program in Computational Biology and Biomedical Informatics, Yale University, New Haven, USA