🤖 AI Summary
Traditional machine-learned interatomic potential (MLIP) training relies on high-fidelity ab initio force labels, incurring prohibitive data generation and computational costs—especially for structural relaxation tasks. This work proposes an end-to-end differentiable relaxation framework: structural relaxation is formulated as an iterative mapping process, and MLIP parameters are optimized directly via backpropagation through time (BPTT), using only the final relaxed structure’s prediction error as the supervisory signal—eliminating dependence on force labels. Crucially, the method fine-tunes the potential energy surface itself without altering the underlying relaxation algorithm or introducing auxiliary learnable modules. Evaluated across diverse material systems, it achieves ~50% average reduction in structural prediction error. Moreover, it exhibits strong robustness to hyperparameter variations and procedural changes, significantly enhancing generalizability and deployment efficiency.
📝 Abstract
Vital to the creation of advanced materials is performing structural relaxations. Traditional approaches built on physics-derived first-principles calculations are computationally expensive, motivating the creation of machine-learning interatomic potentials (MLIPs). Traditional approaches to training MLIPs for structural relaxations involves training models to faithfully reproduce first-principles computed forces. We propose a fine-tuning method to be used on a pretrained MLIP in which we create a fully-differentiable end-to-end simulation loop that optimizes the predicted final structures directly. Trajectories are unrolled and gradients are tracked through the entire relaxation. We show that this method achieves substantial performance gains when applied to pretrained models, leading to a nearly $50%$ reduction in test error across the sample datasets. Interestingly, we show the process is robust to substantial variation in the relaxation setup, achieving negligibly different results across varied hyperparameter and procedural modifications. Experimental results indicate this is due to a ``preference'' of BPTT to modify the MLIP rather than the other trainable parameters. Of particular interest to practitioners is that this approach lowers the data requirements for producing an effective domain-specific MLIP, addressing a common bottleneck in practical deployment.