🤖 AI Summary
Predicting excess Gibbs energy (G<sup>E</sup>) of multicomponent mixtures solely from molecular structures has long suffered from poor thermodynamic consistency and limited extrapolation capability. To address this, we propose HANNA: an end-to-end trainable neural network that intrinsically enforces fundamental thermodynamic laws—particularly the Gibbs–Duhem equation—as hard architectural constraints, and employs geometric projection for parameter-free extrapolation to arbitrary component numbers. Furthermore, we introduce a novel surrogate solver integrating liquid–liquid equilibrium (LLE) data to enrich training information beyond conventional vapor–liquid equilibrium (VLE) measurements. Trained exclusively on binary experimental data, HANNA outperforms state-of-the-art methods in prediction accuracy, thermodynamic consistency, and generalization across mixture composition and component count. The model, source code, and an interactive MLPROP platform are publicly released to accelerate phase-equilibrium prediction and molecular design in chemical engineering.
📝 Abstract
The excess Gibbs energy plays a central role in chemical engineering and chemistry, providing a basis for modeling the thermodynamic properties of liquid mixtures. Predicting the excess Gibbs energy of multi-component mixtures solely from the molecular structures of their components is a long-standing challenge. In this work, we address this challenge by integrating physical laws as hard constraints within a flexible neural network. The resulting model, HANNA, was trained end-to-end on an extensive experimental dataset for binary mixtures from the Dortmund Data Bank, guaranteeing thermodynamically consistent predictions. A novel surrogate solver developed in this work enabled the inclusion of liquid-liquid equilibrium data in the training process. Furthermore, a geometric projection method was applied to enable robust extrapolations to multi-component mixtures, without requiring additional parameters. We demonstrate that HANNA delivers excellent predictions, clearly outperforming state-of-the-art benchmark methods in accuracy and scope. The trained model and corresponding code are openly available, and an interactive interface is provided on our website, MLPROP.