🤖 AI Summary
Existing molecular graph neural networks (GNNs) predominantly rely on XYZ atomic coordinates, overlooking rich textual chemical information—such as IUPAC names, molecular formulas, and physicochemical properties—abundant in databases like PubChem. To address this, we propose a physics-aware multimodal GNN framework that jointly encodes molecular graphs and textual descriptions. Our method integrates BERT-like text embeddings with explicit physical property encodings and introduces a gated attention mechanism to synergistically model geometric structure and chemical semantics. This work is the first to systematically leverage PubChem’s textual metadata to enhance GNN representations; it further reveals a common limitation of mainstream GNNs: their implicit learning of physical representations lacks robustness and task adaptability. On multiple benchmark datasets, our model achieves significant improvements in predicting electronic properties—including band gaps and ionization energies—with gains exhibiting task-specific patterns.
📝 Abstract
Molecular graph neural networks (GNNs) often focus exclusively on XYZ-based geometric representations and thus overlook valuable chemical context available in public databases like PubChem. This work introduces a multimodal framework that integrates textual descriptors, such as IUPAC names, molecular formulas, physicochemical properties, and synonyms, alongside molecular graphs. A gated fusion mechanism balances geometric and textual features, allowing models to exploit complementary information. Experiments on benchmark datasets indicate that adding textual data yields notable improvements for certain electronic properties, while gains remain limited for others. Furthermore, the GNN architectures display similar performance patterns (improving and deteriorating on analogous targets), suggesting they learn comparable representations rather than distinctly different physical insights.