Unified Molecule Generation and Property Prediction

📅 2025-04-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Molecular generation and property prediction have long been treated as disjoint tasks; joint modeling faces challenges from architectural coupling and optimization conflicts between the two objectives. Method: We propose Hyformer—the first Transformer-based unified generative-discriminative model—featuring an alternating attention masking mechanism and a multi-task unified pretraining framework. It jointly optimizes autoregressive generation and property discrimination within a single architecture, enabling bidirectional capability enhancement. Contribution/Results: By modeling the joint distribution to unify representation spaces, Hyformer achieves state-of-the-art performance across diverse downstream tasks—including molecular representation learning, hit compound identification, and antimicrobial peptide design—outperforming dedicated single-task models. It significantly improves both generation quality (e.g., validity, uniqueness, novelty) and prediction accuracy (e.g., regression and classification metrics), demonstrating the efficacy of co-optimizing generative and discriminative objectives in a shared latent space.

Technology Category

Application Category

📝 Abstract
Modeling the joint distribution of the data samples and their properties allows to construct a single model for both data generation and property prediction, with synergistic capabilities reaching beyond purely generative or predictive models. However, training joint models presents daunting architectural and optimization challenges. Here, we propose Hyformer, a transformer-based joint model that successfully blends the generative and predictive functionalities, using an alternating attention mask together with a unified pre-training scheme. We show that Hyformer rivals other joint models, as well as state-of-the-art molecule generation and property prediction models. Additionally, we show the benefits of joint modeling in downstream tasks of molecular representation learning, hit identification and antimicrobial peptide design.
Problem

Research questions and friction points this paper is trying to address.

Model joint distribution of molecules and properties
Overcome architectural and optimization challenges
Improve molecular generation and property prediction
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transformer-based joint model for generation and prediction
Alternating attention mask for blending functionalities
Unified pre-training scheme enhances synergistic capabilities
🔎 Similar Papers
No similar papers found.
A
Adam Izdebski
Institute of AI for Health, Helmholtz Zentrum Munchen; Faculty of Mathematics, Informatics and Mechanics, University of Warsaw
Jan Olszewski
Jan Olszewski
Student, University of Warsaw
Deep Learning
Pankhil Gawade
Pankhil Gawade
Helmholtz Munich
Gen AI for Drug Design
K
Krzysztof Koras
Ardigen SA
V
Valentin Rauscher
Institute of AI for Health, Helmholtz Zentrum Munchen
J
Jakub M Tomczak
Eindhoven University of Technology
Ewa Szczurek
Ewa Szczurek
Associate Professor at University of Warsaw / Institute AI for Health, Helmholtz Zentrum München
computational biologymachine learningartificial intelligence