Crystalite: A Lightweight Transformer for Efficient Crystal Modeling

📅 2026-04-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the inefficiency and limited structural modeling capacity of conventional crystal generation methods that rely on computationally expensive equivariant graph neural networks. The authors propose a lightweight diffusion Transformer that replaces high-dimensional one-hot encodings with subatomic tokenization and introduces a Geometry-Enhanced Module (GEM) to directly inject periodic geometric information from minimum-image pairs into the attention mechanism. This approach preserves the computational efficiency of standard Transformers while effectively integrating crystallographic chemistry and spatial symmetry. The method achieves state-of-the-art performance in both crystal structure prediction and ab initio generation tasks, attaining the highest S.U.N. discovery score and demonstrating significantly faster sampling speeds compared to existing geometry-intensive baselines.
📝 Abstract
Generative models for crystalline materials often rely on equivariant graph neural networks, which capture geometric structure well but are costly to train and slow to sample. We present Crystalite, a lightweight diffusion Transformer for crystal modeling built around two simple inductive biases. The first is Subatomic Tokenization, a compact chemically structured atom representation that replaces high-dimensional one-hot encodings and is better suited to continuous diffusion. The second is the Geometry Enhancement Module (GEM), which injects periodic minimum-image pair geometry directly into attention through additive geometric biases. Together, these components preserve the simplicity and efficiency of a standard Transformer while making it better matched to the structure of crystalline materials. Crystalite achieves state-of-the-art results on crystal structure prediction benchmarks, and de novo generation performance, attaining the best S.U.N. discovery score among the evaluated baselines while sampling substantially faster than geometry-heavy alternatives.
Problem

Research questions and friction points this paper is trying to address.

crystal modeling
generative models
equivariant graph neural networks
sampling efficiency
training cost
Innovation

Methods, ideas, or system contributions that make the work stand out.

Subatomic Tokenization
Geometry Enhancement Module
Diffusion Transformer
Crystal Structure Prediction
De Novo Generation
🔎 Similar Papers
No similar papers found.
T
Tin Hadži Veljković
UvA-Bosch Delta Lab, University of Amsterdam
J
Joshua Rosenthal
UvA-Bosch Delta Lab, University of Amsterdam
I
Ivor Lončarić
Ruđer Bošković Institute
Jan-Willem van de Meent
Jan-Willem van de Meent
University of Amsterdam
Probabilistic ProgrammingMachine LearningArtificial IntelligenceData Science