Physics-inspired transformer quantum states via latent imaginary-time evolution

๐Ÿ“… 2026-02-03
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the limited physical interpretability and difficulty in systematic optimization of conventional Transformer-based neural quantum states. The authors propose a physics-inspired modeling framework that interprets neural quantum states as neural approximations of imaginary-time evolution in a latent space. By introducing a static effective Hamiltonian and employing a Trotterโ€“Suzuki decomposition, they construct an interpretable architecture enhanced with inter-layer weight sharing and a high-precision propagation mechanism. This design improves both expressive power and physical consistency without increasing the number of variational parameters. Numerical experiments on the Jโ‚โ€“Jโ‚‚ Heisenberg model demonstrate that the proposed method achieves or surpasses the accuracy of state-of-the-art Transformer quantum states while using significantly fewer variational parameters.

Technology Category

Application Category

๐Ÿ“ Abstract
Neural quantum states (NQS) are powerful ans\"atze in the variational Monte Carlo framework, yet their architectures are often treated as black boxes. We propose a physically transparent framework in which NQS are treated as neural approximations to latent imaginary-time evolution. This viewpoint suggests that standard Transformer-based NQS (TQS) architectures correspond to physically unmotivated effective Hamiltonians dependent on imaginary time in a latent space. Building on this interpretation, we introduce physics-inspired transformer quantum states (PITQS), which enforce a static effective Hamiltonian by sharing weights across layers and improve propagation accuracy via Trotter-Suzuki decompositions without increasing the number of variational parameters. For the frustrated $J_1$-$J_2$ Heisenberg model, our ans\"atze achieve accuracies comparable to or exceeding state-of-the-art TQS while using substantially fewer variational parameters. This study demonstrates that reinterpreting the deep network structure as a latent cooling process enables a more physically grounded, systematic, and compact design, thereby bridging the gap between black-box expressivity and physically transparent construction.
Problem

Research questions and friction points this paper is trying to address.

neural quantum states
Transformer
imaginary-time evolution
effective Hamiltonian
physical transparency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Physics-inspired Transformer
Neural Quantum States
Imaginary-time Evolution
Trotter-Suzuki Decomposition
Variational Monte Carlo
๐Ÿ”Ž Similar Papers
No similar papers found.
K
Kimihiro Yamazaki
Graduate School of Information Science and Technology, The University of Osaka, 1-5 Yamadaoka, Suita, Osaka, Japan; Fujitsu Laboratories, Fujitsu Limited, 4-1-1 Kamikodanaka, Nakahara-ku, Kawasaki, Kanagawa, Japan
I
Itsushi Sakata
Center for Advanced Intelligence Project, RIKEN, 1-4-1 Nihonbashi, Chuo-ku, Tokyo, Japan
T
Takuya Konishi
Graduate School of Information Science and Technology, The University of Osaka, 1-5 Yamadaoka, Suita, Osaka, Japan; Center for Advanced Intelligence Project, RIKEN, 1-4-1 Nihonbashi, Chuo-ku, Tokyo, Japan
Yoshinobu Kawahara
Yoshinobu Kawahara
The University of Osaka & RIKEN Center for Advanced Intelligence Project
Machine LearningDynamical SystemsNonlinear Dynamics