🤖 AI Summary
This work addresses the challenge of forecasting states for unknown time-varying partial differential equation (PDE) systems characterized exclusively by short trajectory data. Method: We propose a decoupled state prediction framework: first explicitly discovering the governing evolution operator from short sequences, then performing next-state prediction via numerical time integration. Our core innovation is the “operator discovery” paradigm—decoupling dynamical modeling from state evolution—and introducing a hypernetwork to generate lightweight operator-network parameters, enabling cross-physics generalization and efficient pretraining. The methodology comprises multi-physics joint pretraining followed by downstream fine-tuning. Results: Experiments demonstrate state-of-the-art (SOTA) performance across multiple PDE forecasting benchmarks with significantly reduced training epochs. The pre-trained model exhibits strong generalization across diverse physical domains, and fine-tuned variants consistently maintain top-tier accuracy.
📝 Abstract
We address the problem of predicting the next state of a dynamical system governed by unknown temporal partial differential equations (PDEs) using only a short trajectory. While standard transformers provide a natural black-box solution to this task, the presence of a well-structured evolution operator in the data suggests a more tailored and efficient approach. Specifically, when the PDE is fully known, classical numerical solvers can evolve the state accurately with only a few parameters. Building on this observation, we introduce DISCO, a model that uses a large hypernetwork to process a short trajectory and generate the parameters of a much smaller operator network, which then predicts the next state through time integration. Our framework decouples dynamics estimation (i.e., DISCovering an evolution operator from a short trajectory) from state prediction (i.e., evolving this operator). Experiments show that pretraining our model on diverse physics datasets achieves state-of-the-art performance while requiring significantly fewer epochs. Moreover, it generalizes well and remains competitive when fine-tuned on downstream tasks.