Prior-Guided Symbolic Regression: Towards Scientific Consistency in Equation Discovery

📅 2026-02-13
📈 Citations: 0
Influential: 0
📄 PDF

Technology Category

Application Category

📝 Abstract
Symbolic Regression (SR) aims to discover interpretable equations from observational data, with the potential to reveal underlying principles behind natural phenomena. However, existing approaches often fall into the Pseudo-Equation Trap: producing equations that fit observations well but remain inconsistent with fundamental scientific principles. A key reason is that these approaches are dominated by empirical risk minimization, lacking explicit constraints to ensure scientific consistency. To bridge this gap, we propose PG-SR, a prior-guided SR framework built upon a three-stage pipeline consisting of warm-up, evolution, and refinement. Throughout the pipeline, PG-SR introduces a prior constraint checker that explicitly encodes domain priors as executable constraint programs, and employs a Prior Annealing Constrained Evaluation (PACE) mechanism during the evolution stage to progressively steer discovery toward scientifically consistent regions. Theoretically, we prove that PG-SR reduces the Rademacher complexity of the hypothesis space, yielding tighter generalization bounds and establishing a guarantee against pseudo-equations. Experimentally, PG-SR outperforms state-of-the-art baselines across diverse domains, maintaining robustness to varying prior quality, noisy data, and data scarcity.
Problem

Research questions and friction points this paper is trying to address.

Symbolic Regression
Scientific Consistency
Pseudo-Equation Trap
Equation Discovery
Domain Priors
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symbolic Regression
Scientific Consistency
Prior-Guided Learning
Constrained Optimization
Equation Discovery
🔎 Similar Papers
No similar papers found.
J
Jing Xiao
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
X
Xinhai Chen
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
J
Jiaming Peng
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
Qinglin Wang
Qinglin Wang
National University of Defense Technology
Parallel algorithmsHigh Performance ComputingDeep LearningMachine LearningGPU
M
Menghan Jia
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
Z
Zhiquan Lai
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
G
Guangping Yu
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
Dongsheng Li
Dongsheng Li
Professor, School of Computer Science, National University of Defense Technology
Distributed ComputingParallel ComputingCloud ComputingPeer-to-Peer ComputingBig Data
T
Tiejun Li
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China
J
Jie Liu
College of Computer Science and Technology, National University of Defense Technology, Changsha 410073, China