Causal Foundation Models: Disentangling Physics from Instrument Properties

📅 2025-07-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Structured time-series modeling suffers from poor generalization due to entanglement between underlying physical processes and systematic instrument-specific biases—especially in heterogeneous or multi-instrument settings. To address this, we propose a causal-driven dual-encoder foundation model—the first time-series foundation model to explicitly disentangle physical signals from instrument effects. Our method constructs structured contrastive learning objectives from observational triplets and employs latent-variable modeling to achieve separated representations of the two factors. Trained and validated on simulated astronomical time-series data (e.g., TESS variable stars), it significantly outperforms single-latent-space baselines. In downstream forecasting tasks, it enhances few-shot generalization and enables rapid cross-instrument adaptation. This work establishes a new paradigm for interpretable and robust time-series foundation models grounded in causal disentanglement.

Technology Category

Application Category

📝 Abstract
Foundation models for structured time series data must contend with a fundamental challenge: observations often conflate the true underlying physical phenomena with systematic distortions introduced by measurement instruments. This entanglement limits model generalization, especially in heterogeneous or multi-instrument settings. We present a causally-motivated foundation model that explicitly disentangles physical and instrumental factors using a dual-encoder architecture trained with structured contrastive learning. Leveraging naturally occurring observational triplets (i.e., where the same target is measured under varying conditions, and distinct targets are measured under shared conditions) our model learns separate latent representations for the underlying physical signal and instrument effects. Evaluated on simulated astronomical time series designed to resemble the complexity of variable stars observed by missions like NASA's Transiting Exoplanet Survey Satellite (TESS), our method significantly outperforms traditional single-latent space foundation models on downstream prediction tasks, particularly in low-data regimes. These results demonstrate that our model supports key capabilities of foundation models, including few-shot generalization and efficient adaptation, and highlight the importance of encoding causal structure into representation learning for structured data.
Problem

Research questions and friction points this paper is trying to address.

Disentangling physical phenomena from instrument distortions in time series data
Improving model generalization in multi-instrument observational settings
Enhancing few-shot prediction accuracy for structured astronomical data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-encoder architecture disentangles physics and instruments
Structured contrastive learning with observational triplets
Separate latent representations for signals and effects
🔎 Similar Papers
No similar papers found.
J
Jeroen Audenaert
Massachusetts Institute of Technology, Cambridge, MA, USA
Daniel Muthukrishna
Daniel Muthukrishna
Massachusetts Institute of Technology, University of Cambridge, University of Queensland, Australian
AstronomyCosmologyDark EnergySupernovaeMachine Learning
P
Paul F. Gregory
Massachusetts Institute of Technology, Cambridge, MA, USA
D
David W. Hogg
New York University, New York, NY, USA; Flatiron Institute, New York, NY, USA; Max-Planck-Institut für Astronomie, Heidelberg, Germany
V
V. Ashley Villar
Harvard University, Cambridge, MA, USA; The NSF AI Institute for Artificial Intelligence and Fundamental Interactions