🤖 AI Summary
Large language models (LLMs) struggle to effectively model macroeconomic dynamics under small-sample regimes. Method: This paper proposes a novel paradigm—“theory-driven synthetic data + temporal large models”—where million-scale, theory-consistent synthetic panel data are generated via dynamic stochastic general equilibrium (DSGE) modeling and Bayesian posterior sampling, then used to train a temporal Transformer architecture. Contribution/Results: The approach synergistically integrates the structural rigor of macroeconomic theory with the representational power of LLMs, enabling the model to learn a generalizable “macroeconomic language.” Empirical results demonstrate that the resulting hybrid predictor significantly outperforms both purely statistical models and conventional DSGE models in forward-looking forecasts through 2025, achieving, for the first time, a unified balance between theoretical interpretability and data-driven predictive accuracy.
📝 Abstract
We show how state-of-the-art large language models (LLMs), seemingly inapplicable to the small samples typical of macroeconomics, can be trained to learn the language of macroeconomy. We estimate a large-scale dynamic stochastic general equilibrium (DSGE) model on an initial segment of the data and obtain a posterior distribution over structural parameters. We sample from this posterior to generate millions of theory-consistent synthetic panels that, when mixed with actual macroeconomic data, form the training corpus for a time-series transformer with attention. The trained model is then used to forecast out-of-sample through 2025. The results show that this hybrid forecaster, which combines the theoretical coherence of DSGE models with the representational power of modern LLMs, successfully learns the macroeconomic language.