🤖 AI Summary
This study addresses the limitations of traditional factor models, which rely on linear assumptions and homogeneous sampling frequencies, rendering them ill-suited for capturing nonlinear dynamics in high-dimensional mixed-frequency data. To overcome these constraints, the authors propose the Mixed-frequency Panel Transformer Encoder (MPTE), which integrates attention mechanisms into factor modeling. By jointly attending to both temporal and cross-sectional dimensions, MPTE enables context-aware, adaptive signal extraction across frequencies, unifying the treatment of linear and nonlinear relationships. The method extends the generalized principal component analysis framework and is supported by asymptotic theoretical guarantees. Empirical evaluations demonstrate that MPTE significantly outperforms leading benchmarks in simulations and across 13 macroeconomic forecasting tasks, while also effectively identifying influential variables and critical time windows.
📝 Abstract
We propose Mixed-Panels-Transformer Encoder (MPTE), a novel framework for estimating factor models in panel datasets with mixed frequencies and nonlinear signals. Traditional factor models rely on linear signal extraction and require homogeneous sampling frequencies, limiting their applicability to modern high-dimensional datasets where variables are observed at different temporal resolutions. Our approach leverages Transformer-style attention mechanisms to enable context-aware signal construction through flexible, data-dependent weighting schemes that replace fixed linear combinations with adaptive reweighting based on similarity and relevance. We extend classical principal component analysis (PCA) to accommodate general temporal and cross-sectional attention matrices, allowing the model to learn how to aggregate information across frequencies without manual alignment or pre-specified weights. For linear activation functions, we establish consistency and asymptotic normality of factor and loading estimators, showing that our framework nests Target PCA as a special case while providing efficiency gains through transfer learning across auxiliary datasets. The nonlinear extension uses a Transformer architecture to capture complex hierarchical interactions while preserving the theoretical foundations. In simulations, MPTE demonstrates superior performance in nonlinear environments, and in an empirical application to 13 macroeconomic forecasting targets using a selected set of 48 monthly and quarterly series from the FRED-MD and FRED-QD databases, our method achieves competitive performance against established benchmarks. We further analyze attention patterns and systematically ablate model components to assess variable importance and temporal dependence. The resulting patterns highlight which indicators and horizons are most influential for forecasting.