🤖 AI Summary
This paper addresses the substantial bias and limited robustness in causal effect estimation from panel data. We propose the Triple-Robust Panel (TROP) estimator, which integrates three novel mechanisms: (1) low-rank factor modeling of potential outcomes to enhance model flexibility; (2) dynamic weighting of similar treated units; and (3) dynamic weighting of temporally proximate treatment periods. This design jointly accommodates unit-level heterogeneity and temporal dynamics, substantially improving estimation robustness. Across multiple realistic simulation settings—spanning sparse treatments, time-varying confounding, and latent factor structures—TROP consistently outperforms leading methods—including Difference-in-Differences, Synthetic Control, Matrix Completion, and Synthetic DID—in terms of bias reduction, variance control, and confidence interval coverage. Theoretical guarantees support its triple-robustness property: consistency holds if any two of the three underlying models (outcome, unit similarity, timing similarity) are correctly specified. TROP thus provides a theoretically rigorous and empirically effective tool for high-dimensional panel causal inference.
📝 Abstract
This paper studies estimation of causal effects in a panel data setting. We introduce a new estimator, the Triply RObust Panel (TROP) estimator, that combines $(i)$ a flexible model for the potential outcomes based on a low-rank factor structure on top of a two-way-fixed effect specification, with $(ii)$ unit weights intended to upweight units similar to the treated units and $(iii)$ time weights intended to upweight time periods close to the treated time periods. We study the performance of the estimator in a set of simulations designed to closely match several commonly studied real data sets. We find that there is substantial variation in the performance of the estimators across the settings considered. The proposed estimator outperforms two-way-fixed-effect/difference-in-differences, synthetic control, matrix completion and synthetic-difference-in-differences estimators. We investigate what features of the data generating process lead to this performance, and assess the relative importance of the three components of the proposed estimator. We have two recommendations. Our preferred strategy is that researchers use simulations closely matched to the data they are interested in, along the lines discussed in this paper, to investigate which estimators work well in their particular setting. A simpler approach is to use more robust estimators such as synthetic difference-in-differences or the new triply robust panel estimator which we find to substantially outperform two-way fixed effect estimators in many empirically relevant settings.