🤖 AI Summary
Policy interventions often suffer from poor cross-environment generalizability due to two challenges: erroneous generalization arising from effect heterogeneity and “unknown” risks stemming from insufficient evidence. To address these, we propose a generalizable prototype discovery framework that—novelty—incorporates an “acknowledge ignorance” mechanism into prototype learning, jointly optimizing prediction and rejection boundaries. Our method integrates decision-theoretic principles, Bayesian modeling, and frequentist inference, yielding finite-sample regret bounds and asymptotic statistical guarantees. Evaluated on reanalyses of anti-poverty programs across six countries, the framework substantially outperforms forced pooling approaches: it automatically identifies robust, reusable effect prototypes while precisely flagging critical “unknown” contexts where evidence is weak and additional data collection is warranted. This advances evidence-based policymaking by delivering a paradigm that balances statistical reliability with interpretability.
📝 Abstract
When studying policy interventions, researchers are often interested in two related goals: i) learning for which types of individuals the program has the largest effects (heterogeneity) and ii) understanding whether those patterns of treatment effects have predictive power across environments (generalizability). To that end, we develop a framework to learn from the data how to partition observations into groups of individual and environmental characteristics whose effects are generalizable for others - a set of generalizable archetypes. Our view is that implicit in the task of archetypal discovery is detecting those contexts where effects do not generalize and where researchers should collect more evidence before drawing inference on treatment effects. We introduce a method that jointly estimates when and how a prediction can be formed and when, instead, researchers should admit ignorance and elicit further evidence before making predictions. We provide both a decision-theoretic and Bayesian foundation of our procedure. We derive finite-sample (frequentist) regret guarantees, asymptotic theory for inference, and discuss computational properties. We illustrate the benefits of our procedure over existing alternatives that would fail to admit ignorance and force pooling across all units by re-analyzing a multifaceted program targeted towards the poor across six different countries.