🤖 AI Summary
This study addresses the high cost of randomized controlled trials (RCTs) and the challenge of leveraging large-scale observational data—often plagued by bias—for efficient experimental design. The authors propose an active residual learning paradigm that constructs a prior model from observational data and reframes causal effect estimation as the efficient inference of bias-induced residuals. They introduce the R-Design adaptive experimental design framework, whose core innovation is the R-EPIG (Residual Expected Predictive Information Gain) criterion. R-EPIG directly optimizes sample acquisition for improving causal estimators, avoiding the task-irrelevant information commonly collected by conventional approaches. Theoretical analysis demonstrates that residual-based contrastive estimation achieves faster convergence rates than full outcome reconstruction. Extensive experiments on synthetic and semi-synthetic benchmarks show significant improvements over existing methods, confirming that refining a biased model is substantially more efficient than learning from scratch.
📝 Abstract
Randomized Controlled Trials (RCTs) represent the gold standard for causal inference yet remain a scarce resource. While large-scale observational data is often available, it is utilized only for retrospective fusion, and remains discarded in prospective trial design due to bias concerns. We argue this "tabula rasa" data acquisition strategy is fundamentally inefficient. In this work, we propose Active Residual Learning, a new paradigm that leverages the observational model as a foundational prior. This approach shifts the experimental focus from learning target causal quantities from scratch to efficiently estimating the residuals required to correct observational bias. To operationalize this, we introduce the R-Design framework. Theoretically, we establish two key advantages: (1) a structural efficiency gap, proving that estimating smooth residual contrasts admits strictly faster convergence rates than reconstructing full outcomes; and (2) information efficiency, where we quantify the redundancy in standard parameter-based acquisition (e.g., BALD), demonstrating that such baselines waste budget on task-irrelevant nuisance uncertainty. We propose R-EPIG (Residual Expected Predictive Information Gain), a unified criterion that directly targets the causal estimand, minimizing residual uncertainty for estimation or clarifying decision boundaries for policy. Experiments on synthetic and semi-synthetic benchmarks demonstrate that R-Design significantly outperforms baselines, confirming that repairing a biased model is far more efficient than learning one from scratch.