🤖 AI Summary
In small-sample experiments, noisy treatment effect estimates can lead to asymmetric losses in policy generalization and downstream operational decisions. To address this, this work proposes the PATRO method, which retains standard effect estimation while introducing data-agnostic dual adjustments—one for generalization decisions and another for downstream optimization—to minimize Bayes risk. This study is the first to decouple these two adjustments and systematically analyze their complementary or substitutive relationship, yielding a concise, transparent, and approximately Bayes-optimal decision framework. The adjustment parameters are solved via an alternating iterative algorithm that integrates Bayesian decision theory with a plug-in estimation framework. Both theoretical analysis and empirical results demonstrate that PATRO achieves performance close to or matching the Bayes optimum, significantly outperforming conventional approaches that directly plug point estimates into decision pipelines.
📝 Abstract
Firms increasingly use randomized experiments to decide whether to scale up an intervention and, if so, how to re-optimize related operational choices such as inventory, capacity, or pricing. In many settings, experiments are performed on small samples, so the estimated effect of the intervention is uncertain. A common practice is to plug a'significant'estimate of the effect into both (i) the rollout rule and (ii) the downstream optimization. However, this can lead to avoidable losses because the costs of over- versus under-estimating the effect are often asymmetric. The technically ideal approach is to obtain a data-dependent decision rule that minimizes the Bayes risk, but this lacks transparency and requires more computations. We propose Predict-Adjust-Then-Rollout-Optimize (PATRO), a plug-in approach that keeps the standard estimate, but makes data-independent adjustments, respectively, for the two types of decision. We show that the two adjustments can be substitutes or complements and provide an alternating-iteration method to compute the pair. PATRO performs both in theory and numerically close or equivalent to the Bayes-optimal benchmark, making it a simple, effective way to convert noisy experimental results into better rollout and operational decisions.