🤖 AI Summary
In online policy learning, adaptive experimentation induces data dependence, violating the i.i.d. assumption and invalidating classical concentration inequalities—thus undermining standard statistical guarantees. To address this, we develop a novel self-normalized martingale empirical process maximal inequality and construct the first variance-regularized pessimistic policy learning framework applicable to general dependent data. Crucially, we introduce adaptive sample-variance penalization into pessimistic optimization—enabling fast convergence under sequential updates. Theoretically, we establish an excess risk bound that strictly improves upon the standard $1/sqrt{n}$ rate in both parametric and nonparametric settings. Numerical experiments confirm substantial gains in both convergence speed and policy performance over existing methods.
📝 Abstract
Adaptive experiments produce dependent data that break i.i.d. assumptions that underlie classical concentration bounds and invalidate standard learning guarantees. In this paper, we develop a self-normalized maximal inequality for martingale empirical processes. Building on this, we first propose an adaptive sample-variance penalization procedure which balances empirical loss and sample variance, valid for general dependent data. Next, this allows us to derive a new variance-regularized pessimistic off-policy learning objective, for which we establish excess-risk guarantees. Subsequently, we show that, when combined with sequential updates and under standard complexity and margin conditions, the resulting estimator achieves fast convergence rates in both parametric and nonparametric regimes, improving over the usual $1/sqrt{n}$
baseline. We complement our theoretical findings with numerical simulations that illustrate the practical gains of our approach.