Empirical Bayes Estimation and Inference via Smooth Nonparametric Maximum Likelihood

📅 2026-03-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges in nonparametric maximum likelihood estimation (NPMLE) within empirical Bayes frameworks—namely, the difficulty of uncertainty quantification and the logarithmic deconvolution convergence bottleneck arising from discrete mixture representations. To overcome these limitations, the authors introduce a Gaussian location mixture model with smoothness constraints, yielding a smoothed NPMLE formulation that is cast as a convex optimization problem. The proposed method achieves, for the first time, polynomial-rate deconvolution convergence and is shown to be asymptotically minimax optimal over the corresponding function class. The resulting plug-in marginal confidence sets attain near-oracle expected length while providing asymptotically exact coverage. The theoretical analysis encompasses KL projection, weighted total variation convergence, and extensions to heteroscedastic Gaussian observations, and further demonstrates that under model misspecification, the estimator still converges to the pseudo-true posterior at a near-parametric rate.
📝 Abstract
The empirical Bayes $g$-modeling approach via the nonparametric maximum likelihood estimator (NPMLE) is widely used for large-scale estimation and inference in the normal means problem, yet theoretical guarantees for uncertainty quantification remain scarce. A key obstacle is that the NPMLE of the mixing distribution is necessarily discrete, which yields discrete posterior credible sets and a deconvolution rate that is logarithmic. We address both limitations by studying a hierarchical Gaussian smoothing layer that restricts the mixing distribution to a Gaussian location mixture. The resulting smooth NPMLE is computed by solving a convex optimization problem and inherits the near-parametric denoising performance of the classical NPMLE. For deconvolution it achieves a polynomial rate of convergence which we show is asymptotically minimax over the corresponding class. The estimated smooth posteriors converge to the true posteriors at the same polynomial rate in weighted total variation distance. When the model is misspecified, the smooth NPMLE converges to the Kullback-Leibler projection of the true marginal density onto the model class at a nearly parametric rate, and the polynomial deconvolution and posterior convergence rates carry over to this pseudo-true target. Building on this smooth posterior, we characterize optimal marginal coverage sets: the shortest set-valued rules achieving a prescribed marginal coverage probability. Plug-in empirical Bayes marginal coverage sets based on the smooth NPMLE achieve asymptotically exact coverage at a polynomial rate and converge to the oracle optimal set in expected length. All results extend to heteroscedastic Gaussian observations. We also study identifiability of the proposed model and show that the largest Gaussian component of the prior is identifiable, and provide a consistent estimator and a finite-sample upper confidence bound for it.
Problem

Research questions and friction points this paper is trying to address.

empirical Bayes
nonparametric maximum likelihood
uncertainty quantification
deconvolution
posterior inference
Innovation

Methods, ideas, or system contributions that make the work stand out.

smooth NPMLE
empirical Bayes
deconvolution rate
posterior convergence
marginal coverage sets
🔎 Similar Papers
No similar papers found.