Probabilistic Iterative Hard Thresholding for Sparse Learning

๐Ÿ“… 2024-09-02
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
For high-dimensional, small-sample learning problems involving โ„“โ‚€-constrained expected risk minimization, existing stochastic sparse optimization methods struggle to guarantee both convergence and accurate sparse recovery under noisy gradient estimates. To address this, we propose the Probabilistic Iterative Hard Thresholding (PIHT) algorithm, which integrates stochastic optimization, expected risk minimization, and a provably falsifiable probabilistic thresholding update rule. We establish, for the first time, a rigorous convergence theory for โ„“โ‚€-constrained stochastic processes under noisy gradientsโ€”ensuring robust convergence in large-scale, noisy-data regimes. Experiments on sparse linear regression and 1-bit compressed sensing demonstrate that PIHT significantly outperforms state-of-the-art stochastic sparse methods, achieving superior support recovery accuracy and enhanced computational stability.

Technology Category

Application Category

๐Ÿ“ Abstract
For statistical modeling wherein the data regime is unfavorable in terms of dimensionality relative to the sample size, finding hidden sparsity in the ground truth can be critical in formulating an accurate statistical model. The so-called"l0 norm"which counts the number of non-zero components in a vector, is a strong reliable mechanism of enforcing sparsity when incorporated into an optimization problem. However, in big data settings wherein noisy estimates of the gradient must be evaluated out of computational necessity, the literature is scant on methods that reliably converge. In this paper we present an approach towards solving expectation objective optimization problems with cardinality constraints. We prove convergence of the underlying stochastic process, and demonstrate the performance on two Machine Learning problems.
Problem

Research questions and friction points this paper is trying to address.

Finding hidden sparsity in high-dimensional data
Optimizing l0 norm for sparse model accuracy
Ensuring convergence in noisy big data settings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Probabilistic Iterative Hard Thresholding method
Sparse learning with l0 norm
Convergence in noisy gradient settings
M
Matteo Bergamaschi
Department of Mathematics "Tullio Levi-Civita", University of Padua
A
A. Cristofari
Department of Civil Engineering and Computer Science Engineering, University of Rome "Tor Vergata"
Vyacheslav Kungurtsev
Vyacheslav Kungurtsev
Czech Technical University in Prague
Francesco Rinaldi
Francesco Rinaldi
Professor, University of Padova
Nonlinear OptimizationApplied MathematicsOperations ResearchMachine Learning