🤖 AI Summary
To address the high computational complexity—currently optimal at 𝒪(nm log(nm))—of projecting onto the matrix ℓ₁,∞ norm, this paper proposes the first linear-time 𝒪(nm) two-level ℓ₁,∞ projection algorithm, specifically designed for structured sparsification in neural networks. Methodologically, we (1) establish and rigorously prove a key ℓ₁,∞ norm identity, providing the theoretical foundation for linear-time computation; (2) devise a two-level optimization framework that tightly couples exact projection with structured sparsity regularization; and (3) evaluate the algorithm on autoencoder pruning tasks. Results show a 2.5× speedup over the state-of-the-art method, achieving the highest sparsity rate without sacrificing classification accuracy—thereby jointly optimizing sparsity and generalization performance.
📝 Abstract
The $ell_{1,infty}$ norm is an efficient-structured projection, but the complexity of the best algorithm is, unfortunately, $mathcal{O}ig(n m log(n m)ig)$ for a matrix $n imes m$.\ In this paper, we propose a new bi-level projection method, for which we show that the time complexity for the $ell_{1,infty}$ norm is only $mathcal{O}ig(n m ig)$ for a matrix $n imes m$. Moreover, we provide a new $ell_{1,infty}$ identity with mathematical proof and experimental validation. Experiments show that our bi-level $ell_{1,infty}$ projection is $2.5$ times faster than the actual fastest algorithm and provides the best sparsity while keeping the same accuracy in classification applications.