๐ค AI Summary
Sparse Principal Component Analysis (Sparse PCA) is NP-hard; existing exact algorithms suffer from prohibitive computational costs, while approximate methods face trade-offs between accuracy and efficiency. This paper proposes the first Sparse PCA framework based on block-diagonal covariance approximation: it first learns a reordering of variables and a block-diagonal approximation of the covariance matrix; then applies any existing Sparse PCA algorithm independently to each diagonal block; and finally reconstructs a global solution. The method achieves exponential speedup, admits a theoretically guaranteed error bound, and is inherently compatible with diverse baseline algorithms. Experiments on real-world datasets demonstrate that, compared to exact solvers, our approach achieves an average 100.5ร speedup with only 0.61% relative error; compared to state-of-the-art approximate methods, it attains an average 6ร speedup while improving solution qualityโreducing relative error by 0.91%.
๐ Abstract
Sparse Principal Component Analysis (Sparse PCA) is a pivotal tool in data analysis and dimensionality reduction. However, Sparse PCA is a challenging problem in both theory and practice: it is known to be NP-hard and current exact methods generally require exponential runtime. In this paper, we propose a novel framework to efficiently approximate Sparse PCA by (i) approximating the general input covariance matrix with a re-sorted block-diagonal matrix, (ii) solving the Sparse PCA sub-problem in each block, and (iii) reconstructing the solution to the original problem. Our framework is simple and powerful: it can leverage any off-the-shelf Sparse PCA algorithm and achieve significant computational speedups, with a minor additive error that is linear in the approximation error of the block-diagonal matrix. Suppose $g(k, d)$ is the runtime of an algorithm (approximately) solving Sparse PCA in dimension $d$ and with sparsity constant $k$. Our framework, when integrated with this algorithm, reduces the runtime to $mathcal{O}left(frac{d}{d^star} cdot g(k, d^star) + d^2
ight)$, where $d^star leq d$ is the largest block size of the block-diagonal matrix. For instance, integrating our framework with the Branch-and-Bound algorithm reduces the complexity from $g(k, d) = mathcal{O}(k^3cdot d^k)$ to $mathcal{O}(k^3cdot d cdot (d^star)^{k-1})$, demonstrating exponential speedups if $d^star$ is small. We perform large-scale evaluations on many real-world datasets: for exact Sparse PCA algorithm, our method achieves an average speedup factor of 100.50, while maintaining an average approximation error of 0.61%; for approximate Sparse PCA algorithm, our method achieves an average speedup factor of 6.00 and an average approximation error of -0.91%, meaning that our method oftentimes finds better solutions.