Popcorn: Accelerating Kernel K-means on GPUs through Sparse Linear Algebra

📅 2025-01-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Kernel K-means suffers from O(n²) computational and memory complexity in kernel matrix construction, rendering it impractical for large-scale datasets. To address this, we reformulate Kernel K-means as a sparse linear algebra problem—its first such formulation—and introduce a GPU-accelerated paradigm based on sparse matrix–matrix (SpMM) and sparse matrix–vector (SpMV) operations. By leveraging the kernel trick for implicit feature mapping and representing key computations via sparse matrices, our approach circumvents dense kernel matrix materialization, thereby reducing both programming complexity and memory footprint. We release the first open-source GPU implementation of Kernel K-means, achieving second-scale clustering on medium-sized datasets: up to 123.8× speedup over CPU-based implementations and 2.6× over non-sparse GPU variants. Our core contribution lies in transforming high-complexity nonlinear clustering into an efficiently parallelizable sparse computation task, establishing a practical pathway for scalable kernel-based clustering.

Technology Category

Application Category

📝 Abstract
K-means is a popular clustering algorithm with significant applications in numerous scientific and engineering areas. One drawback of K-means is its inability to identify non-linearly separable clusters, which may lead to inaccurate solutions in certain cases. Kernel K-means is a variant of classical K-means that can find non-linearly separable clusters. However, it scales quadratically with respect to the size of the dataset, taking several minutes to cluster even medium-sized datasets on traditional CPU-based machines. In this paper, we present a formulation of Kernel K-means using sparse-dense matrix multiplication (SpMM) and sparse matrix-vector multiplication (SpMV), and we show that our formulation enables the rapid implementation of a fast GPU-based version of Kernel K-means with little programming effort. Our implementation, named Popcorn, is the first open-source GPU-based implementation of Kernel K-means. Popcorn achieves a speedup of up to 123.8x over a CPU implementation of Kernel K-means and a speedup of up to 2.6x over a GPU implementation of Kernel K-means that does not use sparse matrix computations. Our results support the effectiveness of sparse matrices as tools for efficient parallel programming.
Problem

Research questions and friction points this paper is trying to address.

Kernel K-means
large-scale data
computational efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Kernel K-means acceleration
Sparse linear algebra operations
GPU-based computation
🔎 Similar Papers
No similar papers found.