How to Achieve the Intended Aim of Deep Clustering Now, without Deep Learning

📅 2026-02-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses a fundamental limitation of traditional k-means clustering, which struggles to identify clusters of arbitrary shape, scale, and density. While existing deep clustering methods, such as Deep Embedded Clustering (DEC), incorporate deep representations, they generally overlook the intrinsic distributional characteristics of data clusters, thereby failing to overcome this core limitation. The paper is the first to explicitly identify this critical shortcoming and proposes a novel non-deep learning paradigm that explicitly models the distributional properties of clusters to effectively capture complex structures. Extensive experiments on multiple benchmark datasets demonstrate that the proposed method significantly outperforms state-of-the-art deep clustering algorithms, successfully overcoming the inherent constraints of k-means and revealing that deep representations are not a necessary condition for addressing this challenge.

Technology Category

Application Category

📝 Abstract
Deep clustering (DC) is often quoted to have a key advantage over $k$-means clustering. Yet, this advantage is often demonstrated using image datasets only, and it is unclear whether it addresses the fundamental limitations of $k$-means clustering. Deep Embedded Clustering (DEC) learns a latent representation via an autoencoder and performs clustering based on a $k$-means-like procedure, while the optimization is conducted in an end-to-end manner. This paper investigates whether the deep-learned representation has enabled DEC to overcome the known fundamental limitations of $k$-means clustering, i.e., its inability to discover clusters of arbitrary shapes, varied sizes and densities. Our investigations on DEC have a wider implication on deep clustering methods in general. Notably, none of these methods exploit the underlying data distribution. We uncover that a non-deep learning approach achieves the intended aim of deep clustering by making use of distributional information of clusters in a dataset to effectively address these fundamental limitations.
Problem

Research questions and friction points this paper is trying to address.

deep clustering
k-means limitations
cluster shape
data distribution
arbitrary clusters
Innovation

Methods, ideas, or system contributions that make the work stand out.

deep clustering
k-means limitations
distributional information
non-deep learning
cluster shape
🔎 Similar Papers
2024-09-01arXiv.orgCitations: 4
Kai Ming Ting
Kai Ming Ting
Nanjing University
Machine LearningData Mining
Weijie Xu
Weijie Xu
Amazon
LLM FairnessLLM EvaluationTabular LLM
H
Hang Zhang
State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, China; School of Artificial Intelligence, Nanjing University, Nanjing, China