🤖 AI Summary
Deep clustering remains unsuitable for safety-critical time-series applications due to its opaque decision-making process. This paper presents a systematic review of interpretable deep clustering, identifying critical gaps in handling streaming, irregularly sampled, and privacy-sensitive time series, and critically exposing the fundamental flaw of treating interpretability as an ex post-hoc add-on module. We propose six novel research directions: (1) complex-network-driven intrinsically interpretable architectures; (2) unsupervised fidelity evaluation metrics; (3) dynamic data-stream-adaptive explainers; (4) domain-specific explanation schemes; (5) human-in-the-loop clustering optimization; and (6) deep mechanistic analysis of model behavior. Grounded in autoencoder and attention-based foundations—and validated across healthcare, finance, IoT, and climate science—we establish the first comprehensive theoretical framework and technical roadmap for interpretable deep clustering on time series, advancing trustworthy temporal analytics.
📝 Abstract
Deep clustering uncovers hidden patterns and groups in complex time series data, yet its opaque decision-making limits use in safety-critical settings. This survey offers a structured overview of explainable deep clustering for time series, collecting current methods and their real-world applications. We thoroughly discuss and compare peer-reviewed and preprint papers through application domains across healthcare, finance, IoT, and climate science. Our analysis reveals that most work relies on autoencoder and attention architectures, with limited support for streaming, irregularly sampled, or privacy-preserved series, and interpretability is still primarily treated as an add-on. To push the field forward, we outline six research opportunities: (1) combining complex networks with built-in interpretability; (2) setting up clear, faithfulness-focused evaluation metrics for unsupervised explanations; (3) building explainers that adapt to live data streams; (4) crafting explanations tailored to specific domains; (5) adding human-in-the-loop methods that refine clusters and explanations together; and (6) improving our understanding of how time series clustering models work internally. By making interpretability a primary design goal rather than an afterthought, we propose the groundwork for the next generation of trustworthy deep clustering time series analytics.