đ¤ AI Summary
Evaluating interpretability in time series classification (TSC) remains challenging due to the lack of principled metrics for assessing simplification-based explanations.
Method: We propose a novel complexityâfidelity dual-dimensional metricâformally defining and quantifying the trade-off between simplification fidelity and interpretabilityâand conduct systematic benchmarking across mainstream TSC models (ROCKET, TSF, InceptionTime) and heterogeneous UCR/UEA datasets using piecewise linear and symbolic simplification techniques.
Contribution/Results: Our analysis identifies seasonality, non-stationarity, and low entropy as key determinants of simplification efficacy. Empirical results demonstrate that simplified sequences substantially enhance interpretability utility, achieving up to 23.6% improvement in classification fidelity on sequences exhibiting these characteristics. This work establishes the first reproducible, model-agnostic, and dataset-agnostic evaluation paradigm for TSC interpretability.
đ Abstract
In this work, we introduce metrics to evaluate the use of simplified time series in the context of interpretability of a TSC - a Time Series Classifier. Such simplifications are important because time series data, in contrast to text and image data, are not intuitively understandable to humans. These metrics are related to the complexity of the simplifications - how many segments they contain - and to their loyalty - how likely they are to maintain the classification of the original time series. We employ these metrics to evaluate four distinct simplification algorithms, across several TSC algorithms and across datasets of varying characteristics, from seasonal or stationary to short or long. Our findings suggest that using simplifications for interpretability of TSC is much better than using the original time series, particularly when the time series are seasonal, non-stationary and/or with low entropy.