🤖 AI Summary
To address the challenge of diagnosing performance degradation caused by concept drift in dynamic environments, this paper proposes an explainable analytics framework based on temporal evolution of group-level counterfactual explanations. The method innovatively employs evolving cluster centroids and action vectors of group counterfactuals as interpretable proxies for shifts in decision logic, establishing a three-tiered, co-adaptive mechanism linking data, model, and explanation. It integrates counterfactual generation, clustering analysis, distribution shift detection, and prediction inconsistency measurement to enable fine-grained, multi-signal collaborative drift attribution. Experiments demonstrate that the approach effectively distinguishes canonical drift types—including spatial drift and concept re-labeling—while significantly improving explanation transparency and diagnostic accuracy. This work introduces the first traceable and attributable explanatory paradigm for mechanistic analysis of concept drift.
📝 Abstract
Machine learning models in dynamic environments often suffer from concept drift, where changes in the data distribution degrade performance. While detecting this drift is a well-studied topic, explaining how and why the model's decision-making logic changes still remains a significant challenge. In this paper, we introduce a novel methodology to explain concept drift by analyzing the temporal evolution of group-based counterfactual explanations (GCEs). Our approach tracks shifts in the GCEs' cluster centroids and their associated counterfactual action vectors before and after a drift. These evolving GCEs act as an interpretable proxy, revealing structural changes in the model's decision boundary and its underlying rationale. We operationalize this analysis within a three-layer framework that synergistically combines insights from the data layer (distributional shifts), the model layer (prediction disagreement), and our proposed explanation layer. We show that such holistic view allows for a more comprehensive diagnosis of drift, making it possible to distinguish between different root causes, such as a spatial data shift versus a re-labeling of concepts.