🤖 AI Summary
To address the insufficient interpretability of skeleton-based human activity recognition (HAR) models, this paper proposes the first interpretability validation framework integrating SHAP attribution analysis, controllable temporal perturbation, and multi-dimensional fidelity quantification. Methodologically, it couples sliding-window CNN/LSTM architectures with skeleton-specific joint perturbation strategies and systematically evaluates explanation reliability using deletion/insertion curves and Spearman rank correlation. The key contribution lies in the first deep integration of SHAP explanations with dynamic perturbation experiments and quantitative validation—overcoming the limitations of conventional qualitative visualizations. On benchmarks including UCI HAR, SHAP-based explanations achieve an average ΔAUC < 0.08, significantly outperforming LIME and Grad-CAM. This establishes a verifiable interpretability paradigm for trustworthy HAR model deployment in high-stakes applications.