🤖 AI Summary
Existing action assessment methods typically produce a single scalar score, lacking interpretable reasoning and failing to meet real-world demands for fine-grained analysis in domains such as sports science, clinical rehabilitation, and robotics. To address this, we propose a structured action assessment framework. First, we design a stepwise action reasoning mechanism that enables progressive evaluation—from holistic action recognition to fine-grained sub-action decomposition. Second, we introduce hierarchical policy learning that synergistically integrates vision-language models, domain-customized chain-of-thought prompting, and reinforcement learning–based reward optimization to support multi-level semantic modeling. Evaluated on multiple benchmark datasets, our method significantly improves both scoring accuracy and interpretability. It is the first approach to achieve high-accuracy, fine-grained human action assessment with fully transparent, traceable reasoning—thereby enabling trustworthy, decision-ready insights for critical application domains.
📝 Abstract
Evaluating human actions with clear and detailed feedback is important in areas such as sports, healthcare, and robotics, where decisions rely not only on final outcomes but also on interpretable reasoning. However, most existing methods provide only a final score without explanation or detailed analysis, limiting their practical applicability. To address this, we introduce HieroAction, a vision-language model that delivers accurate and structured assessments of human actions. HieroAction builds on two key ideas: (1) Stepwise Action Reasoning, a tailored chain of thought process designed specifically for action assessment, which guides the model to evaluate actions step by step, from overall recognition through sub action analysis to final scoring, thus enhancing interpretability and structured understanding; and (2) Hierarchical Policy Learning, a reinforcement learning strategy that enables the model to learn fine grained sub action dynamics and align them with high level action quality, thereby improving scoring precision. The reasoning pathway structures the evaluation process, while policy learning refines each stage through reward based optimization. Their integration ensures accurate and interpretable assessments, as demonstrated by superior performance across multiple benchmark datasets. Code will be released upon acceptance.