π€ AI Summary
This paper addresses the challenge of simultaneously ensuring physical feasibility and perceptual realism in human motion generation evaluation. To this end, we propose PP-Motionβthe first end-to-end assessment framework that jointly incorporates physical priors and human perceptual modeling. Methodologically, we construct objective ground truth via physics-based annotation and optimize a dual-constraint loss comprising Pearson correlation loss (to enforce dynamical consistency) and perceptual fidelity loss (trained on human motion discrimination). Our key contributions are: (i) introducing minimal physical correction magnitude as a continuous, fine-grained evaluation metric; and (ii) unifying physical plausibility and subjective realism within a single differentiable framework. Experiments demonstrate that PP-Motion significantly outperforms existing metrics in physical consistency and achieves strong correlation (r > 0.92) with human perceptual quality scores. Thus, PP-Motion establishes a novel, interpretable, and differentiable paradigm for assessing generated motion quality.
π Abstract
Human motion generation has found widespread applications in AR/VR, film, sports, and medical rehabilitation, offering a cost-effective alternative to traditional motion capture systems. However, evaluating the fidelity of such generated motions is a crucial, multifaceted task. Although previous approaches have attempted at motion fidelity evaluation using human perception or physical constraints, there remains an inherent gap between human-perceived fidelity and physical feasibility. Moreover, the subjective and coarse binary labeling of human perception further undermines the development of a robust data-driven metric. We address these issues by introducing a physical labeling method. This method evaluates motion fidelity by calculating the minimum modifications needed for a motion to align with physical laws. With this approach, we are able to produce fine-grained, continuous physical alignment annotations that serve as objective ground truth. With these annotations, we propose PP-Motion, a novel data-driven metric to evaluate both physical and perceptual fidelity of human motion. To effectively capture underlying physical priors, we employ Pearson's correlation loss for the training of our metric. Additionally, by incorporating a human-based perceptual fidelity loss, our metric can capture fidelity that simultaneously considers both human perception and physical alignment. Experimental results demonstrate that our metric, PP-Motion, not only aligns with physical laws but also aligns better with human perception of motion fidelity than previous work.