Towards Motion Turing Test: Evaluating Human-Likeness in Humanoid Robots

📅 2026-03-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes the "Action Turing Test" framework to objectively evaluate the human-likeness of humanoid robot motions—specifically, whether human observers can distinguish the source of an action based solely on kinematic information. To facilitate this assessment, the authors introduce HHMotion, the first large-scale, standardized dataset comprising 1,000 motion sequences from both humans and robots, all unified in the SMPL-X representation, along with human-likeness ratings provided by 30 annotators. Experimental results demonstrate that current large language models exhibit limited performance on this task, whereas the proposed lightweight baseline model achieves significantly better results. The dataset, code, and benchmark will be publicly released to advance research in humanoid robot motion evaluation.

Technology Category

Application Category

📝 Abstract
Humanoid robots have achieved significant progress in motion generation and control, exhibiting movements that appear increasingly natural and human-like. Inspired by the Turing Test, we propose the Motion Turing Test, a framework that evaluates whether human observers can discriminate between humanoid robot and human poses using only kinematic information. To facilitate this evaluation, we present the Human-Humanoid Motion (HHMotion) dataset, which consists of 1,000 motion sequences spanning 15 action categories, performed by 11 humanoid models and 10 human subjects. All motion sequences are converted into SMPL-X representations to eliminate the influence of visual appearance. We recruited 30 annotators to rate the human-likeness of each pose on a 0-5 scale, resulting in over 500 hours of annotation. Analysis of the collected data reveals that humanoid motions still exhibit noticeable deviations from human movements, particularly in dynamic actions such as jumping, boxing, and running. Building on HHMotion, we formulate a human-likeness evaluation task that aims to automatically predict human-likeness scores from motion data. Despite recent progress in multimodal large language models, we find that they remain inadequate for assessing motion human-likeness. To address this, we propose a simple baseline model and demonstrate that it outperforms several contemporary LLM-based methods. The dataset, code, and benchmark will be publicly released to support future research in the community.
Problem

Research questions and friction points this paper is trying to address.

Motion Turing Test
human-likeness
humanoid robots
motion evaluation
kinematic information
Innovation

Methods, ideas, or system contributions that make the work stand out.

Motion Turing Test
Humanoid Motion Evaluation
HHMotion Dataset
SMPL-X Representation
Human-Likeness Prediction
🔎 Similar Papers
No similar papers found.
M
Mingzhe Li
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
M
Mengyin Liu
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
Z
Zekai Wu
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
X
Xincheng Lin
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
J
Junsheng Zhang
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
Ming Yan
Ming Yan
Xiamen University
Human Pose EstimationComputer VisionLiDARPoint CloudImbalanced Data
Z
Zengye Xie
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
C
Changwang Zhang
OPPO Research Institute
Chenglu Wen
Chenglu Wen
Professor of Xiamen University
3D visionpoint cloudsmobile mappingrobotics
L
Lan Xu
ShanghaiTech University
Siqi Shen
Siqi Shen
Xiamen University
Reinforcement Learning3D Vision
C
Cheng Wang
Fujian Key Laboratory of Urban Intelligent Sensing and Computing, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University