Generative AI-Driven High-Fidelity Human Motion Simulation

📅 2025-07-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current industrial human motion simulation methods suffer from low fidelity, hindering accurate assessment of operational behavior, safety, and efficiency. To address this, we propose G-AI-HMS—a novel framework integrating large language models (LLMs) with text-to-motion generation to enable end-to-end translation of natural-language task descriptions into high-fidelity 3D motion sequences. Our approach innovatively incorporates MotionGPT for semantic alignment and employs vision-driven pose estimation coupled with spatiotemporal motion similarity metrics—based on geometric and temporal consistency analysis of joint keypoints—for rigorous realism validation. Experiments across eight representative industrial tasks demonstrate statistically significant improvements (p < 0.0001) over manually authored motion descriptions: joint position error is reduced by 32.7%, and temporal misalignment decreases by 41.5%, with marked gains in spatial accuracy, pose alignment, and dynamic temporal fidelity. G-AI-HMS establishes a new paradigm for cost-effective, high-credibility human factors evaluation in industrial settings.

Technology Category

Application Category

📝 Abstract
Human motion simulation (HMS) supports cost-effective evaluation of worker behavior, safety, and productivity in industrial tasks. However, existing methods often suffer from low motion fidelity. This study introduces Generative-AI-Enabled HMS (G-AI-HMS), which integrates text-to-text and text-to-motion models to enhance simulation quality for physical tasks. G-AI-HMS tackles two key challenges: (1) translating task descriptions into motion-aware language using Large Language Models aligned with MotionGPT's training vocabulary, and (2) validating AI-enhanced motions against real human movements using computer vision. Posture estimation algorithms are applied to real-time videos to extract joint landmarks, and motion similarity metrics are used to compare them with AI-enhanced sequences. In a case study involving eight tasks, the AI-enhanced motions showed lower error than human created descriptions in most scenarios, performing better in six tasks based on spatial accuracy, four tasks based on alignment after pose normalization, and seven tasks based on overall temporal similarity. Statistical analysis showed that AI-enhanced prompts significantly (p $<$ 0.0001) reduced joint error and temporal misalignment while retaining comparable posture accuracy.
Problem

Research questions and friction points this paper is trying to address.

Enhancing human motion simulation fidelity using Generative AI
Translating task descriptions into motion-aware language with LLMs
Validating AI-enhanced motions against real human movements
Innovation

Methods, ideas, or system contributions that make the work stand out.

Generative AI integrates text-to-text and text-to-motion models
Large Language Models align with MotionGPT for motion translation
Computer vision validates AI motions against real human movements
🔎 Similar Papers
2024-09-05European Conference on Computer VisionCitations: 4