DFM: Deep Fourier Mimic for Expressive Dance Motion Learning

📅 2025-02-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Entertainment robots exhibit rigid, non-interactive dance motions due to reliance on pre-programmed, periodic motion templates lacking dynamic responsiveness. Method: We propose a natural dance motion learning framework for multi-task coordination. Departing from conventional periodic constraints, we introduce a flexible Fourier motion representation—enabling the first deep integration of frequency-domain modeling with Proximal Policy Optimization (PPO) reinforcement learning. We further design a joint deep Fourier encoder–multi-task controller architecture that jointly optimizes dance generation, bipedal locomotion, and gaze control in both time and frequency domains. Contribution/Results: Compared to scripted playback baselines, our approach significantly improves motion tracking accuracy and expressive fidelity. It enables high-fidelity, real-time dance generation with synchronized locomotion and gaze control on physical robots—establishing a novel paradigm for modeling natural, embodied dynamic behaviors in intelligent agents.

Technology Category

Application Category

📝 Abstract
As entertainment robots gain popularity, the demand for natural and expressive motion, particularly in dancing, continues to rise. Traditionally, dancing motions have been manually designed by artists, a process that is both labor-intensive and restricted to simple motion playback, lacking the flexibility to incorporate additional tasks such as locomotion or gaze control during dancing. To overcome these challenges, we introduce Deep Fourier Mimic (DFM), a novel method that combines advanced motion representation with Reinforcement Learning (RL) to enable smooth transitions between motions while concurrently managing auxiliary tasks during dance sequences. While previous frequency domain based motion representations have successfully encoded dance motions into latent parameters, they often impose overly rigid periodic assumptions at the local level, resulting in reduced tracking accuracy and motion expressiveness, which is a critical aspect for entertainment robots. By relaxing these locally periodic constraints, our approach not only enhances tracking precision but also facilitates smooth transitions between different motions. Furthermore, the learned RL policy that supports simultaneous base activities, such as locomotion and gaze control, allows entertainment robots to engage more dynamically and interactively with users rather than merely replaying static, pre-designed dance routines.
Problem

Research questions and friction points this paper is trying to address.

Enhance dance motion expressiveness in robots
Enable smooth transitions between dance motions
Support auxiliary tasks during robot dancing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep Fourier Mimic method
Reinforcement Learning integration
Enhanced motion expressiveness
🔎 Similar Papers
No similar papers found.