ST-GDance: Long-Term and Collision-Free Group Choreography from Music

📅 2025-07-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of high computational complexity and frequent inter-dancer collisions in music-driven, long-sequence, multi-dancer group dance generation. We propose a novel decoupled spatio-temporal modeling framework: a lightweight, distance-aware graph convolutional network captures spatial topology for collision avoidance, while an accelerated sparse attention mechanism models long-range temporal dependencies for motion coordination. Trained and evaluated on the AIOZ-GDance dataset, our method significantly outperforms prior approaches—enabling generation of sequences exceeding 30 seconds, scaling to over 16 dancers, reducing collision rate by 42.7%, and improving both motion naturalness and audio–motion synchronization. The efficiency and scalability of our approach make it suitable for real-time animation production in film, gaming, and other time-critical applications.

Technology Category

Application Category

📝 Abstract
Group dance generation from music has broad applications in film, gaming, and animation production. However, it requires synchronizing multiple dancers while maintaining spatial coordination. As the number of dancers and sequence length increase, this task faces higher computational complexity and a greater risk of motion collisions. Existing methods often struggle to model dense spatial-temporal interactions, leading to scalability issues and multi-dancer collisions. To address these challenges, we propose ST-GDance, a novel framework that decouples spatial and temporal dependencies to optimize long-term and collision-free group choreography. We employ lightweight graph convolutions for distance-aware spatial modeling and accelerated sparse attention for efficient temporal modeling. This design significantly reduces computational costs while ensuring smooth and collision-free interactions. Experiments on the AIOZ-GDance dataset demonstrate that ST-GDance outperforms state-of-the-art baselines, particularly in generating long and coherent group dance sequences. Project page: https://yilliajing.github.io/ST-GDance-Website/.
Problem

Research questions and friction points this paper is trying to address.

Generating synchronized group dance from music
Reducing motion collisions in multi-dancer choreography
Modeling long-term spatial-temporal interactions efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decouples spatial and temporal dependencies
Uses lightweight graph convolutions
Employs accelerated sparse attention
🔎 Similar Papers
J
Jing Xu
Department of Data Science and AI, Monash University, Melbourne, Australia
W
Weiqiang Wang
Department of Data Science and AI, Monash University, Melbourne, Australia
Cunjian Chen
Cunjian Chen
Monash University
Generative AIComputer VisionDeep Learning
J
Jun Liu
School of Computing and Communications, Lancaster University, Lancaster, UK
Qiuhong Ke
Qiuhong Ke
ARC DECRA Fellow, Senior Lecturer, Monash University
Deep LearningComputer VisionAction RecognitionVideo Understanding