๐ค AI Summary
This work addresses the challenges of long-horizon historical motion modeling, continuous text input, and strict real-time requirements in text-conditioned streaming human pose generation. We propose a diffusion-enhanced autoregressive modeling framework operating in a continuous causal latent space. Unlike diffusion models constrained by fixed-length horizons and GPT-style approaches suffering from latency and error accumulation due to discrete, non-causal tokenization, our method introduces a learnable motion encoder and a streaming decoder that jointly integrate diffusion priors with causal autoregressive dynamics in a continuous latent spaceโenabling millisecond-level low-latency, multi-turn interactive generation. Experiments demonstrate significant improvements over state-of-the-art methods across multiple benchmarks, supporting stable generation of over 1,000 frames and dynamic cross-semantic instruction editing.
๐ Abstract
This paper addresses the challenge of text-conditioned streaming motion generation, which requires us to predict the next-step human pose based on variable-length historical motions and incoming texts. Existing methods struggle to achieve streaming motion generation, e.g., diffusion models are constrained by pre-defined motion lengths, while GPT-based methods suffer from delayed response and error accumulation problem due to discretized non-causal tokenization. To solve these problems, we propose MotionStreamer, a novel framework that incorporates a continuous causal latent space into a probabilistic autoregressive model. The continuous latents mitigate information loss caused by discretization and effectively reduce error accumulation during long-term autoregressive generation. In addition, by establishing temporal causal dependencies between current and historical motion latents, our model fully utilizes the available information to achieve accurate online motion decoding. Experiments show that our method outperforms existing approaches while offering more applications, including multi-round generation, long-term generation, and dynamic motion composition. Project Page: https://zju3dv.github.io/MotionStreamer/