🤖 AI Summary
To address the challenges of unobserved region completion and depth estimation uncertainty in monocular video-based 4D dynamic scene reconstruction, this paper proposes a generative Gaussian optimization framework. Methodologically: (1) A diffusion model synthesizes multi-view-consistent intermediate frames to resolve geometric and appearance ambiguities in Gaussian splatting representations; (2) A frozen “bullet-time” alignment mechanism leverages generated frames as weak supervision for iterative optimization of 4D Gaussian parameters; (3) Differentiable rendering is jointly integrated with multi-view consistency constraints to co-model static and dynamic scene components. Our approach achieves state-of-the-art performance on novel view synthesis and 2D/3D tracking tasks, significantly improving visual fidelity and geometric consistency. It establishes a new paradigm for monocular immersive content generation.
📝 Abstract
Transforming casually captured, monocular videos into fully immersive dynamic experiences is a highly ill-posed task, and comes with significant challenges, e.g., reconstructing unseen regions, and dealing with the ambiguity in monocular depth estimation. In this work we introduce BulletGen, an approach that takes advantage of generative models to correct errors and complete missing information in a Gaussian-based dynamic scene representation. This is done by aligning the output of a diffusion-based video generation model with the 4D reconstruction at a single frozen "bullet-time" step. The generated frames are then used to supervise the optimization of the 4D Gaussian model. Our method seamlessly blends generative content with both static and dynamic scene components, achieving state-of-the-art results on both novel-view synthesis, and 2D/3D tracking tasks.