🤖 AI Summary
To address the inherent trade-off among high computational cost, excessive bandwidth demand, and degraded rendering quality in long-duration dynamic scene novel view synthesis, this paper proposes the first 4D Gaussian-based dynamic reconstruction framework tailored for streaming video. Methodologically, we innovatively integrate Low-Rank Adaptation (LoRA) into 3D Gaussian Splatting to construct an incrementally updatable low-rank deformation model; further, we introduce block-wise spatiotemporal alignment and online optimization to enable continual learning and frame-level incremental reconstruction over arbitrarily long video sequences. Compared to state-of-the-art offline methods, our framework achieves comparable rendering fidelity (PSNR/SSIM), while reducing streaming bandwidth by 90% and substantially lowering GPU memory consumption and computational overhead. This provides a scalable solution for real-time, high-fidelity, long-duration dynamic novel view synthesis.
📝 Abstract
Recent methods have made significant progress in synthesizing novel views with long video sequences. This paper proposes a highly scalable method for dynamic novel view synthesis with continual learning. We leverage the 3D Gaussians to represent the scene and a low-rank adaptation-based deformation model to capture the dynamic scene changes. Our method continuously reconstructs the dynamics with chunks of video frames, reduces the streaming bandwidth by $90%$ while maintaining high rendering quality comparable to the off-line SOTA methods.