🤖 AI Summary
To address temporal flickering and artifacts in online dynamic 3D Gaussian splatting reconstruction caused by sensor noise and observation inconsistency, this paper proposes the first observation error decoupling and compensation framework. It models input observations as the sum of an ideal scene signal and a learnable error term, compensated in real time via a lightweight error modeling network. Integrated into the online optimization pipeline of dynamic Gaussian splatting and augmented with spatiotemporal regularization losses, the method enables incremental reconstruction from streaming video. Evaluated on multiple dynamic datasets, it achieves a 1.8 dB PSNR gain, a 0.04 reduction in LPIPS, and a 62% decrease in temporal jitter over static regions—demonstrating substantial improvements in both rendering quality and temporal consistency over state-of-the-art online methods.
📝 Abstract
Online reconstruction of dynamic scenes is significant as it enables learning scenes from live-streaming video inputs, while existing offline dynamic reconstruction methods rely on recorded video inputs. However, previous online reconstruction approaches have primarily focused on efficiency and rendering quality, overlooking the temporal consistency of their results, which often contain noticeable artifacts in static regions. This paper identifies that errors such as noise in real-world recordings affect temporal inconsistency in online reconstruction. We propose a method that enhances temporal consistency in online reconstruction from observations with temporal inconsistency which is inevitable in cameras. We show that our method restores the ideal observation by subtracting the learned error. We demonstrate that applying our method to various baselines significantly enhances both temporal consistency and rendering quality across datasets. Code, video results, and checkpoints are available at https://bbangsik13.github.io/OR2.