🤖 AI Summary
To address geometric drift and temporal inconsistency in 4D Gaussian Splatting (4DGS) reconstruction for dynamic endoscopic scenes, this paper proposes a jointly trained geometrically guided and temporally aware framework. Methodologically, it introduces confidence-gated monocular depth distillation as a strong geometric prior; designs XYZT spatiotemporal embeddings and screw-motion parameterization to jointly model rigid and non-rigid deformations; and integrates keyframe constraints with streaming optimization, alongside scale-invariant depth loss, depth gradient regularization, and soft warm-up scheduling. Evaluated on EndoNeRF and StereoMIS-P1, the method achieves state-of-the-art monocular 4D reconstruction performance—significantly improving geometric accuracy and long-sequence temporal coherence while maintaining efficient inference.
📝 Abstract
Endoscopic (endo) video exhibits strong view-dependent effects such as specularities, wet reflections, and occlusions. Pure photometric supervision misaligns with geometry and triggers early geometric drift, where erroneous shapes are reinforced during densification and become hard to correct. We ask how to anchor geometry early for 4D Gaussian splatting (4DGS) while maintaining temporal consistency and efficiency in dynamic endoscopic scenes. Thus, we present Endo-G$^{2}$T, a geometry-guided and temporally aware training scheme for time-embedded 4DGS. First, geo-guided prior distillation converts confidence-gated monocular depth into supervision with scale-invariant depth and depth-gradient losses, using a warm-up-to-cap schedule to inject priors softly and avoid early overfitting. Second, a time-embedded Gaussian field represents dynamics in XYZT with a rotor-like rotation parameterization, yielding temporally coherent geometry with lightweight regularization that favors smooth motion and crisp opacity boundaries. Third, keyframe-constrained streaming improves efficiency and long-horizon stability through keyframe-focused optimization under a max-points budget, while non-keyframes advance with lightweight updates. Across EndoNeRF and StereoMIS-P1 datasets, Endo-G$^{2}$T achieves state-of-the-art results among monocular reconstruction baselines.