Catalyst4D: High-Fidelity 3D-to-4D Scene Editing via Dynamic Propagation

📅 2026-03-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses key challenges in dynamic scene editing—such as motion artifacts, temporal flickering, and inconsistent style propagation—by proposing a spatiotemporally coherent, high-fidelity editing framework that transfers high-quality 3D edits to dynamic 4D Gaussian scenes. The approach introduces two core innovations: an anchor-based motion guidance mechanism that effectively mitigates cross-region interference and motion drift, and a color-uncertainty-guided appearance optimization strategy that enhances visual consistency across time. By integrating 3D Gaussian representations, optimal transport, anchor correspondence construction, and selective appearance refinement, the method achieves significant improvements over existing techniques in both visual fidelity and motion coherence, enabling stable and high-quality editing of dynamic scenes.

Technology Category

Application Category

📝 Abstract
Recent advances in 3D scene editing using NeRF and 3DGS enable high-quality static scene editing. In contrast, dynamic scene editing remains challenging, as methods that directly extend 2D diffusion models to 4D often produce motion artifacts, temporal flickering, and inconsistent style propagation. We introduce Catalyst4D, a framework that transfers high-quality 3D edits to dynamic 4D Gaussian scenes while maintaining spatial and temporal coherence. At its core, Anchor-based Motion Guidance (AMG) builds a set of structurally stable and spatially representative anchors from both original and edited Gaussians. These anchors serve as robust region-level references, and their correspondences are established via optimal transport to enable consistent deformation propagation without cross-region interference or motion drift. Complementarily, Color Uncertainty-guided Appearance Refinement (CUAR) preserves temporal appearance consistency by estimating per-Gaussian color uncertainty and selectively refining regions prone to occlusion-induced artifacts. Extensive experiments demonstrate that Catalyst4D achieves temporally stable, high-fidelity dynamic scene editing and outperforms existing methods in both visual quality and motion coherence.
Problem

Research questions and friction points this paper is trying to address.

4D scene editing
temporal coherence
motion artifacts
dynamic scene
spatial consistency
Innovation

Methods, ideas, or system contributions that make the work stand out.

4D scene editing
Anchor-based Motion Guidance
Color Uncertainty-guided Appearance Refinement
dynamic Gaussian splatting
temporal coherence
🔎 Similar Papers
No similar papers found.
S
Shifeng Chen
State Key Laboratory of Complex and Critical Software Environment, Beijing, China; School of Computer Science and Engineering, Beihang University, China
Yihui Li
Yihui Li
Beihang University
J
Jun Liao
School of Artificial Intelligence, Beihang University, China
H
Hongyu Yang
School of Artificial Intelligence, Beihang University, China
Di Huang
Di Huang
Computer Science and Engineering, Beihang University
Computer VisionRepresentation LearningGenerative AIEmbodied AI