๐ค AI Summary
This paper addresses the problem of reconstructing the 3D deformation state of deformable cloth from monocular RGB images. We propose a prediction-update framework: an action-conditioned dynamics model first generates an initial state estimate, which is then refined via gradient-driven optimization in pixel space using differentiable 3D Gaussian rasterization. Our key contribution is the first integration of mesh-based cloth representation with differentiable Gaussian rendering, establishing an end-to-end differentiable mapping from cloth state to imageโtrainable solely with RGB supervision. Evaluated on multiple cloth simulation benchmarks, our method achieves significant improvements in state estimation accuracy (21.3% average error reduction) and faster convergence (37% fewer optimization iterations). This work introduces a new paradigm for depth-free cloth perception and physics-informed simulation.
๐ Abstract
We introduce Cloth-Splatting, a method for estimating 3D states of cloth from RGB images through a prediction-update framework. Cloth-Splatting leverages an action-conditioned dynamics model for predicting future states and uses 3D Gaussian Splatting to update the predicted states. Our key insight is that coupling a 3D mesh-based representation with Gaussian Splatting allows us to define a differentiable map between the cloth state space and the image space. This enables the use of gradient-based optimization techniques to refine inaccurate state estimates using only RGB supervision. Our experiments demonstrate that Cloth-Splatting not only improves state estimation accuracy over current baselines but also reduces convergence time.