🤖 AI Summary
Active object reconstruction suffers from low efficiency due to repeated online viewpoint re-planning. Method: This paper proposes a novel single-shot approach to generate an optimal multi-view sequence. Its core innovation lies in the first use of a 3D diffusion model as a strong geometric-texture joint prior, implicitly encoding object structure and appearance distributions; this is integrated with multi-view geometric constraints and an uncertainty-driven view scoring mechanism to enable efficient, targeted viewpoint planning—particularly for challenging-to-reconstruct regions. The method requires only one forward pass, eliminating iterative optimization. Contribution/Results: Evaluated on both simulation and real robotic platforms, it achieves a 18.7% improvement in reconstruction completeness and reduces viewpoint planning time by 92% compared to baseline methods.
📝 Abstract
Active object reconstruction is crucial for many robotic applications. A key aspect in these scenarios is generating object-specific view configurations to obtain informative measurements for reconstruction. One-shot view planning enables efficient data collection by predicting all views at once, eliminating the need for time-consuming online replanning. Our primary insight is to leverage the generative power of 3D diffusion models as valuable prior information. By conditioning on initial multi-view images, we exploit the priors from the 3D diffusion model to generate an approximate object model, serving as the foundation for our view planning. Our novel approach integrates the geometric and textural distributions of the object model into the view planning process, generating views that focus on the complex parts of the object to be reconstructed. We validate the proposed active object reconstruction system through both simulation and real-world experiments, demonstrating the effectiveness of using 3D diffusion priors for one-shot view planning.