🤖 AI Summary
Existing AI-driven design tools struggle to simultaneously satisfy anatomical constraints and semantic design intent, limiting the functional ergonomics of everyday objects. This paper introduces the first differentiable mesh deformation framework that unifies semantic alignment, contact relationship modeling, and geometric interpenetration constraints—enabling human-aware, text/image/sketch-driven 3D generation. Our method integrates CLIP- and diffusion-based semantic guidance with physics-informed contact modeling and a penetration-aware loss function, facilitating end-to-end unsupervised optimization. Evaluated across diverse household items, our approach yields qualitatively natural human–object interaction postures and quantitatively improves contact accuracy by 23.6% over baselines. It also significantly enhances semantic fidelity and anatomical fit precision, demonstrating superior capability in generating functionally appropriate, user-centered designs.
📝 Abstract
For designing a wide range of everyday objects, the design process should be aware of both the human body and the underlying semantics of the design specification. However, these two objectives present significant challenges to the current AI-based designing tools. In this work, we present a method to synthesize body-aware 3D objects from a base mesh given an input body geometry and either text or image as guidance. The generated objects can be simulated on virtual characters, or fabricated for real-world use. We propose to use a mesh deformation procedure that optimizes for both semantic alignment as well as contact and penetration losses. Using our method, users can generate both virtual or real-world objects from text, image, or sketch, without the need for manual artist intervention. We present both qualitative and quantitative results on various object categories, demonstrating the effectiveness of our approach.