🤖 AI Summary
This work addresses the challenge of understanding physical properties of unknown deformable objects. Conventional approaches, constrained by closed-world assumptions, exhibit poor generalization to unseen objects. To overcome this limitation, we propose a large language model (LLM)-driven embodied probing framework that tightly integrates LLM-based reasoning with robotic interaction—establishing a closed loop of *action generation → physical feedback → symbolic reasoning*. Specifically, the LLM generates exploratory actions (e.g., folding, bending) grounded in prior knowledge; a robot executes them and observes resultant deformations; the system then symbolically models intrinsic properties—such as foldability and bendability—and leverages these representations for attribute-driven task planning (e.g., bin-packing). Experiments demonstrate high-accuracy physical property identification across diverse unseen deformable objects, significantly improving downstream task success rates and cross-object generalization capability. To our knowledge, this is the first work to deeply integrate LLMs with embodied robotic probing for open-world physical property understanding.
📝 Abstract
In this paper, we consider the problem of understanding the physical properties of unseen objects through interactions between the objects and a robot. Handling unseen objects with special properties such as deformability is challenging for traditional task and motion planning approaches as they are often with the closed world assumption. Recent results in Large Language Models (LLMs) based task planning have shown the ability to reason about unseen objects. However, most studies assume rigid objects, overlooking their physical properties. We propose an LLM-based method for probing the physical properties of unseen deformable objects for the purpose of task planning. For a given set of object properties (e.g., foldability, bendability), our method uses robot actions to determine the properties by interacting with the objects. Based on the properties examined by the LLM and robot actions, the LLM generates a task plan for a specific domain such as object packing. In the experiment, we show that the proposed method can identify properties of deformable objects, which are further used for a bin-packing task where the properties take crucial roles to succeed.