๐ค AI Summary
This work addresses the multi-object rearrangement task in partially observable, multi-room environments. We propose HOO-POMDP, a hierarchical planning framework that employs object-oriented state abstraction and online subgoal generation at the high level, while coupling pretrained policy networks for modular execution at the low level. To our knowledge, this is the first approach to integrate object-oriented modeling into hierarchical POMDPs, unifying abstract state representation, interpretable planning, and robust execution under partial observability. Evaluated in the AI2-THOR simulator, HOO-POMDP significantly outperforms existing reinforcement learning and handcrafted planning methods across diverse multi-object, multi-room rearrangement tasksโachieving higher task completion rates, stronger generalization to unseen object configurations and room layouts, and improved adaptability to environmental dynamics.
๐ Abstract
We present an online planning framework for solving multi-object rearrangement problems in partially observable, multi-room environments. Current object rearrangement solutions, primarily based on Reinforcement Learning or hand-coded planning methods, often lack adaptability to diverse challenges. To address this limitation, we introduce a novel Hierarchical Object-Oriented Partially Observed Markov Decision Process (HOO-POMDP) planning approach. This approach comprises of (a) an object-oriented POMDP planner generating sub-goals, (b) a set of low-level policies for sub-goal achievement, and (c) an abstraction system converting the continuous low-level world into a representation suitable for abstract planning. We evaluate our system on varying numbers of objects, rooms, and problem types in AI2-THOR simulated environments with promising results.