🤖 AI Summary
Accurately inferring the world belief state—i.e., Level-1 situation awareness—of human teammates in dynamic, three-dimensional, and partially observable real-world environments remains a key challenge for effective human–robot collaboration. Drawing on theory from mental models, this work proposes the first end-to-end system capable of real-time inference of human belief states in realistic domestic settings by integrating 3D environmental perception, semantic reasoning, and explicit belief modeling. The system is seamlessly embedded within a human–robot teaming framework and operates without requiring frequent explicit communication. Its efficacy has been validated both in high-fidelity simulation and on physical robotic platforms, demonstrating significant improvements in collaborative efficiency during proactive assistance tasks.
📝 Abstract
We investigate estimating a human's world belief state using a robot's observations in a dynamic, 3D, and partially observable environment. The methods are grounded in mental model theory, which posits that human decision making, contextual reasoning, situation awareness, and behavior planning draw from an internal simulation or world belief state. When in teams, the mental model also includes a team model of each teammate's beliefs and capabilities, enabling fluent teamwork without the need for constant and explicit communication. In this work we replicate a core component of the team model by inferring a teammate's belief state, or level one situation awareness, as a human-robot team navigates a household environment. We evaluate our methods in a realistic simulation, extend to a real-world robot platform, and demonstrate a downstream application of the belief state through an active assistance semantic reasoning task.