🤖 AI Summary
This work addresses the challenge of coordinating active perception—such as gaze selection—with language generation in embodied robots under stringent low-latency constraints. To this end, the authors propose a lightweight system framework that tightly integrates a real-time multimodal large language model with attention control mechanisms and interfaces to active perception tools, enabling dynamic perception and context-aware dialogue in a synergistic manner. The approach is validated across six domestic scenarios, demonstrating high accuracy in tool-based decision-making and strong user satisfaction in subjective evaluations. Results show a significant improvement in interaction naturalness and fluency, highlighting the framework’s potential for enabling real-time, high-quality human-robot conversation in complex environments.
📝 Abstract
Situated embodied conversation requires robots to interleave real-time dialogue with active perception: deciding what to look at, when to look, and what to say under tight latency constraints. We present a simple, minimal system recipe that pairs a real-time multimodal language model with a small set of tool interfaces for attention and active perception. We study six home-style scenarios that require frequent attention shifts and increasing perceptual scope. Across four system variants, we evaluate turn-level tool-decision correctness against human annotations and collect subjective ratings of interaction quality. Results indicate that real-time multimodal large language models and tool use for active perception is a promising direction for practical situated embodied conversation.