π€ AI Summary
Bridging the modality gap between vision-action data in embodied tasks and purely text-based large language models (LLMs) remains challenging for zero-training robotic action prediction.
Method: We propose RoboPrompt, a framework that enables direct LLM-driven action generation without fine-tuning or visual encoders. It employs keyframe heuristics for scene selection and lossless textual encoding of end-effector actions and object poses, constructing structured in-context learning (ICL) templates that map multimodal embodied information to LLM-compatible textual prompts end-to-end.
Contribution/Results: Evaluated on both simulation and real-robot tasks, RoboPrompt achieves up to 37% higher action prediction accuracy than zero-shot and conventional ICL baselines. It demonstrates, for the first time, the feasibility of using off-the-shelf, purely textual LLMsβsuch as Llama-3 and GPT-3.5βfor embodied action reasoning without any training, parameter updates, or auxiliary vision modules.
π Abstract
Recently, Large Language Models (LLMs) have achieved remarkable success using in-context learning (ICL) in the language domain. However, leveraging the ICL capabilities within LLMs to directly predict robot actions remains largely unexplored. In this paper, we introduce RoboPrompt, a framework that enables off-the-shelf text-only LLMs to directly predict robot actions through ICL without training. Our approach first heuristically identifies keyframes that capture important moments from an episode. Next, we extract end-effector actions from these keyframes as well as the estimated initial object poses, and both are converted into textual descriptions. Finally, we construct a structured template to form ICL demonstrations from these textual descriptions and a task instruction. This enables an LLM to directly predict robot actions at test time. Through extensive experiments and analysis, RoboPrompt shows stronger performance over zero-shot and ICL baselines in simulated and real-world settings.