AgentRob: From Virtual Forum Agents to Hijacked Physical Robots

πŸ“… 2026-02-14
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work proposes a novel three-layer framework that enables community-driven, language-mediated coordination between multiple large language model (LLM)-based agents and physical robotsβ€”a capability largely absent in current systems that rely on direct control. By leveraging an online forum as the interaction medium, agents interpret user instructions via @-mention mechanisms, then orchestrate Unitree Go2/G1 robots through vision-language model (VLM)-powered controllers integrated with the Model Context Protocol (MCP). The architecture supports asynchronous, persistent, and identity-aware cross-domain interactions, allowing multiple embodied agents to operate concurrently within a unified forum environment. Experimental results demonstrate the feasibility and effectiveness of this paradigm, marking the first realization of a multi-agent, multi-robot collaborative system coordinated entirely through natural language in a public, community-oriented setting.

Technology Category

Application Category

πŸ“ Abstract
Large Language Model (LLM)-powered autonomous agents have demonstrated significant capabilities in virtual environments, yet their integration with the physical world remains narrowly confined to direct control interfaces. We present AgentRob, a framework that bridges online community forums, LLM-powered agents, and physical robots through the Model Context Protocol (MCP). AgentRob enables a novel paradigm where autonomous agents participate in online forums--reading posts, extracting natural language commands, dispatching physical robot actions, and reporting results back to the community. The system comprises three layers: a Forum Layer providing asynchronous, persistent, multi-agent interaction; an Agent Layer with forum agents that poll for @mention-targeted commands; and a Robot Layer with VLM-driven controllers and Unitree Go2/G1 hardware that translate commands into robot primitives via iterative tool calling. The framework supports multiple concurrent agents with distinct identities and physical embodiments coexisting in the same forum, establishing the feasibility of forum-mediated multi-agent robot orchestration.
Problem

Research questions and friction points this paper is trying to address.

LLM-powered agents
physical robots
online forums
multi-agent orchestration
autonomous agent integration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Model Context Protocol
LLM-powered agents
forum-mediated orchestration
VLM-driven robot control
multi-agent physical embodiment
πŸ”Ž Similar Papers
No similar papers found.
Wenrui Liu
Wenrui Liu
Zhejiang University
time seriesmulti-modalLLM
Yaxuan Wang
Yaxuan Wang
PhD Student of Computer Science, University of California, Santa Curz
machine learning
X
Xun Zhang
Peking University, China
Yanshu Wang
Yanshu Wang
Tsinghua University
Computer&Economy
J
Jiashen Wei
Peking University, China
Y
Yifan Xiang
Peking University, China
Y
Yuhang Wang
Peking University, China
M
Mingshen Ye
Peking University, China
E
Elsie Dai
Peking University, China
Z
Zhiqi Liu
Peking University, China
Yingjie Xu
Yingjie Xu
Hong Kong University of Science and Technology(Guang Zhou))
Computer Vision
Xinyang Chen
Xinyang Chen
Associate Professor, Harbin Institute of Technology (Shenzhen)
machine learningmultimodal learningtransfer learning
H
Hengzhe Sun
Peking University, China
J
Jiyu Shen
Peking University, China
J
Jingjing He
Peking University, China
Tong Yang
Tong Yang
Peking University, Beijing, China. PKU. εŒ—δΊ¬ε€§ε­¦
SketchNetwork measurementBloom filterIP lookupHash Table