MochiSwarm: A testbed for robotic blimps in realistic environments

📅 2025-03-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Autonomous operation of aerial robots—particularly airships—in real-world environments remains challenging under infrastructure-free conditions (i.e., without GPS or motion-capture systems), hindering scalable deployment for logistics delivery and surveillance. Method: This paper introduces the first lightweight, modular, fully autonomous, open-source airship swarm testbed. Its software stack integrates embedded visual servoing, onboard SLAM-compatible perception, distributed communication, and multi-agent cooperative control, enabling plug-and-play sensing and reconfigurable actuation (e.g., visual-servoed differential drive) for infrastructure-free, scene-adaptive autonomy. Contribution/Results: We experimentally validate end-to-end pick-and-deliver operations with a 12-airship swarm in unstructured real-world environments. The platform significantly improves scalability, robustness, and full autonomy while enabling low-power, wide-area aerial robot swarm deployment—a novel paradigm for infrastructure-agnostic aerial robotics.

Technology Category

Application Category

📝 Abstract
Testing aerial robots in tasks such as pickup-and-delivery and surveillance significantly benefits from high energy efficiency and scalability of the deployed robotic system. This paper presents MochiSwarm, an open-source testbed of light-weight robotic blimps, ready for multi-robot operation without external localization. We introduce the system design in hardware, software, and perception, which capitalizes on modularity, low cost, and light weight. The hardware allows for rapid modification, which enables the integration of additional sensors to enhance autonomy for different scenarios. The software framework supports different actuation models and communication between the base station and multiple blimps. The detachable perception module allows independent blimps to perform tasks that involve detection and autonomous actuation. We showcase a differential-drive module as an example, of which the autonomy is enabled by visual servoing using the perception module. A case study of pickup-and-delivery tasks with up to 12 blimps highlights the autonomy of the MochiSwarm without external infrastructures.
Problem

Research questions and friction points this paper is trying to address.

Develops MochiSwarm for testing robotic blimps in realistic environments.
Enables multi-robot operations without external localization support.
Supports modular, low-cost, lightweight designs for enhanced autonomy.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Open-source testbed for lightweight robotic blimps
Modular hardware for rapid sensor integration
Detachable perception module enabling autonomous tasks
🔎 Similar Papers
No similar papers found.
J
Jiawei Xu
Autonomous and Intelligent Robotics Laboratory (AIRLab), Lehigh University, PA, 18015, USA
T
Thong Vu
Autonomous and Intelligent Robotics Laboratory (AIRLab), Lehigh University, PA, 18015, USA
D
Diego S. D’Antonio
Autonomous and Intelligent Robotics Laboratory (AIRLab), Lehigh University, PA, 18015, USA
David Saldaña
David Saldaña
Assistant Professor at Lehigh University
RoboticsModular Aerial RobotsRobot SwarmsMulti-Robot Systems