Learning to Crawl: Latent Model-Based Reinforcement Learning for Soft Robotic Adaptive Locomotion

📅 2025-10-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Soft-bodied crawling robots struggle to achieve adaptive locomotion under model inaccuracies, sensor noise, and unknown gaits. To address this, we propose a model-based reinforcement learning (MB-RL) framework grounded in latent-state dynamical modeling. Our method employs variational inference to learn a low-dimensional latent dynamics model solely from noisy on-board IMU and time-of-flight (ToF) distance measurements, enabling short-horizon motion prediction and end-to-end gait policy optimization. The key innovation lies in the tight integration of implicit dynamic modeling with an actor-critic RL architecture, eliminating reliance on precise physical models or external perception systems. Simulation results demonstrate that the proposed approach robustly generates adaptive crawling gaits under strong sensor noise, significantly enhancing the autonomy and locomotive adaptability of soft robots in uncertain environments.

Technology Category

Application Category

📝 Abstract
Soft robotic crawlers are mobile robots that utilize soft body deformability and compliance to achieve locomotion through surface contact. Designing control strategies for such systems is challenging due to model inaccuracies, sensor noise, and the need to discover locomotor gaits. In this work, we present a model-based reinforcement learning (MB-RL) framework in which latent dynamics inferred from onboard sensors serve as a predictive model that guides an actor-critic algorithm to optimize locomotor policies. We evaluate the framework on a minimal crawler model in simulation using inertial measurement units and time-of-flight sensors as observations. The learned latent dynamics enable short-horizon motion prediction while the actor-critic discovers effective locomotor policies. This approach highlights the potential of latent-dynamics MB-RL for enabling embodied soft robotic adaptive locomotion based solely on noisy sensor feedback.
Problem

Research questions and friction points this paper is trying to address.

Developing control strategies for soft robotic crawlers
Addressing model inaccuracies and sensor noise in locomotion
Discovering adaptive locomotor gaits using reinforcement learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Latent dynamics model guides actor-critic learning
Onboard sensors provide observations for locomotion control
Model-based reinforcement learning optimizes soft robot policies
🔎 Similar Papers
No similar papers found.
V
Vaughn Gzenda
Embodied Learning and Intelligence for eXploration and Innovative soft Robotics (ELIXIR) Lab, Toronto Metropolitan University, Toronto, ON, Canada
Robin Chhabra
Robin Chhabra
Professor of Robotics & Mechatronics, Toronto Metropolitan University
Soft RoboticsEmbodied AIMulti-Robot SystemsRobotic Self PerceptionGeometric Mechanics