🤖 AI Summary
Soft-bodied crawling robots struggle to achieve adaptive locomotion under model inaccuracies, sensor noise, and unknown gaits. To address this, we propose a model-based reinforcement learning (MB-RL) framework grounded in latent-state dynamical modeling. Our method employs variational inference to learn a low-dimensional latent dynamics model solely from noisy on-board IMU and time-of-flight (ToF) distance measurements, enabling short-horizon motion prediction and end-to-end gait policy optimization. The key innovation lies in the tight integration of implicit dynamic modeling with an actor-critic RL architecture, eliminating reliance on precise physical models or external perception systems. Simulation results demonstrate that the proposed approach robustly generates adaptive crawling gaits under strong sensor noise, significantly enhancing the autonomy and locomotive adaptability of soft robots in uncertain environments.
📝 Abstract
Soft robotic crawlers are mobile robots that utilize soft body deformability and compliance to achieve locomotion through surface contact. Designing control strategies for such systems is challenging due to model inaccuracies, sensor noise, and the need to discover locomotor gaits. In this work, we present a model-based reinforcement learning (MB-RL) framework in which latent dynamics inferred from onboard sensors serve as a predictive model that guides an actor-critic algorithm to optimize locomotor policies. We evaluate the framework on a minimal crawler model in simulation using inertial measurement units and time-of-flight sensors as observations. The learned latent dynamics enable short-horizon motion prediction while the actor-critic discovers effective locomotor policies. This approach highlights the potential of latent-dynamics MB-RL for enabling embodied soft robotic adaptive locomotion based solely on noisy sensor feedback.