🤖 AI Summary
Traditional frame-based robotic navigation suffers from motion blur and high latency due to fixed-frame-rate sampling. To address this, we propose a real-time, human-centered navigation framework that fuses asynchronous event-camera sensing with multimodal perception (event streams + distance sensors). Methodologically, we adopt a two-stage reinforcement learning strategy: first pretraining the policy network via imitation learning, then refining it end-to-end using Deep Deterministic Policy Gradient (DDPG). This work is the first to integrate the event camera’s ultra-low latency and high dynamic range with a staged policy learning paradigm, significantly improving response speed, environmental adaptability, and sample efficiency. Simulation results demonstrate robust real-time pedestrian following and dynamic obstacle avoidance, validating the framework’s effectiveness and generalizability in highly agile scenarios.
📝 Abstract
This work introduces a robot navigation controller that combines event cameras and other sensors with reinforcement learning to enable real-time human-centered navigation and obstacle avoidance. Unlike conventional image-based controllers, which operate at fixed rates and suffer from motion blur and latency, this approach leverages the asynchronous nature of event cameras to process visual information over flexible time intervals, enabling adaptive inference and control. The framework integrates event-based perception, additional range sensing, and policy optimization via Deep Deterministic Policy Gradient, with an initial imitation learning phase to improve sample efficiency. Promising results are achieved in simulated environments, demonstrating robust navigation, pedestrian following, and obstacle avoidance. A demo video is available at the project website.