Incremental Residual Reinforcement Learning Toward Real-World Learning for Social Navigation

πŸ“… 2026-04-09
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenges of insufficient simulation coverage due to regional variations in real-world social navigation and the limited computational resources and low learning efficiency on edge devices. To tackle these issues, the authors propose a lightweight online learning framework that integrates incremental learning with residual reinforcement learning. The method eliminates the need for a replay buffer and achieves efficient adaptation through residual policy optimization built upon a base policy. By innovatively combining incremental learning with residual reinforcement learning, the approach matches the performance of conventional replay-buffer-based methods in simulation and outperforms existing incremental learning strategies. Real-world experiments further demonstrate the robot’s rapid adaptation to unseen environments, significantly enhancing generalization and learning efficiency under edge deployment constraints.
πŸ“ Abstract
As the demand for mobile robots continues to increase, social navigation has emerged as a critical task, driving active research into deep reinforcement learning (RL) approaches. However, because pedestrian dynamics and social conventions vary widely across different regions, simulations cannot easily encompass all possible real-world scenarios. Real-world RL, in which agents learn while operating directly in physical environments, presents a promising solution to this issue. Nevertheless, this approach faces significant challenges, particularly regarding constrained computational resources on edge devices and learning efficiency. In this study, we propose incremental residual RL (IRRL). This method integrates incremental learning, which is a lightweight process that operates without a replay buffer or batch updates, with residual RL, which enhances learning efficiency by training only on the residuals relative to a base policy. Through the simulation experiments, we demonstrated that, despite lacking a replay buffer, IRRL achieved performance comparable to those of conventional replay buffer-based methods and outperformed existing incremental learning approaches. Furthermore, the real-world experiments confirmed that IRRL can enable robots to effectively adapt to previously unseen environments through the real-world learning.
Problem

Research questions and friction points this paper is trying to address.

social navigation
real-world reinforcement learning
computational constraints
learning efficiency
mobile robots
Innovation

Methods, ideas, or system contributions that make the work stand out.

incremental learning
residual reinforcement learning
real-world reinforcement learning
social navigation
edge-device RL
πŸ”Ž Similar Papers
No similar papers found.
H
Haruto Nagahisa
School of Engineering, Kyushu University, Fukuoka, Japan
Kohei Matsumoto
Kohei Matsumoto
Kyushu University
Intelligent RoboticsService RobotSpatial IntelligenceMachine Learning
Y
Yuki Tomita
School of Information Science and Electrical Engineering, Kyushu University, Fukuoka, Japan
Y
Yuki Hyodo
School of Information Science and Electrical Engineering, Kyushu University, Fukuoka, Japan
Ryo Kurazume
Ryo Kurazume
Professor of Computer Science, Kyushu University
RoboticsComputer vision