Dyna-LfLH: Learning Agile Navigation in Dynamic Environments from Learned Hallucination

📅 2024-03-25
🏛️ arXiv.org
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
To address the poor generalization and data scarcity challenges of motion planners in dynamic, densely cluttered environments, this paper proposes a self-supervised motion planning learning framework. Methodologically, it introduces “learning-based hallucination” for dynamic obstacle modeling—novelly employing learnable latent variable distributions to physically plausible generate diverse obstacle trajectories, thereby automatically constructing high-difficulty training scenarios and overcoming the fundamental limitation of conventional learning-from-hindsight (LfH) approaches in handling moving obstacles. The framework integrates implicit modeling, an end-to-end planning network, and simulation-to-real co-training. Extensive evaluations on both simulated and real-world robotic platforms demonstrate that our method achieves up to a 25% improvement in task success rate over state-of-the-art learning-based and classical planners, while significantly enhancing navigation safety and agility.

Technology Category

Application Category

📝 Abstract
This paper presents a self-supervised learning method to safely learn a motion planner for ground robots to navigate environments with dense and dynamic obstacles. When facing highly-cluttered, fast-moving, hard-to-predict obstacles, classical motion planners may not be able to keep up with limited onboard computation. For learning-based planners, high-quality demonstrations are difficult to acquire for imitation learning while reinforcement learning becomes inefficient due to the high probability of collision during exploration. To safely and efficiently provide training data, the Learning from Hallucination (LfH) approaches synthesize difficult navigation environments based on past successful navigation experiences in relatively easy or completely open ones, but unfortunately cannot address dynamic obstacles. In our new Dynamic Learning from Learned Hallucination (Dyna-LfLH), we design and learn a novel latent distribution and sample dynamic obstacles from it, so the generated training data can be used to learn a motion planner to navigate in dynamic environments. Dyna-LfLH is evaluated on a ground robot in both simulated and physical environments and achieves up to 25% better success rate compared to baselines.
Problem

Research questions and friction points this paper is trying to address.

Training motion planners for dense dynamic obstacle navigation
Overcoming learning inefficiencies in imitation and reinforcement learning
Generating dynamic obstacles from learned latent distributions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised training with dynamic obstacle hallucination
Generating obstacles from learned latent distribution
Improving navigation success rate in dynamic environments
🔎 Similar Papers
No similar papers found.
S
Saad Abdul Ghani
George Mason University
Zizhao Wang
Zizhao Wang
UT Austin
reinforcement learning
P
Peter Stone
The University of Texas at Austin, Sony AI
X
Xuesu Xiao
George Mason University