🤖 AI Summary
Current autonomous drones struggle to interpret and execute natural-language instructions in real time within unstructured environments, primarily due to reliance on handcrafted skills, labor-intensive hyperparameter tuning, or excessive computational overhead—hindering onboard deployment. This paper introduces the first fully onboard, real-time lightweight Vision-Language-Action (VLA) framework. It innovatively integrates 3D Gaussian Splatting (3DGS)-based simulation, differentiable reinforcement learning (DiffRL), and a Mixture-of-Experts (MoE) action head to jointly enhance generalization and continual learning capability. In simulation, task success rates reach 83% (seen) and 75% (unseen); on physical drones, they achieve 67% and 50%, respectively. Cross-environment average success rates are 81% in simulation and 67% on hardware—demonstrating substantial improvements in practicality and robustness of language-guided navigation.
📝 Abstract
Autonomous drones capable of interpreting and executing high-level language instructions in unstructured environments remain a long-standing goal. Yet existing approaches are constrained by their dependence on hand-crafted skills, extensive parameter tuning, or computationally intensive models unsuitable for onboard use. We introduce GRaD-Nav++, a lightweight Vision-Language-Action (VLA) framework that runs fully onboard and follows natural-language commands in real time. Our policy is trained in a photorealistic 3D Gaussian Splatting (3DGS) simulator via Differentiable Reinforcement Learning (DiffRL), enabling efficient learning of low-level control from visual and linguistic inputs. At its core is a Mixture-of-Experts (MoE) action head, which adaptively routes computation to improve generalization while mitigating forgetting. In multi-task generalization experiments, GRaD-Nav++ achieves a success rate of 83% on trained tasks and 75% on unseen tasks in simulation. When deployed on real hardware, it attains 67% success on trained tasks and 50% on unseen ones. In multi-environment adaptation experiments, GRaD-Nav++ achieves an average success rate of 81% across diverse simulated environments and 67% across varied real-world settings. These results establish a new benchmark for fully onboard Vision-Language-Action (VLA) flight and demonstrate that compact, efficient models can enable reliable, language-guided navigation without relying on external infrastructure.