🤖 AI Summary
This work addresses the generalization bottleneck in real-time target search and path planning for autonomous systems operating in unknown environments. We propose a Vector Symbolic Architecture-based Occupancy Grid Map (VSA-OGM), the first approach to integrate biologically inspired hyperdimensional computing into the perception–decision closed loop of reinforcement learning. VSA-OGM replaces conventional Bayesian Hilbert Maps (BHM) and enables end-to-end compatibility with downstream deep RL policies. Experiments on the F1Tenth simulation platform demonstrate that VSA-OGM improves task success rate by 47% in unseen environments. While achieving comparable sample efficiency to BHM under single- and multi-scenario training, VSA-OGM exhibits significantly superior cross-scenario transfer performance. These results validate its enhanced robustness to environmental variations and stronger generalization capability—critical for deploying autonomous agents beyond training conditions.
📝 Abstract
Real-time autonomous systems utilize multi-layer computational frameworks to perform critical tasks such as perception, goal finding, and path planning. Traditional methods implement perception using occupancy grid mapping (OGM), segmenting the environment into discretized cells with probabilistic information. This classical approach is well-established and provides a structured input for downstream processes like goal finding and path planning algorithms. Recent approaches leverage a biologically inspired mathematical framework known as vector symbolic architectures (VSA), commonly known as hyperdimensional computing, to perform probabilistic OGM in hyperdimensional space. This approach, VSA-OGM, provides native compatibility with spiking neural networks, positioning VSA-OGM as a potential neuromorphic alternative to conventional OGM. However, for large-scale integration, it is essential to assess the performance implications of VSA-OGM on downstream tasks compared to established OGM methods. This study examines the efficacy of VSA-OGM against a traditional OGM approach, Bayesian Hilbert Maps (BHM), within reinforcement learning based goal finding and path planning frameworks, across a controlled exploration environment and an autonomous driving scenario inspired by the F1-Tenth challenge. Our results demonstrate that VSA-OGM maintains comparable learning performance across single and multi-scenario training configurations while improving performance on unseen environments by approximately 47%. These findings highlight the increased generalizability of policy networks trained with VSA-OGM over BHM, reinforcing its potential for real-world deployment in diverse environments.