🤖 AI Summary
In structured outdoor environments such as vineyards, LiDAR-based localization often fails due to geometric repetition between rows and perceptual ambiguity. To address this, we propose a semantic-aware particle filter localization method. Our approach fuses LiDAR measurements with object-level semantic landmarks—specifically vine trunks and support posts—and introduces a novel “semantic wall” mechanism that constructs pseudo-structural constraints from adjacent semantic landmarks to suppress inter-row ambiguity. In semantically sparse regions, it adaptively incorporates low-accuracy GPS priors to ensure global consistency. Experimental evaluation in real-world vineyards demonstrates stable intra-row localization, robust recovery when AMCL fails, and significantly improved accuracy and robustness compared to vision-based SLAM systems such as RTAB-Map.
📝 Abstract
Accurate localisation is critical for mobile robots in structured outdoor environments, yet LiDAR-based methods often fail in vineyards due to repetitive row geometry and perceptual aliasing. We propose a semantic particle filter that incorporates stable object-level detections, specifically vine trunks and support poles into the likelihood estimation process. Detected landmarks are projected into a birds eye view and fused with LiDAR scans to generate semantic observations. A key innovation is the use of semantic walls, which connect adjacent landmarks into pseudo-structural constraints that mitigate row aliasing. To maintain global consistency in headland regions where semantics are sparse, we introduce a noisy GPS prior that adaptively supports the filter. Experiments in a real vineyard demonstrate that our approach maintains localisation within the correct row, recovers from deviations where AMCL fails, and outperforms vision-based SLAM methods such as RTAB-Map.