🤖 AI Summary
Blind individuals face significant challenges in “last-mile” navigation, including difficulty locating building entrances, weak obstacle detection, and insufficient comprehension of complex spatial layouts. This paper presents the first systematic comparative study of head-mounted versus cane-mounted cameras for blind navigation, evaluating their respective field-of-view efficacy. Through synchronized dual-perspective field data collection, user studies, and multimodal behavioral analysis—integrated with SLAM-based mapping and NeRF-based 3D reconstruction—we empirically demonstrate complementary perceptual capabilities: the head-mounted view improves global localization accuracy, while the cane-mounted view enhances ground-level detail coverage and scene reconstruction fidelity. We propose a head-cane collaborative sensor configuration that significantly outperforms single-view baselines across all metrics—including localization precision, 3D reconstruction quality, and task completion rate. This work provides the first empirical, viewpoint-driven design principle for sensor placement in wearable assistive navigation systems.
📝 Abstract
Blind individuals face persistent challenges in last-mile navigation, including locating entrances, identifying obstacles, and navigating complex or cluttered spaces. Although wearable cameras are increasingly used in assistive systems, there has been no systematic, vantage-focused comparison to guide their design. This paper addresses that gap through a two-part investigation. First, we surveyed ten experienced blind cane users, uncovering navigation strategies, pain points, and technology preferences. Participants stressed the importance of multi-sensory integration, destination-focused travel, and assistive tools that complement (rather than replace) the cane's tactile utility. Second, we conducted controlled data collection with a blind participant navigating five real-world environments using synchronized head- and cane-mounted cameras, isolating vantage placement as the primary variable. To assess how each vantage supports spatial perception, we evaluated SLAM performance (for localization and mapping) and NeRF-based 3D reconstruction (for downstream scene understanding). Head-mounted sensors delivered superior localization accuracy, while cane-mounted views offered broader ground-level coverage and richer environmental reconstructions. A combined (head+cane) configuration consistently outperformed both. These results highlight the complementary strengths of different sensor placements and offer actionable guidance for developing hybrid navigation aids that are perceptive, robust, and user-aligned.