🤖 AI Summary
Service robots operating in close proximity to humans require low-cost, omnidirectional human perception. Method: We propose the first self-supervised method for human detection and 2D pose estimation using only a single-line (1D) LiDAR. Cross-modal knowledge distillation from RGB-D detection transfers narrow-field supervision to omnidirectional 1D point sequences without manual annotation; a lightweight CNN-LSTM architecture models temporal point sequences and enables environment-adaptive training. Contribution/Results: Our approach achieves 71% precision and 80% recall in novel scenes, with mean errors of 13 cm in distance and 44° in orientation—trained autonomously on just 70 minutes of collected data. This work is the first to demonstrate that a single-line LiDAR, under self-supervision, can support robust spatial human perception—substantially reducing hardware cost and enhancing deployment generalizability.
📝 Abstract
Localizing humans is a key prerequisite for any service robot operating in proximity to people. In these scenarios, robots rely on a multitude of state-of-the-art detectors usually designed to operate with RGB-D cameras or expensive 3D LiDARs. However, most commercially available service robots are equipped with cameras with a narrow field of view, making them blind when a user is approaching from other directions, or inexpensive 1D LiDARs whose readings are difficult to interpret. To address these limitations, we propose a self-supervised approach to detect humans and estimate their 2D pose from 1D LiDAR data, using detections from an RGB-D camera as a supervision source. Our approach aims to provide service robots with spatial awareness of nearby humans. After training on 70 minutes of data autonomously collected in two environments, our model is capable of detecting humans omnidirectionally from 1D LiDAR data in a novel environment, with 71% precision and 80% recall, while retaining an average absolute error of 13 cm in distance and 44{deg} in orientation.