🤖 AI Summary
To address the challenge of sparse radar signals and missing joint detections in millimeter-wave (mmWave) radar-based human pose estimation (HPE) caused by mirror-like reflections, this paper proposes an end-to-end spatiotemporal joint modeling framework. Methodologically, it pioneers the integration of the Mamba state-space model into radar HPE; introduces a Cross-View Fusion Mamba encoder for effective dual-radar cross-view feature fusion; and designs a Spatio-Temporal-Cross Attention decoder to jointly model multi-frame temporal dynamics. Additionally, a velocity-constrained loss is incorporated to enhance motion smoothness. Evaluated on the TransHuPR and HuPR benchmarks, the method achieves absolute AP improvements of 11.0 and 14.6 over strong baselines, respectively—demonstrating substantial gains in accuracy and robustness while maintaining computationally efficient inference.
📝 Abstract
Millimeter-wave radar offers a privacy-preserving and lighting-invariant alternative to RGB sensors for Human Pose Estimation (HPE) task. However, the radar signals are often sparse due to specular reflection, making the extraction of robust features from radar signals highly challenging. To address this, we present milliMamba, a radar-based 2D human pose estimation framework that jointly models spatio-temporal dependencies across both the feature extraction and decoding stages. Specifically, given the high dimensionality of radar inputs, we adopt a Cross-View Fusion Mamba encoder to efficiently extract spatio-temporal features from longer sequences with linear complexity. A Spatio-Temporal-Cross Attention decoder then predicts joint coordinates across multiple frames. Together, this spatio-temporal modeling pipeline enables the model to leverage contextual cues from neighboring frames and joints to infer missing joints caused by specular reflections. To reinforce motion smoothness, we incorporate a velocity loss alongside the standard keypoint loss during training. Experiments on the TransHuPR and HuPR datasets demonstrate that our method achieves significant performance improvements, exceeding the baselines by 11.0 AP and 14.6 AP, respectively, while maintaining reasonable complexity. Code: https://github.com/NYCU-MAPL/milliMamba