TDFANet: Encoding Sequential 4D Radar Point Clouds Using Trajectory-Guided Deformable Feature Aggregation for Place Recognition

📅 2025-04-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenges of place recognition under adverse weather conditions—caused by sparsity, high noise, and low resolution in 4D radar data—this paper proposes the first end-to-end radar-based pose estimation framework. Methodologically, we introduce a novel trajectory-guided deformable feature aggregation mechanism: radar-derived ego-velocity estimates are used to predict motion trajectories and correct temporal misalignment in bird’s-eye-view (BEV) features; this is combined with dynamic point filtering, BEV grid-based encoding, optical-flow-inspired feature alignment, and multi-scale spatiotemporal attention aggregation. Evaluated on a real-world automotive 4D radar dataset, our method achieves a 21.3% improvement in place recognition accuracy under dynamic scenes, significantly outperforming existing approaches. Extensive experiments demonstrate strong robustness and practicality in challenging conditions including rain, fog, and nighttime operation.

Technology Category

Application Category

📝 Abstract
Place recognition is essential for achieving closed-loop or global positioning in autonomous vehicles and mobile robots. Despite recent advancements in place recognition using 2D cameras or 3D LiDAR, it remains to be seen how to use 4D radar for place recognition - an increasingly popular sensor for its robustness against adverse weather and lighting conditions. Compared to LiDAR point clouds, radar data are drastically sparser, noisier and in much lower resolution, which hampers their ability to effectively represent scenes, posing significant challenges for 4D radar-based place recognition. This work addresses these challenges by leveraging multi-modal information from sequential 4D radar scans and effectively extracting and aggregating spatio-temporal features.Our approach follows a principled pipeline that comprises (1) dynamic points removal and ego-velocity estimation from velocity property, (2) bird's eye view (BEV) feature encoding on the refined point cloud, (3) feature alignment using BEV feature map motion trajectory calculated by ego-velocity, (4) multi-scale spatio-temporal features of the aligned BEV feature maps are extracted and aggregated.Real-world experimental results validate the feasibility of the proposed method and demonstrate its robustness in handling dynamic environments. Source codes are available.
Problem

Research questions and friction points this paper is trying to address.

How to use 4D radar for robust place recognition
Overcoming sparsity and noise in 4D radar data
Extracting spatio-temporal features from sequential radar scans
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic points removal and ego-velocity estimation
Bird's eye view feature encoding on refined points
Multi-scale spatio-temporal feature extraction and aggregation
🔎 Similar Papers
No similar papers found.
Shouyi Lu
Shouyi Lu
TongJi University
AIGCImage EditingPoint Cloud Super-ResolutionPose EstimationPlace Recognition
G
Guirong Zhuo
School of Automotive Studies, Tongji University, Shanghai, China
H
Haitao Wang
Shanghai Geometrical Perception and Learning Co., Ltd., Shanghai, China
Q
Quan Zhou
School of Automotive Studies, Tongji University, Shanghai, China
Huanyu Zhou
Huanyu Zhou
Vehicle Engineering, Tongji University
4D Radar
Renbo Huang
Renbo Huang
Tongji University
Cooperated Perception、Motion Prediction
M
Minqing Huang
School of Automotive Studies, Tongji University, Shanghai, China
Lianqing Zheng
Lianqing Zheng
Tongji University Ph.D student
BEV/OCCVLA4D Radar PerceptionMultimodal FusionData Closed-Loop
Q
Qiang Shu
Shanghai Tongyu Automotive Technology Co., Ltd., Shanghai, China