The RoboSense Challenge: Sense Anything, Navigate Anywhere, Adapt Across Platforms

📅 2026-01-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenges of robotic perception in open, dynamic environments—such as sensor noise, environmental variability, and platform heterogeneity—by introducing the first unified benchmark framework that integrates multidimensional tasks. The framework encompasses five core tracks: language-guided decision-making, socially compliant navigation, sensor configuration generalization, cross-view/cross-modal alignment, and cross-platform 3D perception. Through standardized datasets and evaluation protocols, it synergistically combines techniques in language-vision alignment, multimodal fusion, domain adaptation, and social behavior modeling to advance reproducible and robust perception research. The initiative attracted participation from 143 teams across 85 institutions in 16 countries, with 23 winning solutions yielding common methodologies and design principles that significantly accelerate progress in cross-platform robust perception.

Technology Category

Application Category

📝 Abstract
Autonomous systems are increasingly deployed in open and dynamic environments -- from city streets to aerial and indoor spaces -- where perception models must remain reliable under sensor noise, environmental variation, and platform shifts. However, even state-of-the-art methods often degrade under unseen conditions, highlighting the need for robust and generalizable robot sensing. The RoboSense 2025 Challenge is designed to advance robustness and adaptability in robot perception across diverse sensing scenarios. It unifies five complementary research tracks spanning language-grounded decision making, socially compliant navigation, sensor configuration generalization, cross-view and cross-modal correspondence, and cross-platform 3D perception. Together, these tasks form a comprehensive benchmark for evaluating real-world sensing reliability under domain shifts, sensor failures, and platform discrepancies. RoboSense 2025 provides standardized datasets, baseline models, and unified evaluation protocols, enabling large-scale and reproducible comparison of robust perception methods. The challenge attracted 143 teams from 85 institutions across 16 countries, reflecting broad community engagement. By consolidating insights from 23 winning solutions, this report highlights emerging methodological trends, shared design principles, and open challenges across all tracks, marking a step toward building robots that can sense reliably, act robustly, and adapt across platforms in real-world environments.
Problem

Research questions and friction points this paper is trying to address.

robust perception
domain shift
sensor noise
platform generalization
environmental variation
Innovation

Methods, ideas, or system contributions that make the work stand out.

robust perception
cross-platform adaptation
domain generalization
multimodal correspondence
autonomous navigation
🔎 Similar Papers
Lingdong Kong
Lingdong Kong
National University of Singapore
Computer VisionDeep Learning
Shaoyuan Xie
Shaoyuan Xie
University of California, Irvine
AI SecurityMachine LearningAutonomous Driving
Zeying Gong
Zeying Gong
The Hong Kong University of Science and Technology (Guangzhou)
ForecastingEmbodied AI
Y
Ye Li
Challenge & Workshop Organizers
M
Meng Chu
Challenge & Workshop Organizers
A
Ao Liang
Challenge & Workshop Organizers
Yuhao Dong
Yuhao Dong
Tsinghua University, Nanyang Technological University
Multi-modal LearningComputer Vision
Tianshuai Hu
Tianshuai Hu
Ph.D. student in HKUST
RoboticsAutonomous Driving
Ronghe Qiu
Ronghe Qiu
Hong Kong University of Science and Technology
Embodied AIMobile Manipulation
Rong Li
Rong Li
PhD student, HKUST (GZ)
Computer VisionEmbodied AI
Hanjiang Hu
Hanjiang Hu
Carnegie Mellon University
Machine LearningControlRobotics
Dongyue Lu
Dongyue Lu
National University of Singapore
Computer Vision
Wei Yin
Wei Yin
Staff Research Scientist, Horizon Robotics
World ModelGenerative AIPhysical AI
Wenhao Ding
Wenhao Ding
Research Scientist, NVIDIA Research
ReasoningSafetyReinforcement LearningSimulation
L
Linfeng Li
Challenge & Workshop Organizers
Hang Song
Hang Song
Associate Professor, Hiroshima University
Wireless sensingCommunication networkMicrowave imagingAcoustic/optical technology
Wenwei Zhang
Wenwei Zhang
Shanghai AI Laboratory
Large Language ModelScalable OversightArtificial Intelligence
Yue Ma
Yue Ma
Bytedance
NLPDialogue SystemLLM
Junwei Liang
Junwei Liang
Assistant Professor, HKUST (Guangzhou) | CSE, HKUST | Ph.D. @CMU
Computer VisionRoboticsEmbodied AITrajectory Prediction
Zhedong Zheng
Zhedong Zheng
University of Macau | NUS | UTS | Fudan
AIGCData-centric AISpatial IntelligenceObject Re-identificationDomain Adaptation
L
Lai Xing Ng
Challenge & Workshop Organizers
B
Benoit R. Cottereau
Challenge & Workshop Organizers
Wei Tsang Ooi
Wei Tsang Ooi
National University of Singapore
Multimedia SystemsInteractive SystemsIntelligent Systems
Ziwei Liu
Ziwei Liu
Associate Professor, Nanyang Technological University
Computer VisionMachine LearningComputer Graphics