Scaling Spatial Intelligence with Multimodal Foundation Models

📅 2025-11-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current multimodal foundation models exhibit significant limitations in spatial intelligence—encompassing geometric reasoning, 3D understanding, and spatial relational modeling. To address this, we introduce the SenseNova-SI series, built upon architectures including Qwen3-VL, InternVL3, and Bagel, and trained via large-scale supervised fine-tuning under a unified spatial intelligence paradigm spanning visual understanding and generation. We propose a systematic taxonomy of spatial capabilities, curate 8 million high-quality, diverse spatial data samples, and incorporate data debiasing and structured annotation strategies. Experiments demonstrate state-of-the-art performance on spatial intelligence benchmarks—including VSI-Bench, MMSI, and MindCube—while preserving strong general multimodal competence. Notably, we report the first empirical observation of diversity-driven emergent spatial generalization and provide preliminary validation of spatial chain-of-thought reasoning. All models and datasets are publicly released.

Technology Category

Application Category

📝 Abstract
Despite remarkable progress, multimodal foundation models still exhibit surprising deficiencies in spatial intelligence. In this work, we explore scaling up multimodal foundation models to cultivate spatial intelligence within the SenseNova-SI family, built upon established multimodal foundations including visual understanding models (i.e., Qwen3-VL and InternVL3) and unified understanding and generation models (i.e., Bagel). We take a principled approach to constructing high-performing and robust spatial intelligence by systematically curating SenseNova-SI-8M: eight million diverse data samples under a rigorous taxonomy of spatial capabilities. SenseNova-SI demonstrates unprecedented performance across a broad range of spatial intelligence benchmarks: 68.7% on VSI-Bench, 43.3% on MMSI, 85.6% on MindCube, 54.6% on ViewSpatial, and 50.1% on SITE, while maintaining strong general multimodal understanding (e.g., 84.9% on MMBench-En). More importantly, we analyze the impact of data scaling, discuss early signs of emergent generalization capabilities enabled by diverse data training, analyze the risk of overfitting and language shortcuts, present a preliminary study on spatial chain-of-thought reasoning, and validate the potential downstream application. SenseNova-SI is an ongoing project, and this report will be updated continuously. All newly trained multimodal foundation models are publicly released to facilitate further research in this direction.
Problem

Research questions and friction points this paper is trying to address.

Addressing spatial intelligence deficiencies in multimodal foundation models
Scaling up models to cultivate spatial reasoning capabilities systematically
Developing robust spatial intelligence through diverse data curation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Scaling multimodal models for spatial intelligence
Creating eight million diverse spatial data samples
Achieving top performance across spatial benchmarks
🔎 Similar Papers
No similar papers found.
Z
Zhongang Cai
SenseTime Research
R
Ruisi Wang
SenseTime Research
Chenyang Gu
Chenyang Gu
Undergraduate, Peking University
Embodied AIRobotic Manipulation
Fanyi Pu
Fanyi Pu
MMLab@NTU, Singapore
Machine LearningComputer VisionNature Language Processing
J
Junxiang Xu
SenseTime Research
Y
Yubo Wang
SenseTime Research
Wanqi Yin
Wanqi Yin
SenseTime Research
Computer VisionMotion CaptureDigital Human
Z
Zhitao Yang
SenseTime Research
C
Chen Wei
SenseTime Research
Q
Qingping Sun
SenseTime Research
T
Tongxi Zhou
SenseTime Research
J
Jiaqi Li
SenseTime Research
H
Hui En Pang
Nanyang Technological University
O
Oscar Qian
SenseTime Research, Nanyang Technological University
Y
Yukun Wei
SenseTime Research
Z
Zhiqian Lin
SenseTime Research
X
Xuanke Shi
SenseTime Research
K
Kewang Deng
SenseTime Research
X
Xiaoyang Han
SenseTime Research
Z
Zukai Chen
SenseTime Research
X
Xiangyu Fan
SenseTime Research
Hanming Deng
Hanming Deng
Unknown affiliation
Deep LearningObject Detection
Lewei Lu
Lewei Lu
Research Director (We're Hiring, luotto@sensetime.com) @ SenseTime Research
Computer VisionDeep Learning
L
Liang Pan
SenseTime Research
B
Bo Li
Nanyang Technological University