🤖 AI Summary
This study systematically investigates the key factors in data curation for multimodal reasoning under fixed model architectures and training protocols. Framed within the NeurIPS 2025 DCVLR Challenge, the work proposes a difficulty-aware sampling strategy grounded in alignment with foundational datasets and conducts ablation studies to assess the impact of data scale, diversity, and synthetic augmentation. The findings reveal that sample difficulty is the dominant driver of performance gains; merely increasing data volume reduces variance without necessarily improving accuracy, while data diversity and synthetic augmentation offer limited benefits. The proposed approach secured first place in the challenge, underscoring the critical role of alignment and difficulty-aware sampling in data-efficient multimodal reasoning.
📝 Abstract
We study data curation for multimodal reasoning through the NeurIPS 2025 Data Curation for Vision-Language Reasoning (DCVLR) challenge, which isolates dataset selection by fixing the model and training protocol. Using a compact curated dataset derived primarily from Walton Multimodal Cold Start, our submission placed first in the challenge. Through post-competition ablations, we show that difficulty-based example selection on an aligned base dataset is the dominant driver of performance gains. Increasing dataset size does not reliably improve mean accuracy under the fixed training recipe, but mainly reduces run-to-run variance, while commonly used diversity and synthetic augmentation heuristics provide no additional benefit and often degrade performance. These results characterize DCVLR as a saturation-regime evaluation and highlight the central role of alignment and difficulty in data-efficient multimodal reasoning.