🤖 AI Summary
This work addresses the lack of comprehensive evaluation frameworks for multimodal reasoning in real-world and domain-specific applications. Methodologically, it introduces the MARS2 2025 Multimodal Reasoning Challenge benchmark—the first to systematically extend multimodal reasoning to specialized domains such as advertising video understanding. It releases two novel datasets: Lens (covering 12 everyday scenarios) and AdsQA (focused on advertising video reasoning), and defines three core tracks: real-world visual grounding, spatial-aware visual question answering, and advertising video reasoning. The benchmark integrates techniques spanning visual grounding, spatial relation modeling, and cross-modal reasoning, incorporating over 40 baseline models and 15 team submissions. It attracted 76 international teams, yielding 40+ valid submissions. All data, code, and leaderboards are fully open-sourced, establishing the first reproducible, general-purpose yet domain-inclusive multimodal reasoning evaluation framework.
📝 Abstract
This paper reviews the MARS2 2025 Challenge on Multimodal Reasoning. We aim to bring together different approaches in multimodal machine learning and LLMs via a large benchmark. We hope it better allows researchers to follow the state-of-the-art in this very dynamic area. Meanwhile, a growing number of testbeds have boosted the evolution of general-purpose large language models. Thus, this year's MARS2 focuses on real-world and specialized scenarios to broaden the multimodal reasoning applications of MLLMs. Our organizing team released two tailored datasets Lens and AdsQA as test sets, which support general reasoning in 12 daily scenarios and domain-specific reasoning in advertisement videos, respectively. We evaluated 40+ baselines that include both generalist MLLMs and task-specific models, and opened up three competition tracks, i.e., Visual Grounding in Real-world Scenarios (VG-RS), Visual Question Answering with Spatial Awareness (VQA-SA), and Visual Reasoning in Creative Advertisement Videos (VR-Ads). Finally, 76 teams from the renowned academic and industrial institutions have registered and 40+ valid submissions (out of 1200+) have been included in our ranking lists. Our datasets, code sets (40+ baselines and 15+ participants' methods), and rankings are publicly available on the MARS2 workshop website and our GitHub organization page https://github.com/mars2workshop/, where our updates and announcements of upcoming events will be continuously provided.