🤖 AI Summary
Large vision-language models (LVLMs) exhibit limited efficacy of chain-of-thought prompting on complex visual reasoning tasks. Method: We propose a training-free inference enhancement framework that, for the first time, integrates Monte Carlo Tree Search (MCTS) with a multimodal self-reward mechanism to jointly evaluate subproblem utility, answer correctness, and cross-modal cue relevance—without requiring auxiliary discriminative models. We further incorporate reasoning path modeling and test-time scaling to improve inference depth and robustness. Contribution/Results: Our method achieves state-of-the-art performance on three multimodal mathematical reasoning benchmarks. Crucially, it provides the first empirical validation of the test-time scaling law in multimodal reasoning, demonstrating predictable performance gains with increased inference budget. The framework is fully plug-and-play, requires no fine-tuning or external supervision, and generalizes across diverse LVLM backbones.
📝 Abstract
Large Vision-Language Models (LVLMs) have shown exceptional performance in multimodal tasks, but their effectiveness in complex visual reasoning is still constrained, especially when employing Chain-of-Thought prompting techniques. In this paper, we propose VReST, a novel training-free approach that enhances Reasoning in LVLMs through Monte Carlo Tree Search and Self-Reward mechanisms. VReST meticulously traverses the reasoning landscape by establishing a search tree, where each node encapsulates a reasoning step, and each path delineates a comprehensive reasoning sequence. Our innovative multimodal Self-Reward mechanism assesses the quality of reasoning steps by integrating the utility of sub-questions, answer correctness, and the relevance of vision-language clues, all without the need for additional models. VReST surpasses current prompting methods and secures state-of-the-art performance across three multimodal mathematical reasoning benchmarks. Furthermore, it substantiates the efficacy of test-time scaling laws in multimodal tasks, offering a promising direction for future research.