🤖 AI Summary
This work addresses the Test-time Compute Optimization (TCO) problem: improving answer accuracy of large language models under a fixed inference budget, specifically by decoupling external generation and selection phases. We propose Compute-Aware Tree Search (CATS), an actor-critic–style framework that jointly performs multi-path generation, tree search, and reinforcement learning. Crucially, CATS is the first to establish a PAC-Bayes theoretical analysis linking reward model (PRM) generalization error to sample complexity, enabling dynamic budget allocation guided by reward distribution statistics and sparsity. The method balances accuracy and efficiency without increasing model parameters or training overhead. On the MATH and AIME benchmarks, CATS significantly outperforms existing external test-time search methods—achieving substantial accuracy gains under identical compute budgets. These results empirically validate that controlling PRM generalization error is critical for effective TCO.
📝 Abstract
External test-time reasoning enhances large language models (LLMs) by decoupling generation and selection. At inference time, the model generates multiple reasoning paths, and an auxiliary process reward model (PRM) is used to score and select the best one. A central challenge in this setting is test-time compute optimality (TCO), i.e., how to maximize answer accuracy under a fixed inference budget. In this work, we establish a theoretical framework to analyze how the generalization error of the PRM affects compute efficiency and reasoning performance. Leveraging PAC-Bayes theory, we derive generalization bounds and show that a lower generalization error of PRM leads to fewer samples required to find correct answers. Motivated by this analysis, we propose Compute-Aware Tree Search (CATS), an actor-critic framework that dynamically controls search behavior. The actor outputs sampling hyperparameters based on reward distributions and sparsity statistics, while the critic estimates their utility to guide budget allocation. Experiments on the MATH and AIME benchmarks with various LLMs and PRMs demonstrate that CATS consistently outperforms other external TTS methods, validating our theoretical predictions.