🤖 AI Summary
Current open-source reasoning models (e.g., DeepSeek-R1-Distill series, QwQ-32B) exhibit substantial instability in mathematical, scientific, and programming benchmark evaluations; minor variations in prompt design, sampling strategies, scoring protocols, or data preprocessing induce large score fluctuations, undermining the reproducibility of claimed performance gains. Method: We conduct the first systematic investigation into how evaluation design biases can be strategically exploited to inflate reported reasoning capabilities, and propose a rigorous, reproducibility-centered evaluation paradigm. Through multidimensional controlled experiments, we empirically analyze key assessment variables—including prompt formatting, decoding parameters, and answer extraction heuristics—to quantify their impact on measured performance. Contribution/Results: Our analysis reveals that prevailing benchmarks lack robustness: under standardized evaluation conditions, purported advantages of multiple models significantly diminish or vanish entirely. This work establishes a methodological foundation and practical guidelines for reliable LLM reasoning evaluation.
📝 Abstract
Reasoning models represented by the Deepseek-R1-Distill series have been widely adopted by the open-source community due to their strong performance in mathematics, science, programming, and other domains. However, our study reveals that their benchmark evaluation results are subject to significant fluctuations caused by various factors. Subtle differences in evaluation conditions can lead to substantial variations in results. Similar phenomena are observed in other open-source inference models fine-tuned based on the Deepseek-R1-Distill series, as well as in the QwQ-32B model, making their claimed performance improvements difficult to reproduce reliably. Therefore, we advocate for the establishment of a more rigorous paradigm for model performance evaluation and present our empirical assessments of the Deepseek-R1-Distill series models.