When to Trust the Cheap Check: Weak and Strong Verification for Reasoning

πŸ“… 2026-02-19
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the lack of an effective decision mechanism for balancing weak verification (fast but unreliable) and strong verification (reliable but costly) in large language model inference. It formalizes, for the first time, a weak–strong verification framework and proposes a dynamic two-threshold policy that decides, based on the confidence of weak verification, whether to accept, reject, or escalate a query to strong verification. The approach makes no distributional assumptions about the query stream, models, or verifiers, and integrates online learning to rigorously control population-level false acceptance and rejection rates while minimizing reliance on strong verification. Theoretical analysis reveals that the optimal policy inherently exhibits a two-threshold structure and elucidates how calibration and sharpness govern the utility of weak verification.

Technology Category

Application Category

πŸ“ Abstract
Reasoning with LLMs increasingly unfolds inside a broader verification loop. Internally, systems use cheap checks, such as self-consistency or proxy rewards, which we call weak verification. Externally, users inspect outputs and steer the model through feedback until results are trustworthy, which we call strong verification. These signals differ sharply in cost and reliability: strong verification can establish trust but is resource-intensive, while weak verification is fast and scalable but noisy and imperfect. We formalize this tension through weak--strong verification policies, which decide when to accept or reject based on weak verification and when to defer to strong verification. We introduce metrics capturing incorrect acceptance, incorrect rejection, and strong-verification frequency. Over population, we show that optimal policies admit a two-threshold structure and that calibration and sharpness govern the value of weak verifiers. Building on this, we develop an online algorithm that provably controls acceptance and rejection errors without assumptions on the query stream, the language model, or the weak verifier.
Problem

Research questions and friction points this paper is trying to address.

weak verification
strong verification
trust calibration
verification policy
LLM reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

weak verification
strong verification
verification policy
online algorithm
threshold structure
πŸ”Ž Similar Papers
No similar papers found.