🤖 AI Summary
This work addresses the unreliability of multimodal reasoning, which often leads to error propagation and hallucination due to insufficient uncertainty calibration. The authors propose a novel executable reasoning graph framework that models perceptual and logical operations as nodes producing conformal prediction sets, thereby providing calibrated, stepwise uncertainty guarantees. A lightweight controller dynamically schedules tool invocations based on available computational budget. This approach establishes, for the first time, a compositional conformal guarantee mechanism that enables verifiable, evidence-backed reasoning, suppresses error accumulation, and allows controllable trade-offs between computation and accuracy. Experiments demonstrate consistent superiority over strong baselines across document, chart, and multi-image question-answering benchmarks in terms of performance, reliability, and computational efficiency.
📝 Abstract
We present Proof-of-Perception (PoP), a tool-using framework that casts multimodal reasoning as an executable graph with explicit reliability guarantees. Each perception or logic node outputs a conformal set, yielding calibrated, stepwise uncertainty; a lightweight controller uses these certificates to allocate compute under a budget, expanding with extra tool calls only when needed and stopping early otherwise. This grounds answers in verifiable evidence, reduces error compounding and hallucinations, and enables principled accuracy-compute trade-offs. Across document, chart, and multi-image QA benchmarks, PoP improves performance and reliability over strong chain-of-thought, ReAct-style, and program-of-thought baselines while using computation more efficiently.