Answering the Unanswerable Is to Err Knowingly: Analyzing and Mitigating Abstention Failures in Large Reasoning Models

📅 2025-08-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large reasoning models (LRMs) frequently fail to abstain appropriately when confronted with intrinsically unsolvable problems—e.g., those with insufficient conditions—exposing a systemic inconsistency between internal cognitive states and external responses. This paper presents the first systematic characterization of this misalignment. We propose a lightweight, two-stage method: (1) cognitive monitoring, which models the model’s internal uncertainty via auxiliary calibration; and (2) dynamic output intervention, which guides the model to proactively abstain during inference for unsolvable queries. Our approach requires only minimal labeled data and low-cost fine-tuning, without architectural modifications. Experiments demonstrate a substantial 32.7% increase in abstention rate on unsolvable problems, while preserving near-original performance on standard complex reasoning benchmarks (accuracy drop <0.5%). This work establishes a new paradigm for building trustworthy, interpretable reasoning AI systems grounded in calibrated self-awareness and principled abstention.

Technology Category

Application Category

📝 Abstract
Large reasoning models (LRMs) have shown remarkable progress on complex reasoning tasks. However, some questions posed to LRMs are inherently unanswerable, such as math problems lacking sufficient conditions. We find that LRMs continually fail to provide appropriate abstentions when confronted with these unanswerable questions. In this paper, we systematically analyze, investigate, and resolve this issue for trustworthy AI. We first conduct a detailed analysis of the distinct response behaviors of LRMs when facing unanswerable questions. Then, we show that LRMs possess sufficient cognitive capabilities to recognize the flaws in these questions. However, they fail to exhibit appropriate abstention behavior, revealing a misalignment between their internal cognition and external response. Finally, to resolve this issue, we propose a lightweight, two-stage method that combines cognitive monitoring with inference-time intervention. Experimental results demonstrate that our method significantly improves the abstention rate while maintaining the overall reasoning performance.
Problem

Research questions and friction points this paper is trying to address.

Analyzing abstention failures in large reasoning models
Mitigating inappropriate responses to unanswerable questions
Aligning internal cognition with external response behavior
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cognitive monitoring for abstention detection
Inference-time intervention to align responses
Lightweight two-stage method improves abstention
🔎 Similar Papers
No similar papers found.
Y
Yi Liu
State Key Laboratory for Novel Software Technology, Nanjing University, China
X
Xiangyu Liu
State Key Laboratory for Novel Software Technology, Nanjing University, China
Zequn Sun
Zequn Sun
Nanjing University
Knowledge GraphLarge Language Model
W
Wei Hu
State Key Laboratory for Novel Software Technology, Nanjing University, China; National Institute of Healthcare Data Science, Nanjing University, China