π€ AI Summary
Current language models lack intrinsic risk-awareness mechanisms, rendering them incapable of autonomously deciding when to answer, refuse, or abstain from speculative responses under uncertainty.
Method: We formally define the βrisk-aware decision-makingβ task and propose a dynamic, skill-decomposition-based framework that employs chain-of-thought prompting to jointly model multi-level reasoning and refusal behavior, complemented by a risk-sensitive evaluation protocol.
Contribution/Results: Experiments reveal that even state-of-the-art reasoning models reliably perform risk-informed decisions only when guided by explicit prompt chains; static refusal strategies exhibit fundamental limitations. Our work establishes the first principled benchmark for risk-adaptive decision-making, providing both theoretical foundations and scalable technical pathways toward building truly autonomous and safety-aware intelligent agents.
π Abstract
Knowing when to answer or refuse is crucial for safe and reliable decision-making language agents. Although prior work has introduced refusal strategies to boost LMs' reliability, how these models adapt their decisions to different risk levels remains underexplored. We formalize the task of risk-aware decision-making, expose critical weaknesses in existing LMs, and propose skill-decomposition solutions to mitigate them. Our findings show that even cutting-edge LMs--both regular and reasoning models--still require explicit prompt chaining to handle the task effectively, revealing the challenges that must be overcome to achieve truly autonomous decision-making agents.