🤖 AI Summary
This study challenges the dynamic guidance fidelity of chain-of-thought (CoT) reasoning in soft reasoning tasks—such as analytical and commonsense reasoning—questioning whether CoT reflects genuine internal reasoning or merely post-hoc, unfaithful rationalization. Method: Through dynamic behavioral analysis and attribution-based evaluation, we systematically compare CoT usage across instruction-tuned, dedicated reasoning, and distillation-based large language models. Contribution/Results: We find that CoT yields limited practical improvement for soft reasoning and frequently diverges from the model’s actual inference process. Moreover, the extent of CoT influence and user trust exhibit significant misalignment across model families. To our knowledge, this is the first work to empirically identify a pervasive “guidance failure” phenomenon of CoT in soft reasoning, directly challenging its default assumption as a reliable reasoning mechanism. Our findings provide novel empirical evidence and methodological insights for interpretable AI and trustworthy reasoning evaluation.
📝 Abstract
Recent work has demonstrated that Chain-of-Thought (CoT) often yields limited gains for soft-reasoning problems such as analytical and commonsense reasoning. CoT can also be unfaithful to a model's actual reasoning. We investigate the dynamics and faithfulness of CoT in soft-reasoning tasks across instruction-tuned, reasoning and reasoning-distilled models. Our findings reveal differences in how these models rely on CoT, and show that CoT influence and faithfulness are not always aligned.