Therefore I am. I Think

πŸ“… 2026-04-01
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This study investigates whether large language reasoning models form decision intent prior to generating reasoning chains. Employing linear probing and activation steering techniques, the work provides the first evidence that models encode detectable decision representations before emitting any reasoning tokens. Experiments demonstrate that linear probes can decode final decisions with high confidence at this early stage, and that steering intermediate activations reverses model behavior in 7%–79% of samples, with subsequent reasoning chains often post-hoc rationalizing the altered decisions. These findings challenge the prevailing assumption that models β€œthink before deciding,” instead revealing a phenomenon of decision precommitment embedded within the reasoning process.
πŸ“ Abstract
We consider the question: when a large language reasoning model makes a choice, did it think first and then decide to, or decide first and then think? In this paper, we present evidence that detectable, early-encoded decisions shape chain-of-thought in reasoning models. Specifically, we show that a simple linear probe successfully decodes tool-calling decisions from pre-generation activations with very high confidence, and in some cases, even before a single reasoning token is produced. Activation steering supports this causally: perturbing the decision direction leads to inflated deliberation, and flips behavior in many examples (between 7 - 79% depending on model and benchmark). We also show through behavioral analysis that, when steering changes the decision, the chain-of-thought process often rationalizes the flip rather than resisting it. Together, these results suggest that reasoning models can encode action choices before they begin to deliberate in text.
Problem

Research questions and friction points this paper is trying to address.

large language models
reasoning
decision-making
chain-of-thought
pre-generation activations
Innovation

Methods, ideas, or system contributions that make the work stand out.

chain-of-thought
decision encoding
activation steering
linear probing
reasoning models
πŸ”Ž Similar Papers
No similar papers found.
E
Esakkivel Esakkiraja
Khoury College of Computer Sciences, Northeastern University
Sai Rajeswar
Sai Rajeswar
Staff Research Scientist, Adjunct Professor, Mila, ServiceNow
machine learninggenerative modelsreinforcement learning
D
Denis Akhiyarov
ServiceNow
R
Rajagopal Venkatesaramani
Khoury College of Computer Sciences, Northeastern University