🤖 AI Summary
Current large language models (LLMs) are fundamentally constrained by autoregressive architectures and statistical modeling paradigms, rendering them incapable of decisional, intrinsically goal-directed “thinking” in feature space. Method: We formally define computable thinking for LLMs—requiring goal-driven state evolution and strategic state transitions within latent space—and rigorously prove, via cognitive process modeling and computability analysis, that mainstream architectures (e.g., Transformer) cannot satisfy this definition in principle. We then propose three foundational principles for a “Thought-Ready” architecture: (i) controllable latent-space dynamics, (ii) explicit goal embedding and representation, and (iii) intervenable reasoning paths. Contribution/Results: This work establishes the first formal, computability-grounded theory of thinking for LLMs, refutes the theoretical feasibility of thinking in existing architectures, and provides both a rigorous theoretical foundation and a concrete design paradigm for next-generation AI systems endowed with genuine reasoning agency.
📝 Abstract
This paper elucidates that current state-of-the-art Large Language Models (LLMs) are fundamentally incapable of making decisions or developing"thoughts"within the feature space due to their architectural constraints. We establish a definition of"thought"that encompasses traditional understandings of that term and adapt it for application to LLMs. We demonstrate that the architectural design and language modeling training methodology of contemporary LLMs inherently preclude them from engaging in genuine thought processes. Our primary focus is on this theoretical realization rather than practical insights derived from experimental data. Finally, we propose solutions to enable thought processes within the feature space and discuss the broader implications of these architectural modifications.