Universal computation is intrinsic to language model decoding

📅 2026-01-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates whether large language models possess universal computational capabilities and whether such capabilities arise from training or are inherent to their architecture. Through formal analysis of the autoregressive decoding mechanism, the study establishes—for the first time—that even randomly initialized, untrained language models are theoretically Turing-complete. This demonstrates that universal computation stems from intrinsic properties of the decoding process rather than being acquired through training. The research reframes the problem of computational controllability as one of prompt programmability, revealing that the primary role of training is to enhance the model’s ability to reliably and efficiently invoke its innate computational capacities via natural language prompts.

Technology Category

Application Category

📝 Abstract
Language models now provide an interface to express and often solve general problems in natural language, yet their ultimate computational capabilities remain a major topic of scientific debate. Unlike a formal computer, a language model is trained to autoregressively predict successive elements in human-generated text. We prove that chaining a language model's autoregressive output is sufficient to perform universal computation. That is, a language model can simulate the execution of any algorithm on any input. The challenge of eliciting desired computational behaviour can thus be reframed in terms of programmability: the ease of finding a suitable prompt. Strikingly, we demonstrate that even randomly initialized language models are capable of universal computation before training. This implies that training does not give rise to computational expressiveness -- rather, it improves programmability, enabling a natural language interface for accessing these intrinsic capabilities.
Problem

Research questions and friction points this paper is trying to address.

universal computation
language models
computational capabilities
programmability
autoregressive decoding
Innovation

Methods, ideas, or system contributions that make the work stand out.

universal computation
language models
autoregressive decoding
programmability
random initialization
🔎 Similar Papers
2024-03-04Computer Vision and Pattern RecognitionCitations: 3