🤖 AI Summary
This work addresses the limitations of explicit symbolic spaces in language models—namely linguistic redundancy, discretization bottlenecks, sequential inefficiency, and semantic loss—which constrain computational efficiency and expressive capacity. The study systematically reviews advances in latent space research and introduces, for the first time, a five-dimensional analytical framework tailored to language models: foundations, evolution, mechanisms, capabilities, and outlook. This unified framework integrates architecture, representation, computation, and optimization, while linking technical pathways to higher-order abilities such as reasoning, memory, and embodiment. By synthesizing and categorizing cutting-edge research, the paper elucidates the pivotal role of latent spaces in enhancing model efficiency and capability, and clearly identifies key challenges and promising directions for future work.
📝 Abstract
Latent space is rapidly emerging as a native substrate for language-based models. While modern systems are still commonly understood through explicit token-level generation, an increasing body of work shows that many critical internal processes are more naturally carried out in continuous latent space than in human-readable verbal traces. This shift is driven by the structural limitations of explicit-space computation, including linguistic redundancy, discretization bottlenecks, sequential inefficiency, and semantic loss. This survey aims to provide a unified and up-to-date landscape of latent space in language-based models. We organize the survey into five sequential perspectives: Foundation, Evolution, Mechanism, Ability, and Outlook. We begin by delineating the scope of latent space, distinguishing it from explicit or verbal space and from the latent spaces commonly studied in generative visual models. We then trace the field's evolution from early exploratory efforts to the current large-scale expansion. To organize the technical landscape, we examine existing work through the complementary lenses of mechanism and ability. From the perspective of Mechanism, we identify four major lines of development: Architecture, Representation, Computation, and Optimization. From the perspective of Ability, we show how latent space supports a broad capability spectrum spanning Reasoning, Planning, Modeling, Perception, Memory, Collaboration, and Embodiment. Beyond consolidation, we discuss the key open challenges, and outline promising directions for future research. We hope this survey serves not only as a reference for existing work, but also as a foundation for understanding latent space as a general computational and systems paradigm for next-generation intelligence.