Enabling Agents to Communicate Entirely in Latent Space

πŸ“… 2025-11-12
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Natural language as a communication medium among LLM agents suffers from fundamental limitations: discrete tokenization severely compresses and distorts rich latent state information, hindering collaborative reasoning. To address this, we propose Interlatβ€”the first full-latent-space agent communication paradigm leveraging the final-layer hidden states of LLMs, eliminating natural language generation entirely and enabling direct information compression, transmission, and joint reasoning in continuous latent space. Its core innovation is the first end-to-end, tokenization-free direct latent-state interconnection between agents. Experiments demonstrate that Interlat significantly outperforms fine-tuned chain-of-thought prompting and single-agent baselines, accelerating inference while maintaining or even improving task performance. Moreover, it preserves critical semantic content and reasoning pathways more completely than language-based alternatives.

Technology Category

Application Category

πŸ“ Abstract
While natural language is the de facto communication medium for LLM-based agents, it presents a fundamental constraint. The process of downsampling rich, internal latent states into discrete tokens inherently limits the depth and nuance of information that can be transmitted, thereby hindering collaborative problem-solving. Inspired by human mind-reading, we propose Interlat (Inter-agent Latent Space Communication), a paradigm that leverages the last hidden states of an LLM as a representation of its mind for direct transmission (termed latent communication). An additional compression process further compresses latent communication via entirely latent space reasoning. Experiments demonstrate that Interlat outperforms both fine-tuned chain-of-thought (CoT) prompting and single-agent baselines, promoting more exploratory behavior and enabling genuine utilization of latent information. Further compression not only substantially accelerates inference but also maintains competitive performance through an efficient information-preserving mechanism. We position this work as a feasibility study of entirely latent space inter-agent communication, and our results highlight its potential, offering valuable insights for future research.
Problem

Research questions and friction points this paper is trying to address.

Overcoming information loss from latent state discretization
Enhancing collaborative problem-solving through direct latent transmission
Accelerating inference while preserving compressed latent information
Innovation

Methods, ideas, or system contributions that make the work stand out.

Direct latent space communication between agents
Compression via latent space reasoning for efficiency
Utilizing hidden states as mind representations
πŸ”Ž Similar Papers
No similar papers found.
Z
Zhuoyun Du
Zhejiang University
R
Runze Wang
Taobao & Tmail Group of Alibaba
H
Huiyu Bai
Nanyang Technological University
Zouying Cao
Zouying Cao
Shanghai Jiao Tong University
Natural Language ProcessingLarge Language ModelsReinforcement Learning
Xiaoyong Zhu
Xiaoyong Zhu
Jiangsu University
Electrical MachinesElectrical Vehicle
B
Bo Zheng
Taobao & Tmail Group of Alibaba
W
Wei Chen
Zhejiang University
H
Haochao Ying
Zhejiang University