AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees

📅 2025-12-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the quadratic computational complexity of self-attention in large language models (LLMs), which impedes efficient long-context processing, and to overcome limitations of existing compression methods—namely poor semantic fidelity, positional bias, and inadequate long-range dependency modeling—this paper proposes an adaptive hierarchical compression framework. The method dynamically identifies information-dense regions to construct a semantic binary tree, where variable-length gist tokens serve as leaf nodes, and introduces a lightweight hierarchical aggregation mechanism. Critically, it preserves both fine-grained local details and global semantic consistency without modifying the frozen backbone model. Compared to state-of-the-art explicit and implicit compression approaches, our framework substantially reduces computational overhead while improving inference efficiency and context fidelity across multiple long-text benchmarks; the additional parameter count is negligible.

Technology Category

Application Category

📝 Abstract
The quadratic complexity of self-attention constrains Large Language Models (LLMs) in processing long contexts, a capability essential for many advanced applications. Context compression aims to alleviate this computational bottleneck while retaining critical semantic information. However, existing approaches often fall short: explicit methods may compromise local detail, whereas implicit methods can suffer from positional biases, information degradation, or an inability to capture long-range semantic dependencies. We propose AdmTree, a novel framework for adaptive, hierarchical context compression with a central focus on preserving high semantic fidelity while maintaining efficiency. AdmTree dynamically segments input based on information density, utilizing gist tokens to summarize variable-length segments as the leaves of a semantic binary tree. This structure, together with a lightweight aggregation mechanism and a frozen backbone LLM (thereby minimizing new trainable parameters), enables efficient hierarchical abstraction of the context. By preserving fine-grained details alongside global semantic coherence, mitigating positional bias, and dynamically adapting to content, AdmTree robustly retains the semantic information of long contexts.
Problem

Research questions and friction points this paper is trying to address.

Compress long contexts to reduce computational complexity in LLMs
Preserve semantic fidelity while mitigating positional biases and information loss
Enable hierarchical abstraction with minimal trainable parameters for efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive hierarchical semantic tree compression
Dynamic segmentation using gist token leaves
Lightweight aggregation with frozen backbone LLM
🔎 Similar Papers
No similar papers found.
Y
Yangning Li
Shenzhen International Graduate School, Tsinghua University
S
Shaoshen Chen
Shenzhen International Graduate School, Tsinghua University
Y
Yinghui Li
Shenzhen International Graduate School, Tsinghua University
Yankai Chen
Yankai Chen
Postdoctoral Associate, Cornell University
Information RetrievalKnowledge MiningLarge Language ModelsAgentic AI
H
Hai-Tao Zheng
Shenzhen International Graduate School, Tsinghua University
H
Hui Wang
Peng Cheng Laboratory
Wenhao Jiang
Wenhao Jiang
GML, Tencent, PolyU
Computer VisionMachine LearningFoundation Models
Philip S. Yu
Philip S. Yu
Professor of Computer Science, University of Illinons at Chicago
Data miningDatabasePrivacy