Learning to Decide with Just Enough: Information-Theoretic Context Summarization for CDMPs

📅 2025-10-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address weak generalization, high computational overhead, and unstable performance of Contextual Markov Decision Processes (CMDPs) under high-dimensional or unstructured contexts, this paper proposes an information-theoretic framework for contextual semantic compression. Methodologically, it leverages large language models to extract low-dimensional, semantically rich context summaries that enhance state representation while preserving decision-critical information. Theoretically, it introduces the novel concept of *approximate contextual sufficiency*, establishing a formal regret bound for CMDPs grounded in a delay–entropy trade-off. Empirically, the approach achieves significant improvements in reward, success rate, and sample efficiency across diverse tasks—including discrete/continuous control, visual navigation, and recommendation—while reducing inference latency and memory footprint. These results demonstrate its effectiveness, scalability, and interpretability in resource-constrained settings.

Technology Category

Application Category

📝 Abstract
Contextual Markov Decision Processes (CMDPs) offer a framework for sequential decision-making under external signals, but existing methods often fail to generalize in high-dimensional or unstructured contexts, resulting in excessive computation and unstable performance. We propose an information-theoretic summarization approach that uses large language models (LLMs) to compress contextual inputs into low-dimensional, semantically rich summaries. These summaries augment states by preserving decision-critical cues while reducing redundancy. Building on the notion of approximate context sufficiency, we provide, to our knowledge, the first regret bounds and a latency-entropy trade-off characterization for CMDPs. Our analysis clarifies how informativeness impacts computational cost. Experiments across discrete, continuous, visual, and recommendation benchmarks show that our method outperforms raw-context and non-context baselines, improving reward, success rate, and sample efficiency, while reducing latency and memory usage. These findings demonstrate that LLM-based summarization offers a scalable and interpretable solution for efficient decision-making in context-rich, resource-constrained environments.
Problem

Research questions and friction points this paper is trying to address.

Compressing high-dimensional contexts into low-dimensional summaries for decision-making
Addressing computational inefficiency and instability in contextual Markov processes
Balancing information preservation with computational cost in resource-constrained environments
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLMs compress contextual inputs into summaries
Summaries preserve decision-critical cues and reduce redundancy
Method improves reward, efficiency, and reduces latency
🔎 Similar Papers
No similar papers found.
Peidong Liu
Peidong Liu
Westlake University
3D computer visionRobotics
J
Junjiang Lin
University of Toronto
Shaowen Wang
Shaowen Wang
Professor, University of Illinois Urbana-Champaign
CyberGISGeospatial Data ScienceSpatial AISpatial AnalysisSustainability
Y
Yao Xu
University of Illinois at Urbana-Champaign
H
Haiqing Li
University of Texas at Arlington
X
Xuhao Xie
University of Texas at Arlington
Siyi Wu
Siyi Wu
University of Toronto
Climate InformaticsHuman-Computer InteractionHuman-AI Collaboration
H
Hao Li
University of Arizona