Learning How to Remember: A Meta-Cognitive Management Method for Structured and Transferable Agent Memory

📅 2026-01-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of existing large language model agents in long-horizon decision-making, which rely on fixed memory representations and consequently exhibit poor generalization under distributional shifts. To overcome this, the authors propose Meta-Cognitive Memory Abstraction (MCMA), a novel framework that formulates memory management as a learnable metacognitive capability. MCMA decouples task execution from memory by freezing the task model and introducing a trainable memory collaborator, enabling the construction of a multi-level abstract memory hierarchy. A task-similarity-driven mechanism selectively reuses stored memories, and the memory collaborator is trained via Direct Preference Optimization. Experiments demonstrate that MCMA significantly outperforms multiple baselines on ALFWorld, ScienceWorld, and BabyAI, achieving superior performance in task completion, out-of-distribution generalization, and cross-task transfer.

Technology Category

Application Category

📝 Abstract
Large language model (LLM) agents increasingly rely on accumulated memory to solve long-horizon decision-making tasks. However, most existing approaches store memory in fixed representations and reuse it at a single or implicit level of abstraction, which limits generalization and often leads to negative transfer when distribution shift. This paper proposes the Meta-Cognitive Memory Abstraction method (MCMA), which treats memory abstraction as a learnable cognitive skill rather than a fixed design choice. MCMA decouples task execution from memory management by combining a frozen task model with a learned memory copilot. The memory copilot is trained using direct preference optimization, it determines how memories should be structured, abstracted, and reused. Memories are further organized into a hierarchy of abstraction levels, enabling selective reuse based on task similarity. When no memory is transferable, MCMA transfers the ability to abstract and manage memory by transferring the memory copilot. Experiments on ALFWorld, ScienceWorld, and BabyAI demonstrate substantial improvements in performance, out-of-distribution generalization, and cross-task transfer over several baselines.
Problem

Research questions and friction points this paper is trying to address.

agent memory
memory abstraction
transfer learning
distribution shift
generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Meta-Cognitive Memory Abstraction
Memory Copilot
Direct Preference Optimization
Hierarchical Memory Abstraction
Cross-Task Transfer
🔎 Similar Papers
No similar papers found.
S
Sirui Liang
Institute of Automation, CAS
Pengfei Cao
Pengfei Cao
Institute of Automation, Chinese Academy of Sciences
Natural Language ProcessingLarge Language ModelsInformation Extraction
Jian Zhao
Jian Zhao
Zhongguancun Institute of Artificial Intelligence
Reinforcement LearningMulti-Agent System
W
Wenhao Teng
Department of Gastrointestinal Surgery, Fujian Provincial Cancer Hospital
X
Xiangwen Liao
College of Computer and Data Science, Fuzhou University
J
Jun Zhao
Institute of Automation, CAS
K
Kang Liu
Institute of Automation, CAS