MARS: Memory-Enhanced Agents with Reflective Self-improvement

📅 2025-03-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitations of large language models (LLMs) in long-term memory retention, constrained context windows, and sustained decision-making under dynamic environments, this paper proposes a tri-agent collaborative framework—comprising User, Assistant, and Verifier agents—that integrates reflective reasoning, iterative feedback-driven policy optimization, and forgetting-aware memory management. Crucially, it introduces, for the first time, the Ebbinghaus forgetting curve into LLM memory retrieval and update mechanisms. We further design a reflective self-improvement mechanism that unifies multi-task coordination and long-horizon information preservation. Experiments on multi-turn dialogue and complex planning benchmarks demonstrate significant improvements: task completion rate and cross-turn consistency increase by 12.6% over strong baselines, validating the framework’s effectiveness in sustaining coherent, adaptive behavior over extended interactions.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have made significant advances in the field of natural language processing, but they still face challenges such as continuous decision-making, lack of long-term memory, and limited context windows in dynamic environments. To address these issues, this paper proposes an innovative framework Memory-Enhanced Agents with Reflective Self-improvement. The MARS framework comprises three agents: the User, the Assistant, and the Checker. By integrating iterative feedback, reflective mechanisms, and a memory optimization mechanism based on the Ebbinghaus forgetting curve, it significantly enhances the agents capabilities in handling multi-tasking and long-span information.
Problem

Research questions and friction points this paper is trying to address.

Addresses continuous decision-making challenges in LLMs
Solves lack of long-term memory in dynamic environments
Overcomes limited context windows for multi-tasking
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates iterative feedback for multi-tasking
Uses reflective mechanisms for self-improvement
Optimizes memory via Ebbinghaus forgetting curve
🔎 Similar Papers
X
Xuechen Liang
East China Jiaotong University, Nanchang, Jiangxi Province, China
Meiling Tao
Meiling Tao
University of Electronic Science and Technology of China
NLPLLM
Yinghui Xia
Yinghui Xia
HKUST(GZ)
Jianhui Wang
Jianhui Wang
University of Electronic Science and Technology of China, Chengdu, China
K
Kun Li
Xiamen University, Xiamen, Fujian 361005, China
Yijin Wang
Yijin Wang
undergraduate,Xidian University
machine learning
J
Jingsong Yang
AutoAgents Co., Ltd., Beijing, China
Tianyu Shi
Tianyu Shi
University of Toronto
Reinforcement learningIntelligent Transportation SystemLarge Language ModelsAILLM agent
Y
Yuantao Wang
Beijing University of Technology, Beijing, China
M
Miao Zhang
Shenzhen International Graduate School, Tsinghua University, Shenzhen, China
Xueqian Wang
Xueqian Wang
Tsinghua University
Information FusionTarget DetectionRadar ImagingImage Processing