MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models

📅 2025-05-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current large language models (LLMs) lack a unified, structured memory architecture—relying solely on parametric memory (learned weights) and transient activation memory (contextual states)—while external memory approaches like retrieval-augmented generation (RAG) lack lifecycle management and multimodal support, hindering long-term knowledge evolution. To address this, we propose MemOS, a memory operating system for LLMs that elevates memory to a first-class runtime resource, unifying parametric, activation-state, and plaintext memory. Its core contributions are: (1) a standardized memory abstraction, MemCube; (2) end-to-end memory lifecycle governance and cross-modal integration; and (3) a centralized memory execution framework. Experiments demonstrate that MemOS significantly enhances LLMs’ capabilities in long-term knowledge evolution, personalized adaptation, and cross-platform collaboration. By enabling persistent, structured, and multimodal memory management, MemOS establishes a foundational paradigm for continual learning and AGI infrastructure.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have emerged as foundational infrastructure in the pursuit of Artificial General Intelligence (AGI). Despite their remarkable capabilities in language perception and generation, current LLMs fundamentally lack a unified and structured architecture for handling memory. They primarily rely on parametric memory (knowledge encoded in model weights) and ephemeral activation memory (context-limited runtime states). While emerging methods like Retrieval-Augmented Generation (RAG) incorporate plaintext memory, they lack lifecycle management and multi-modal integration, limiting their capacity for long-term knowledge evolution. To address this, we introduce MemOS, a memory operating system designed for LLMs that, for the first time, elevates memory to a first-class operational resource. It builds unified mechanisms for representation, organization, and governance across three core memory types: parametric, activation, and plaintext. At its core is the MemCube, a standardized memory abstraction that enables tracking, fusion, and migration of heterogeneous memory, while offering structured, traceable access across tasks and contexts. MemOS establishes a memory-centric execution framework with strong controllability, adaptability, and evolvability. It fills a critical gap in current LLM infrastructure and lays the groundwork for continual adaptation, personalized intelligence, and cross-platform coordination in next-generation intelligent systems.
Problem

Research questions and friction points this paper is trying to address.

LLMs lack unified memory architecture for handling diverse memory types
Current methods miss lifecycle management and multi-modal memory integration
MemOS introduces structured memory governance for LLM evolution and control
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unified memory architecture for LLMs
MemCube abstraction for heterogeneous memory
Memory-centric execution framework
Zhiyu Li
Zhiyu Li
Tianjin University
Robust controlattitude control
S
Shichao Song
Renmin University of China
H
Hanyu Wang
Renmin University of China
S
Simin Niu
Renmin University of China
Ding Chen
Ding Chen
Postdoctoral Scholar, University of Texas Southwestern Medical Center
J
Jiawei Yang
MemTensor (Shanghai) Technology Co., Ltd.
Chenyang Xi
Chenyang Xi
Beijing Institute of Technology
Reinforcement Learning
H
Huayi Lai
Renmin University of China
J
Jihao Zhao
Renmin University of China
Y
Yezhaohui Wang
MemTensor (Shanghai) Technology Co., Ltd.
J
Junpeng Ren
MemTensor (Shanghai) Technology Co., Ltd.
Z
Zehao Lin
MemTensor (Shanghai) Technology Co., Ltd.
Jiahao Huo
Jiahao Huo
Tongji University
Multimodal AIInterpretabilityNatural Language Processing
T
Tianyi Chen
Shanghai Jiao Tong University
K
Kai Chen
MemTensor (Shanghai) Technology Co., Ltd.
K
Ke-Rong Li
Shanghai Jiao Tong University
Z
Zhiqiang Yin
Renmin University of China
Q
Qingchen Yu
MemTensor (Shanghai) Technology Co., Ltd.
B
Bo Tang
MemTensor (Shanghai) Technology Co., Ltd.
H
Hongkang Yang
MemTensor (Shanghai) Technology Co., Ltd.
Z
Zhi-Qin John Xu
Shanghai Jiao Tong University
Feiyu Xiong
Feiyu Xiong
MemTensor (Shanghai) Technology Co., Ltd.
Machine LearningNLPLLM