Mem-Gallery: Benchmarking Multimodal Long-Term Conversational Memory for MLLM Agents

๐Ÿ“… 2026-01-07
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 1
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing benchmarks struggle to effectively evaluate the ability of multimodal large language models to retain, organize, and evolve visual and textual information over extended dialogues. To address this gap, this work proposes Mem-Galleryโ€”the first comprehensive benchmark specifically designed for assessing multimodal long-term dialogue memory. It features high-quality, multi-turn image-text conversations and introduces a three-dimensional evaluation framework encompassing memory retrieval and adaptation, memory-based reasoning, and knowledge management. Evaluations of 13 representative memory systems using Mem-Gallery reveal critical bottlenecks in current modelsโ€™ capacity for organizing and reasoning over multimodal memories, underscoring the necessity of explicitly preserving and structurally organizing multimodal information to support robust long-term dialogue capabilities.

Technology Category

Application Category

๐Ÿ“ Abstract
Long-term memory is a critical capability for multimodal large language model (MLLM) agents, particularly in conversational settings where information accumulates and evolves over time. However, existing benchmarks either evaluate multi-session memory in text-only conversations or assess multimodal understanding within localized contexts, failing to evaluate how multimodal memory is preserved, organized, and evolved across long-term conversational trajectories. Thus, we introduce Mem-Gallery, a new benchmark for evaluating multimodal long-term conversational memory in MLLM agents. Mem-Gallery features high-quality multi-session conversations grounded in both visual and textual information, with long interaction horizons and rich multimodal dependencies. Building on this dataset, we propose a systematic evaluation framework that assesses key memory capabilities along three functional dimensions: memory extraction and test-time adaptation, memory reasoning, and memory knowledge management. Extensive benchmarking across thirteen memory systems reveals several key findings, highlighting the necessity of explicit multimodal information retention and memory organization, the persistent limitations in memory reasoning and knowledge management, as well as the efficiency bottleneck of current models.
Problem

Research questions and friction points this paper is trying to address.

multimodal long-term memory
conversational memory
MLLM agents
memory benchmarking
multimodal understanding
Innovation

Methods, ideas, or system contributions that make the work stand out.

multimodal long-term memory
MLLM agents
conversational benchmark
memory reasoning
memory knowledge management
๐Ÿ”Ž Similar Papers
No similar papers found.