🤖 AI Summary
This work addresses critical limitations of multimodal large language models (MLLMs) in realistic, sustained dialogues—including memory decay, delayed knowledge updating, error accumulation in reasoning, and weak refusal capability. To systematically evaluate these issues, we introduce MMRC, the first benchmark specifically designed for open-ended, multi-turn multimodal interaction. MMRC is grounded in real-world scenarios and comprises 5,120 multi-turn dialogues and 28,720 human-annotated questions, enabling comprehensive assessment across six core capabilities: information extraction, multi-turn reasoning, knowledge updating, image management, memory recall, and refusal. Extensive experiments reveal significant performance degradation across 20 state-of-the-art MLLMs in long-horizon interactions and identify four prevalent failure patterns. We further propose NOTE-TAKING, a lightweight memory-augmentation strategy, which yields an average 12.3% improvement in key capabilities across six models. MMRC establishes a new standard for evaluating MLLMs in practical deployment and provides an extensible pathway for optimization.
📝 Abstract
Recent multimodal large language models (MLLMs) have demonstrated significant potential in open-ended conversation, generating more accurate and personalized responses. However, their abilities to memorize, recall, and reason in sustained interactions within real-world scenarios remain underexplored. This paper introduces MMRC, a Multi-Modal Real-world Conversation benchmark for evaluating six core open-ended abilities of MLLMs: information extraction, multi-turn reasoning, information update, image management, memory recall, and answer refusal. With data collected from real-world scenarios, MMRC comprises 5,120 conversations and 28,720 corresponding manually labeled questions, posing a significant challenge to existing MLLMs. Evaluations on 20 MLLMs in MMRC indicate an accuracy drop during open-ended interactions. We identify four common failure patterns: long-term memory degradation, inadequacies in updating factual knowledge, accumulated assumption of error propagation, and reluctance to say no. To mitigate these issues, we propose a simple yet effective NOTE-TAKING strategy, which can record key information from the conversation and remind the model during its responses, enhancing conversational capabilities. Experiments across six MLLMs demonstrate significant performance improvements.