🤖 AI Summary
Current multimodal large language model (MLLM) evaluations rely heavily on static benchmarks, which are vulnerable to data contamination and performance saturation—leading to distorted and unreliable assessments. To address these limitations, we propose Knowledge-Enhanced Benchmark Evolution (KBE), a dynamic evaluation framework that introduces graph-structured knowledge modeling for the first time in multimodal assessment, explicitly capturing visual–textual semantic associations. KBE supports dynamic question generation via vision-based re-sampling and external knowledge injection, enabling controllable difficulty adjustment and continuous benchmark evolution. By transcending static benchmark constraints, KBE significantly mitigates data contamination risks, delays benchmark saturation, and enhances fine-grained capability discrimination across MLLMs. Extensive experiments demonstrate that KBE establishes a more reliable, comprehensive, and sustainably evolving evaluation paradigm for multimodal foundation models.
📝 Abstract
The rapid progress of multimodal large language models (MLLMs) calls for more reliable evaluation protocols. Existing static benchmarks suffer from the potential risk of data contamination and saturation, leading to inflated or misleading performance evaluations. To address these issues, we first apply Graph formulation to represent a static or dynamic VQA sample. With the formulation, we propose Knowledge-enhanced Benchmark Evolution(KBE), a dynamic multimodal evaluation framework. KBE first analyzes the original static benchmark, then expands it by integrating multimodal knowledge, transforming the static benchmark into a controllable, dynamic evolving version. Crucially, KBE can both reconstruct questions by Re-selecting visual information in the original image and expand existing questions with external textual knowledge. It enables difficulty-controllable evaluation by adjusting the degree of question exploration. Extensive experiments demonstrate that KBE alleviates the risk of data contamination, data saturation, and provides a more comprehensive assessment of MLLM capabilities.