🤖 AI Summary
This work identifies a systematic deficiency in multimodal large language models (MLLMs): their inability to effectively compose cross-modal skills—specifically, to sequentially coordinate vision and language capabilities for composite tasks requiring structured multimodal skill dependencies.
Method: The authors introduce three benchmark task categories with explicit skill dependency graphs, compare direct prompting against two-stage cascaded inference, and propose a hybrid approach combining chain-of-thought prompting with lightweight, skill-composition-oriented fine-tuning.
Contribution/Results: Experiments show that while the proposed method significantly improves performance, all mainstream MLLMs exhibit a persistent and non-negligible compositional generalization gap. This study is the first to systematically characterize the cross-modal skill composition bottleneck in MLLMs, establishing a novel, structured evaluation benchmark and a reproducible methodology for assessing and enhancing the compositional reasoning capabilities of multimodal models.
📝 Abstract
Skill composition is the ability to combine previously learned skills to solve new tasks. As neural networks acquire increasingly complex skills during their pretraining, it is not clear how successfully they can compose them. In this paper, we focus on Multimodal Large Language Models (MLLM), and study their ability to compose skills across modalities. To this end, we design three evaluation tasks which can be solved sequentially composing two modality-dependent skills, and evaluate several open MLLMs under two main settings: i) prompting the model to directly solve the task, and ii) using a two-step cascaded inference approach, which manually enforces the composition of the two skills for a given task. Even with these straightforward compositions, we find that all evaluated MLLMs exhibit a significant cross-modality skill composition gap. To mitigate the aforementioned gap, we explore two alternatives: i) use chain-of-thought prompting to explicitly instruct MLLMs for skill composition and ii) a specific fine-tuning recipe to promote skill composition. Although those strategies improve model performance, they still exhibit significant skill composition gaps, suggesting that more research is needed to improve cross-modal skill composition in MLLMs.