🤖 AI Summary
To address the challenges of heterogeneous multimodal data fusion and high computational overhead in resource-constrained embodied robotics scenarios, this paper proposes LCMF, a lightweight cascaded attention framework. LCMF innovatively integrates cross-attention with selective state space models (SSMs) and introduces a multi-level cross-modal parameter-sharing mechanism to achieve semantic complementary alignment between images/videos and text. Compared to mainstream approaches, LCMF reduces model parameters to 166.51M (image–text) and 219M (video–text), and cuts FLOPs by 4.35×. On visual question answering (VQA), it achieves 74.29% accuracy; on embodied question answering (EQA) with video inputs, its performance matches that of large language model–based agents. These results demonstrate substantial improvements in cross-modal understanding efficiency and deployment feasibility for edge-constrained robotic systems.
📝 Abstract
Multimodal semantic learning plays a critical role in embodied intelligence, especially when robots perceive their surroundings, understand human instructions, and make intelligent decisions. However, the field faces technical challenges such as effective fusion of heterogeneous data and computational efficiency in resource-constrained environments. To address these challenges, this study proposes the lightweight LCMF cascaded attention framework, introducing a multi-level cross-modal parameter sharing mechanism into the Mamba module. By integrating the advantages of Cross-Attention and Selective parameter-sharing State Space Models (SSMs), the framework achieves efficient fusion of heterogeneous modalities and semantic complementary alignment. Experimental results show that LCMF surpasses existing multimodal baselines with an accuracy of 74.29% in VQA tasks and achieves competitive mid-tier performance within the distribution cluster of Large Language Model Agents (LLM Agents) in EQA video tasks. Its lightweight design achieves a 4.35-fold reduction in FLOPs relative to the average of comparable baselines while using only 166.51M parameters (image-text) and 219M parameters (video-text), providing an efficient solution for Human-Robot Interaction (HRI) applications in resource-constrained scenarios with strong multimodal decision generalization capabilities.