π€ AI Summary
To address the strong modality heterogeneity and fusion difficulty in multimodal 3D understanding, this paper proposes a Mixture-of-Experts (MoE)-based Transformer architecture. Methodologically, it introduces specialized expert networks for unimodal processing and cross-modal interaction, employs a Top-1 gating mechanism for efficient feature routing, and designs a multimodal information aggregation module; additionally, it adopts a progressive pretraining strategy guided by 2D semantic priors to improve initialization quality and cross-modal adaptation. The core contribution lies in the systematic integration of MoE into 3D multimodal fusion, significantly enhancing the modelβs capacity to capture modality-specific characteristics and discrepancies. The approach achieves state-of-the-art performance across four mainstream 3D understanding benchmarks, notably outperforming prior work by 6.1 mIoU on Multi3DRefer, thereby demonstrating both effectiveness and generalizability.
π Abstract
Multi-modal 3D understanding is a fundamental task in computer vision. Previous multi-modal fusion methods typically employ a single, dense fusion network, struggling to handle the significant heterogeneity and complexity across modalities, leading to suboptimal performance. In this paper, we propose MoE3D, which integrates Mixture of Experts (MoE) into the multi-modal learning framework. The core is that we deploy a set of specialized "expert" networks, each adept at processing a specific modality or a mode of cross-modal interaction. Specifically, the MoE-based transformer is designed to better utilize the complementary information hidden in the visual features. Information aggregation module is put forward to further enhance the fusion performance. Top-1 gating is employed to make one expert process features with expert groups, ensuring high efficiency. We further propose a progressive pre-training strategy to better leverage the semantic and 2D prior, thus equipping the network with good initialization. Our MoE3D achieves competitive performance across four prevalent 3D understanding tasks. Notably, our MoE3D surpasses the top-performing counterpart by 6.1 mIoU on Multi3DRefer.