CoCoAFusE: Beyond Mixtures of Experts via Model Fusion

πŸ“… 2025-05-02
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper addresses the dual challenges of poor interpretability and low uncertainty quantification (UQ) accuracy in deep learning for multimodal tasks with covariate-dependent uncertainty. We propose the Competitive/Cooperative Adaptive Fusion of Experts (CoCoAFusE), a novel mixture-of-experts (MoE) framework that introduces *distribution-level fusion*β€”replacing conventional weighted averaging with cooperative probabilistic modeling across expert outputs. This enables locally interpretable, mechanism-aware uncertainty estimation. A key innovation is the elimination of multimodal artifacts in smooth transition regions, thereby improving prediction interval tightness and calibration. Extensive experiments on multiple synthetic and real-world benchmarks demonstrate that CoCoAFusE significantly enhances UQ performance: it yields narrower, better-calibrated prediction intervals while maintaining high predictive accuracy and preserving local interpretability grounded in underlying data mechanisms.

Technology Category

Application Category

πŸ“ Abstract
Many learning problems involve multiple patterns and varying degrees of uncertainty dependent on the covariates. Advances in Deep Learning (DL) have addressed these issues by learning highly nonlinear input-output dependencies. However, model interpretability and Uncertainty Quantification (UQ) have often straggled behind. In this context, we introduce the Competitive/Collaborative Fusion of Experts (CoCoAFusE), a novel, Bayesian Covariates-Dependent Modeling technique. CoCoAFusE builds on the very philosophy behind Mixtures of Experts (MoEs), blending predictions from several simple sub-models (or"experts") to achieve high levels of expressiveness while retaining a substantial degree of local interpretability. Our formulation extends that of a classical Mixture of Experts by contemplating the fusion of the experts' distributions in addition to their more usual mixing (i.e., superimposition). Through this additional feature, CoCoAFusE better accommodates different scenarios for the intermediate behavior between generating mechanisms, resulting in tighter credible bounds on the response variable. Indeed, only resorting to mixing, as in classical MoEs, may lead to multimodality artifacts, especially over smooth transitions. Instead, CoCoAFusE can avoid these artifacts even under the same structure and priors for the experts, leading to greater expressiveness and flexibility in modeling. This new approach is showcased extensively on a suite of motivating numerical examples and a collection of real-data ones, demonstrating its efficacy in tackling complex regression problems where uncertainty is a key quantity of interest.
Problem

Research questions and friction points this paper is trying to address.

Enhances model interpretability and uncertainty quantification in deep learning
Extends Mixtures of Experts by fusing expert distributions for better flexibility
Addresses multimodality artifacts in smooth transitions between generating mechanisms
Innovation

Methods, ideas, or system contributions that make the work stand out.

Bayesian Covariates-Dependent Modeling technique
Competitive/Collaborative Fusion of Experts
Fusion of experts' distributions for flexibility
πŸ”Ž Similar Papers
No similar papers found.
A
Aurelio Raffa Ugolini
Dip. di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 (Italy)
M
M. Tanelli
Dip. di Elettronica, Informazione e Bioingegneria, Politecnico di Milano, 20133 (Italy)
Valentina Breschi
Valentina Breschi
Assistant Professor @ TU/e
System identificationData-driven controlCollaborative estimation and control