Beyond Mixtures and Products for Ensemble Aggregation: A Likelihood Perspective on Generalized Means

📅 2026-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of effectively aggregating predictive distributions in deep ensembles to enhance both performance and reliability. From a log-likelihood perspective, the authors systematically analyze generalized mean-normalized aggregation and establish, for the first time, a unified theoretical framework that characterizes the behavior of aggregation across different orders \( r \). They rigorously prove that the aggregated prediction strictly outperforms any individual model if and only if \( r \in [0,1] \), thereby providing a solid theoretical foundation for the empirical success of linear pooling (\( r=1 \)) and geometric pooling (\( r \to 0 \)). Extensive experiments on image and text classification benchmarks confirm the practical relevance and effectiveness of the proposed theory.

Technology Category

Application Category

📝 Abstract
Density aggregation is a central problem in machine learning, for instance when combining predictions from a Deep Ensemble. The choice of aggregation remains an open question with two commonly proposed approaches being linear pooling (probability averaging) and geometric pooling (logit averaging). In this work, we address this question by studying the normalized generalized mean of order $r \in \mathbb{R} \cup \{-\infty,+\infty\}$ through the lens of log-likelihood, the standard evaluation criterion in machine learning. This provides a unifying aggregation formalism and shows different optimal configurations for different situations. We show that the regime $r \in [0,1]$ is the only range ensuring systematic improvements relative to individual distributions, thereby providing a principled justification for the reliability and widespread practical use of linear ($r=1$) and geometric ($r=0$) pooling. In contrast, we show that aggregation rules with $r \notin [0,1]$ may fail to provide consistent gains with explicit counterexamples. Finally, we corroborate our theoretical findings with empirical evaluations using Deep Ensembles on image and text classification benchmarks.
Problem

Research questions and friction points this paper is trying to address.

density aggregation
ensemble methods
generalized means
log-likelihood
Deep Ensembles
Innovation

Methods, ideas, or system contributions that make the work stand out.

generalized mean
density aggregation
log-likelihood
ensemble methods
pooling
🔎 Similar Papers
2024-03-22arXiv.orgCitations: 7
R
Raphaël Razafindralambo
Université Côte d’Azur, Inria, CNRS, I3S/LJAD, Maasai, Nice, France
R
Rémy Sun
Université Côte d’Azur, Inria, CNRS, I3S/LJAD, Maasai, Nice, France
F
Frédéric Precioso
Université Côte d’Azur, Inria, CNRS, I3S/LJAD, Maasai, Nice, France
Damien Garreau
Damien Garreau
Professor for the Theory of Machine Learning, Julius-Maximilians-Universität Würzburg
Explainable AIensembleschange-point detectioncomparison-based learning
Pierre-Alexandre Mattei
Pierre-Alexandre Mattei
Research scientist, Inria, Université Côte d'Azur
StatisticsMachine learningLatent variable modelsDeep generative models