π€ AI Summary
This work addresses the challenge of deploying medical imaging models when fine-tuned on local private data yet evaluated on unlabeled, out-of-distribution test batches from unseen clinical sitesβa scenario where conventional mean ensembling fails due to domain shift. To overcome this, the authors propose an entropy-driven online model fusion method that dynamically constructs a batch-specific ensemble during forward propagation. By decoupling the encoder and classifier head and fusing them separately, the approach achieves efficient unsupervised domain adaptation without requiring backpropagation. Evaluated across nine medical and natural image classification benchmarks, the method consistently outperforms existing techniques under both standard and challenging conditions, while preserving the inference efficiency of a single model.
π Abstract
Model merging under unseen test-time distribution shifts often renders naive strategies, such as mean averaging unreliable. This challenge is especially acute in medical imaging, where models are fine-tuned locally at clinics on private data, producing domain-specific models that differ by scanner, protocol, and population. When deployed at an unseen clinical site, test cases arrive in unlabeled, non-i.i.d. batches, and the model must adapt immediately without labels. In this work, we introduce an entropy-adaptive, fully online model-merging method that yields a batch-specific merged model via only forward passes, effectively leveraging target information. We further demonstrate why mean merging is prone to failure and misaligned under heterogeneous domain shifts. Next, we mitigate encoder classifier mismatch by decoupling the encoder and classification head, merging with separate merging coefficients. We extensively evaluate our method with state-of-the-art baselines using two backbones across nine medical and natural-domain generalization image classification datasets, showing consistent gains across standard evaluation and challenging scenarios. These performance gains are achieved while retaining single-model inference at test-time, thereby demonstrating the effectiveness of our method.