🤖 AI Summary
This work addresses negative transfer in multi-source transfer learning, which often arises from heterogeneity among source tasks. Existing approaches typically optimize either source weights or the number of transferred samples in isolation, neglecting their joint optimization. The authors formulate this challenge as a parameter estimation problem based on KL divergence–derived generalization error and propose a unified asymptotic analysis framework to simultaneously optimize both source weights and sample sizes. Theoretically, they establish—for the first time—that after proper weight adaptation, utilizing all available source samples is optimal, and they provide a closed-form solution for the single-source case along with a convex optimization algorithm for the multi-source setting. Extensive experiments on benchmarks such as DomainNet and Office-Home demonstrate significant improvements over strong baselines, corroborating both the theoretical predictions and practical efficacy of the proposed method.
📝 Abstract
Transfer learning plays a vital role in improving model performance in data-scarce scenarios. However, naive uniform transfer from multiple source tasks may result in negative transfer, highlighting the need to properly balance the contributions of heterogeneous sources. Moreover, existing transfer learning methods typically focus on optimizing either the source weights or the amount of transferred samples, while largely neglecting the joint consideration of the other. In this work, we propose a theoretical framework, Unified Optimization of Weights and Quantities (UOWQ), which formulates multi-source transfer learning as a parameter estimation problem grounded in an asymptotic analysis of a Kullback-Leibler divergence-based generalization error measure. The proposed framework jointly determines the optimal source weights and optimal transfer quantities for each source task. Firstly, we prove that using all available source samples is always optimal once the weights are properly adjusted, and we provide a theoretical explanation for this phenomenon. Moreover, to determine the optimal transfer weights, our analysis yields closed-form solutions in the single-source setting and develops a convex optimization-based numerical procedure for the multi-source case. Building on the theoretical results, we further propose practical algorithms for both multi-source transfer learning and multi-task learning settings. Extensive experiments on real-world benchmarks, including DomainNet and Office-Home, demonstrate that UOWQ consistently outperforms strong baselines. The results validate both the theoretical predictions and the practical effectiveness of our framework.