🤖 AI Summary
Conventional static core-set selection fails to adapt to the heterogeneous requirements across different training stages. Method: This paper proposes a dynamic multi-objective adaptive core-set selection framework that dynamically switches sampling strategies according to training progression—emphasizing class balance in early stages, feature diversity in mid-stages, and prediction uncertainty in late stages—thereby enabling the first training-process-aware, multi-objective co-optimization. Contribution/Results: We theoretically establish a (1−1/e)-approximation guarantee. By integrating submodular optimization, active learning, and representation analysis, our method achieves O(n log n) computational efficiency. Empirically, it attains full-dataset accuracy on multiple benchmarks while significantly reducing memory overhead. Moreover, it is the first work to quantitatively characterize the dynamic evolution of data utility throughout training.
📝 Abstract
We present Mode(Multi-Objective adaptive Data Efficiency), a framework that dynamically combines coreset selection strategies based on their evolving contribution to model performance. Unlike static methods, mode adapts selection criteria to training phases: emphasizing class balance early, diversity during representation learning, and uncertainty at convergence. We show that MODE achieves (1-1/e)-approximation with O(n log n) complexity and demonstrates competitive accuracy while providing interpretable insights into data utility evolution. Experiments show mode reduces memory requirements