🤖 AI Summary
Deep learning training incurs substantial computational costs, prolonged training time, and high carbon emissions. To address this, this paper proposes a dynamic data reduction method that adaptively prunes input data in real time during image classification training. Unlike static cropping, our approach jointly optimizes data reduction and model accuracy: it evaluates sample importance via gradient sensitivity analysis and integrates online sample selection with adaptive batch-size reduction to dynamically adjust the volume of training data per iteration. Experiments on CIFAR-10, CIFAR-100, and an ImageNet subset demonstrate that the method preserves original model accuracy while reducing both training time and carbon emissions by approximately 50%. This yields significant improvements in training efficiency and environmental sustainability without compromising model performance.
📝 Abstract
We propose a novel method for training a neural network for image classification to reduce input data dynamically, in order to reduce the costs of training a neural network model. As Deep Learning tasks become more popular, their computational complexity increases, leading to more intricate algorithms and models which have longer runtimes and require more input data. The result is a greater cost on time, hardware, and environmental resources. By using data reduction techniques, we reduce the amount of work performed, and therefore the environmental impact of AI techniques, and with dynamic data reduction we show that accuracy may be maintained while reducing runtime by up to 50%, and reducing carbon emission proportionally.