DecTrain: Deciding When to Train a DNN Online

πŸ“… 2024-10-03
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Monocular depth estimation models suffer from accuracy degradation on out-of-distribution data and prohibitively high computational overhead under continuous online training. To address this, we propose DecTrainβ€”a sparse online training framework that dynamically triggers lightweight self-supervised fine-tuning based on predictive cost-benefit analysis of accuracy gain versus training cost. DecTrain is the first method enabling compact models to approach the generalization performance of large models through intelligent training scheduling. Experiments demonstrate that DecTrain achieves comparable accuracy to full-time online training while reducing training frequency by 44%. When applied to a small model, DecTrain recovers 97% of the accuracy attainable by a large baseline model, with overall computation reduced by 56% relative to that large model.

Technology Category

Application Category

πŸ“ Abstract
Deep neural networks (DNNs) can deteriorate in accuracy when deployment data differs from training data. While performing online training at all timesteps can improve accuracy, it is computationally expensive. We propose DecTrain, a new algorithm that decides when to train a monocular depth DNN online using self-supervision with low overhead. To make the decision at each timestep, DecTrain compares the cost of training with the predicted accuracy gain. We evaluate DecTrain on out-of-distribution data, and find DecTrain maintains accuracy compared to online training at all timesteps, while training only 44% of the time on average. We also compare the recovery of a low inference cost DNN using DecTrain and a more generalizable high inference cost DNN on various sequences. DecTrain recovers the majority (97%) of the accuracy gain of online training at all timesteps while reducing computation compared to the high inference cost DNN which recovers only 66%. With an even smaller DNN, we achieve 89% recovery while reducing computation by 56%. DecTrain enables low-cost online training for a smaller DNN to have competitive accuracy with a larger, more generalizable DNN at a lower overall computational cost.
Problem

Research questions and friction points this paper is trying to address.

Monocular Depth Perception
Online Learning
Computational Cost
Innovation

Methods, ideas, or system contributions that make the work stand out.

DecTrain
Online Learning
Computational Efficiency
πŸ”Ž Similar Papers
No similar papers found.
Z
Zih-Sing Fu
Massachusetts Institute of Technology, Cambridge, MA 02139, USA
S
Soumya Sudhakar
Massachusetts Institute of Technology, Cambridge, MA 02139, USA
S
S. Karaman
Massachusetts Institute of Technology, Cambridge, MA 02139, USA
Vivienne Sze
Vivienne Sze
Professor, EECS at MIT
VLSILow-Power DesignMachine LearningRoboticsVideo Coding