PRIOT: Pruning-Based Integer-Only Transfer Learning for Embedded Systems

πŸ“… 2025-03-21
πŸ›οΈ IEEE Embedded Systems Letters
πŸ“ˆ Citations: 1
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Traditional fine-tuning fails under static quantization on resource-constrained embedded devices lacking floating-point units (e.g., Raspberry Pi Pico), while dynamic scale computation incurs prohibitive overhead. Method: We propose the first integer-only edge-side transfer learning framework, abandoning weight updates in favor of a novel integer training paradigm based on edge pruning. We introduce Edge-Popup, a heuristic pruning mechanism, and its memory-optimized variant PRIOT-Sβ€”enabling efficient, low-memory integer fine-tuning under static quantization. Contributions/Results: On rotated MNIST and rotated CIFAR-10, our method improves accuracy by 8.08–33.75 percentage points. PRIOT-S reduces memory footprint significantly, with <1% accuracy degradation, and supports C/C++ deployment on ARM Cortex-M0+ microcontrollers. This work establishes the first practical, fully integer, statically quantized fine-tuning approach for ultra-low-power edge devices.

Technology Category

Application Category

πŸ“ Abstract
On-device transfer learning is crucial for adapting a common backbone model to the unique environment of each edge device. Tiny microcontrollers, such as the Raspberry Pi Pico, are key targets for on-device learning but often lack floating-point units, necessitating integer-only training. Dynamic computation of quantization scale factors, which is adopted in former studies, incurs high computational costs. Therefore, this study focuses on integer-only training with static scale factors, which is challenging with existing training methods. We propose a new training method named PRIOT, which optimizes the network by pruning selected edges rather than updating weights, allowing effective training with static scale factors. The pruning pattern is determined by the edge-popup algorithm, which trains a parameter named score assigned to each edge instead of the original parameters and prunes the edges with low scores before inference. Additionally, we introduce a memory-efficient variant, PRIOT-S, which only assigns scores to a small fraction of edges. We implement PRIOT and PRIOT-S on the Raspberry Pi Pico and evaluate their accuracy and computational costs using a tiny CNN model on the rotated MNIST dataset and the VGG11 model on the rotated CIFAR-10 dataset. Our results demonstrate that PRIOT improves accuracy by 8.08 to 33.75 percentage points over existing methods, while PRIOT-S reduces memory footprint with minimal accuracy loss.
Problem

Research questions and friction points this paper is trying to address.

Enables integer-only training on microcontrollers without floating-point units
Reduces computational costs by using static quantization scale factors
Improves accuracy and memory efficiency in on-device transfer learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Pruning edges instead of updating weights
Static scale factors for integer-only training
Memory-efficient variant with selective scoring
πŸ”Ž Similar Papers
No similar papers found.
H
Honoka Anada
Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
S
Sefutsu Ryu
Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
M
Masayuki Usui
Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
T
Tatsuya Kaneko
Graduate School of Information Science and Technology, The University of Tokyo, Tokyo 113-8656, Japan
Shinya Takamaeda-Yamazaki
Shinya Takamaeda-Yamazaki
The University of Tokyo
Computer ArchitectureHigh-Level Synthesis CompilerFPGA SystemAlgorithm/Hardware Co-designMachine Learning Acceleration