ShiftAddNet: A Hardware-Inspired Deep Network

📅 2020-10-24
🏛️ Neural Information Processing Systems
📈 Citations: 72
Influential: 15
📄 PDF
🤖 AI Summary
To address the high energy consumption of deep neural networks (DNNs) on edge devices—primarily caused by computationally expensive multiplication operations—this paper proposes ShiftAddNet, a novel neural network architecture that entirely eliminates multiplications, relying solely on bit-shifts and additions. Its core innovation is the first fully shift-add parameterization scheme, enabling end-to-end energy-efficiency–aware training and fine-grained, adjustable trade-offs between accuracy and efficiency. This design significantly enhances model robustness to quantization and pruning. Evaluated on FPGA hardware with rigorous energy quantification, ShiftAddNet reduces hardware-quantized energy consumption by over 80% on image classification tasks compared to standard DNNs, while maintaining or even improving accuracy. The architecture thus offers a practical, hardware-efficient alternative for ultra-low-power edge inference without sacrificing performance.
📝 Abstract
Multiplication (e.g., convolution) is arguably a cornerstone of modern deep neural networks (DNNs). However, intensive multiplications cause expensive resource costs that challenge DNNs' deployment on resource-constrained edge devices, driving several attempts for multiplication-less deep networks. This paper presented ShiftAddNet, whose main inspiration is drawn from a common practice in energy-efficient hardware implementation, that is, multiplication can be instead performed with additions and logical bit-shifts. We leverage this idea to explicitly parameterize deep networks in this way, yielding a new type of deep network that involves only bit-shift and additive weight layers. This hardware-inspired ShiftAddNet immediately leads to both energy-efficient inference and training, without compromising the expressive capacity compared to standard DNNs. The two complementary operation types (bit-shift and add) additionally enable finer-grained control of the model's learning capacity, leading to more flexible trade-off between accuracy and (training) efficiency, as well as improved robustness to quantization and pruning. We conduct extensive experiments and ablation studies, all backed up by our FPGA-based ShiftAddNet implementation and energy measurements. Compared to existing DNNs or other multiplication-less models, ShiftAddNet aggressively reduces over 80% hardware-quantified energy cost of DNNs training and inference, while offering comparable or better accuracies. Codes and pre-trained models are available at this https URL.
Problem

Research questions and friction points this paper is trying to address.

Reduces energy costs in DNNs using bit-shift and additive layers
Enables efficient DNN deployment on resource-constrained edge devices
Improves trade-offs between accuracy, efficiency, and robustness
Innovation

Methods, ideas, or system contributions that make the work stand out.

Replaces multiplications with additions and bit-shifts
Enables energy-efficient DNN training and inference
Improves robustness to quantization and pruning
🔎 Similar Papers
No similar papers found.
Haoran You
Haoran You
Georgia Tech
Efficient MLAlg-HW Co-Design
Xiaohan Chen
Xiaohan Chen
University of Cyprus
Transfer learningFault diagnosisTime-series analysis
Yongan Zhang
Yongan Zhang
Georgia Institute of Technology
Deep LearningAI Accelerators
Chaojian Li
Chaojian Li
Hong Kong University of Science and Technology
Efficient AIHardware / software codesign
S
Sicheng Li
Alibaba DAMO Academy
Z
Zihao Liu
Alibaba DAMO Academy
Z
Zhangyang Wang
Department of Electrical and Computer Engineering, The University of Texas at Austin
Y
Yingyan Lin
Department of Electrical and Computer Engineering, Rice University