DNA: Differentiable Network-Accelerator Co-Search

πŸ“… 2020-10-28
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 16
✨ Influential: 3
πŸ“„ PDF
πŸ€– AI Summary
Co-optimizing deep neural networks (DNNs) and hardware accelerators remains challenging due to the discrete, coupled nature of neural architecture and accelerator microarchitecture/mapping space. Method: This paper proposes DNA, a differentiable neural-accelerator co-search frameworkβ€”the first to enable end-to-end joint optimization of DNN topology and accelerator microarchitecture/mapping strategy. It introduces a unified, FPGA/ASIC-compatible accelerator design space; extends DARTS with hardware-aware gradient propagation; and establishes a PyTorch-based differentiable modeling infrastructure supporting closed-loop optimization driven by FPGA measurements and ASIC synthesis. Contribution/Results: DNA overcomes the differentiability bottleneck inherent in prior discrete co-search approaches. On ImageNet, it achieves a 3.04Γ— throughput improvement and a 5.46% accuracy gain. Search efficiency improves by 1234.3Γ— over state-of-the-art methods, consistently outperforming ten baseline models across accuracy, latency, energy, and hardware feasibility.
πŸ“ Abstract
Powerful yet complex deep neural networks (DNNs) have fueled a booming demand for efficient DNN solutions to bring DNN-powered intelligence into numerous applications. Jointly optimizing the networks and their accelerators are promising in providing optimal performance. However, the great potential of such solutions have yet to be unleashed due to the challenge of simultaneously exploring the vast and entangled, yet different design spaces of the networks and their accelerators. To this end, we propose DNA, a Differentiable Network-Accelerator co-search framework for automatically searching for matched networks and accelerators to maximize both the task accuracy and acceleration efficiency. Specifically, DNA integrates two enablers: (1) a generic design space for DNN accelerators that is applicable to both FPGA- and ASIC-based DNN accelerators and compatible with DNN frameworks such as PyTorch to enable algorithmic exploration for more efficient DNNs and their accelerators; and (2) a joint DNN network and accelerator co-search algorithm that enables simultaneously searching for optimal DNN structures and their accelerators' micro-architectures and mapping methods to maximize both the task accuracy and acceleration efficiency. Experiments and ablation studies based on FPGA measurements and ASIC synthesis show that the matched networks and accelerators generated by DNA consistently outperform state-of-the-art (SOTA) DNNs and DNN accelerators (e.g., 3.04x better FPS with a 5.46% higher accuracy on ImageNet), while requiring notably reduced search time (up to 1234.3x) over SOTA co-exploration methods, when evaluated over ten SOTA baselines on three datasets. All codes will be released upon acceptance.
Problem

Research questions and friction points this paper is trying to address.

Complex DNN Optimization
Accelerator Design
Efficiency Enhancement
Innovation

Methods, ideas, or system contributions that make the work stand out.

DNA System
Co-optimization
Accelerator Design
πŸ”Ž Similar Papers
No similar papers found.
Yongan Zhang
Yongan Zhang
Georgia Institute of Technology
Deep LearningAI Accelerators
Y
Y. Fu
Department of Electrical and Computer Engineering, Rice University
Weiwen Jiang
Weiwen Jiang
George Mason University
Quantum ComputingAI AcceleratorsHW-SW Co-Design
Chaojian Li
Chaojian Li
Hong Kong University of Science and Technology
Efficient AIHardware / software codesign
Haoran You
Haoran You
Georgia Tech
Efficient MLAlg-HW Co-Design
M
Meng Li
Facebook, Inc.
V
V. Chandra
Facebook, Inc.
Y
Yingyan Lin
Department of Electrical and Computer Engineering, Rice University