Mitigating Communication Costs in Neural Networks: The Role of Dendritic Nonlinearity

📅 2023-06-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
High inter-layer communication overhead severely limits the energy efficiency of artificial neural networks. Method: Inspired by nonlinear dendritic computation in biological neurons, we propose a differentiable dendritic computing unit that embeds local feature aggregation within individual neurons. Our approach integrates computational neuroscience modeling with end-to-end machine learning training to construct a dendrite-enhanced network architecture supporting gradient backpropagation. Contribution/Results: While dendritic nonlinearity alone does not enhance baseline learning capability, performing nonlinear aggregation of multi-branch inputs *within* each neuron significantly expands representational capacity and reduces inter-layer data transmission by up to 62% (on ImageNet using ResNet-18), without compromising inference accuracy. This work establishes a “communication-minimization” paradigm for neuromorphic accelerator design targeting low-power edge AI.
📝 Abstract
Our understanding of biological neuronal networks has profoundly influenced the development of artificial neural networks (ANNs). However, neurons utilized in ANNs differ considerably from their biological counterparts, primarily due to the absence of complex dendritic trees with local nonlinearities. Early studies have suggested that dendritic nonlinearities could substantially improve the learning capabilities of neural network models. In this study, we systematically examined the role of nonlinear dendrites within neural networks. Utilizing machine-learning methodologies, we assessed how dendritic nonlinearities influence neural network performance. Our findings demonstrate that dendritic nonlinearities do not substantially affect learning capacity; rather, their primary benefit lies in enabling network capacity expansion while minimizing communication costs through effective localized feature aggregation. This research provides critical insights with significant implications for designing future neural network accelerators aimed at reducing communication overhead during neural network training and inference.
Problem

Research questions and friction points this paper is trying to address.

Investigates dendritic nonlinearities' impact on neural networks
Explores communication cost reduction in neural networks
Assesses localized feature aggregation for network efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Incorporates dendritic nonlinearities in neural networks
Minimizes communication costs via localized aggregation
Evaluates impact using machine-learning methodologies
X
Xundong Wu
Zhejiang Lab, China
Pengfei Zhao
Pengfei Zhao
ATB Potsdam
LLMCompressionXAIMechanistic Interpretability
Z
Zilin Yu
Bytedance, China
L
Lei Ma
Beijing Academy of Artificial Intelligence, China; National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, China; National Biomedical Imaging Center, Peking University, China
K
Ka-Wa Yip
Zhejiang Lab, China
Huajin Tang
Huajin Tang
Zhejiang University, China
Brain-inspired AIneuroroboticsspiking neural networksbrain-inspired computing
Gang Pan
Gang Pan
Tianjin University
Computer visionMultimodalAI
P
Poirazi Panayiota
Tiejun Huang
Tiejun Huang
Professor,School of Computer Science, Peking University
Visual Information Processing