Expanding-and-Shrinking Binary Neural Networks

📅 2025-03-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Binary neural networks (BNNs) suffer from limited feature representation capacity due to binary weights and activations, resulting in substantially lower accuracy than full-precision models on complex tasks. To address this, we propose a differentiable, lightweight Expand-Squeeze operation—the first method to break the feature map value bottleneck in BNNs without increasing binary hardware overhead. Our approach enhances binarization via gradient approximation and channel-wise adaptive scaling, thereby improving feature diversity. The operation is architecture-agnostic, seamlessly integrating into both CNNs and Transformers, and compatible with standard BNN training pipelines. Extensive experiments demonstrate state-of-the-art performance across diverse tasks—including image classification, object detection, and diffusion models—achieving absolute accuracy gains of 3.2–5.7 percentage points over prior methods, while incurring less than 0.5% additional computational cost.

Technology Category

Application Category

📝 Abstract
While binary neural networks (BNNs) offer significant benefits in terms of speed, memory and energy, they encounter substantial accuracy degradation in challenging tasks compared to their real-valued counterparts. Due to the binarization of weights and activations, the possible values of each entry in the feature maps generated by BNNs are strongly constrained. To tackle this limitation, we propose the expanding-and-shrinking operation, which enhances binary feature maps with negligible increase of computation complexity, thereby strengthening the representation capacity. Extensive experiments conducted on multiple benchmarks reveal that our approach generalizes well across diverse applications ranging from image classification, object detection to generative diffusion model, while also achieving remarkable improvement over various leading binarization algorithms based on different architectures including both CNNs and Transformers.
Problem

Research questions and friction points this paper is trying to address.

Reduces accuracy loss in binary neural networks
Enhances binary feature maps efficiently
Improves performance across diverse applications
Innovation

Methods, ideas, or system contributions that make the work stand out.

Expanding-and-shrinking operation enhances binary feature maps
Negligible computation complexity increase for improved representation
Generalizes across image classification, detection, diffusion models
🔎 Similar Papers
No similar papers found.
X
Xulong Shi
QCraft
C
Caiyi Sun
Southeast University
Z
Zhi Qi
Southeast University
Liu Hao
Liu Hao
The Hong Kong University of Science and Technology (Guangzhou)
LLM AgentsUrban Foundation ModelSpatio-temporal Data MiningUrban Intelligence
X
Xiaodong Yang
QCraft