E-ConvNeXt: A Lightweight and Efficient ConvNeXt Variant with Cross-Stage Partial Connections

📅 2025-08-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high parameter count and computational complexity of ConvNeXt in resource-constrained scenarios, this paper proposes E-ConvNeXt—a lightweight variant that replaces Layer Scale with cross-stage partial connections (CSP), depthwise separable convolutions, and channel-wise attention, while also redesigning the Stem and Block architectures. These modifications significantly reduce model complexity without compromising representational capacity, achieving an effective balance between feature expressiveness and computational efficiency. Experiments demonstrate that E-ConvNeXt-mini achieves 78.3% Top-1 accuracy on ImageNet with only 0.9 GFLOPs—reducing computation by ~80% over the original ConvNeXt-tiny—while E-ConvNeXt-small attains 81.9% accuracy at 3.1 GFLOPs. Moreover, E-ConvNeXt exhibits strong transferability to downstream tasks such as object detection. This work establishes a scalable architectural paradigm for lightweight, high-performance CNN design.

Technology Category

Application Category

📝 Abstract
Many high-performance networks were not designed with lightweight application scenarios in mind from the outset, which has greatly restricted their scope of application. This paper takes ConvNeXt as the research object and significantly reduces the parameter scale and network complexity of ConvNeXt by integrating the Cross Stage Partial Connections mechanism and a series of optimized designs. The new network is named E-ConvNeXt, which can maintain high accuracy performance under different complexity configurations. The three core innovations of E-ConvNeXt are : (1) integrating the Cross Stage Partial Network (CSPNet) with ConvNeXt and adjusting the network structure, which reduces the model's network complexity by up to 80%; (2) Optimizing the Stem and Block structures to enhance the model's feature expression capability and operational efficiency; (3) Replacing Layer Scale with channel attention. Experimental validation on ImageNet classification demonstrates E-ConvNeXt's superior accuracy-efficiency balance: E-ConvNeXt-mini reaches 78.3% Top-1 accuracy at 0.9GFLOPs. E-ConvNeXt-small reaches 81.9% Top-1 accuracy at 3.1GFLOPs. Transfer learning tests on object detection tasks further confirm its generalization capability.
Problem

Research questions and friction points this paper is trying to address.

Reducing ConvNeXt's parameter scale and complexity
Maintaining high accuracy in lightweight configurations
Enhancing feature expression and operational efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrating Cross Stage Partial Connections mechanism
Optimizing Stem and Block structures for efficiency
Replacing Layer Scale with channel attention
🔎 Similar Papers
No similar papers found.
Fang Wang
Fang Wang
Postdoc, Stanford University
Reading acquisitiondyslexiacross-linguistic researchbilingualismcognitive neuroscience
Huitao Li
Huitao Li
Duke-Nus Medical School
Medical Informatics
W
Wenhan Chao
School of Computer Science and Engineering, Beihang University, Beijing, 100191, People’s Republic of China
Z
Zheng Zhuo
College of Information Engineering, Beijing Institute of Petrochemical Technology, Beijing, 102617, People’s Republic of China
Y
Yiran Ji
College of Information Engineering, Beijing Institute of Petrochemical Technology, Beijing, 102617, People’s Republic of China
C
Chang Peng
College of Information Engineering, Beijing Institute of Petrochemical Technology, Beijing, 102617, People’s Republic of China
Y
Yupeng Sun
College of Information Engineering, Beijing Institute of Petrochemical Technology, Beijing, 102617, People’s Republic of China