Effect of Convolutional Depth on Image Recognition Performance: VGG vs. ResNet vs. GoogLeNet

๐Ÿ“… 2026-02-09
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This study investigates the genuine impact of network depth on image recognition performance by introducing the concept of โ€œeffective depthโ€ to distinguish between nominal layer count and the actual depth through which information propagates. Under a unified training protocol, controlled comparative experiments are conducted across VGG, ResNet, and GoogLeNet architectures to quantitatively assess how depth influences accuracy, convergence stability, and computational efficiency. The findings reveal that residual connections and Inception modules substantially enhance depth utilization efficiency, achieving superior performance with lower effective depth, whereas conventional deep networks often suffer from accuracy saturation and optimization difficulties. This work underscores the critical role of architectural design in modulating the benefits of depth, moving beyond the simplistic paradigm of merely stacking more layers.

Technology Category

Application Category

๐Ÿ“ Abstract
Increasing convolutional depth has been central to advances in image recognition, yet deeper networks do not uniformly yield higher accuracy, stable optimization, or efficient computation. We present a controlled comparative study of three canonical convolutional neural network architectures - VGG, ResNet, and GoogLeNet - to isolate how depth influences classification performance, convergence behavior, and computational efficiency. By standardizing training protocols and explicitly distinguishing between nominal and effective depth, we show that the benefits of depth depend critically on architectural mechanisms that constrain its effective manifestation during training rather than on nominal depth alone. Although plain deep networks exhibit early accuracy saturation and optimization instability, residual and inception-based architectures consistently translate additional depth into improved accuracy at lower effective depth and favorable accuracy-compute trade-offs. These findings demonstrate that effective depth, not nominal depth, is the operative quantity governing depth's role as a productive scaling dimension in convolutional networks.
Problem

Research questions and friction points this paper is trying to address.

convolutional depth
image recognition
network architecture
effective depth
nominal depth
Innovation

Methods, ideas, or system contributions that make the work stand out.

effective depth
convolutional neural networks
architectural mechanisms
image recognition
depth scaling
๐Ÿ”Ž Similar Papers
2024-08-29Medical Imaging 2025: Digital and Computational PathologyCitations: 1