π€ AI Summary
This work investigates the intrinsic relationship between the double-descent phenomenon and the dynamic evolution of shape/texture bias during CNN training. Using CIFAR-10/100 and Stylized-Image benchmarks, we employ gradient attribution, feature disentanglement analysis, and stage-wise model distillation to establish, for the first time, a quantitative correspondence between double-descent curves and texture-bias migration. We find that texture bias markedly intensifies at the double-descent peak, revealing a coupled mechanism among model capacity, dataset size, and inductive bias. We propose a bias-evolution analytical framework and demonstrate that early stopping effectively mitigates this bias, improving cross-domain generalization robustness by 12.3%. Our results provide novel insights into implicit regularization and generalization behavior in deep learning, linking architectural and data-driven factors to learned representational priors.