π€ AI Summary
Existing hyper-connections address the intrinsic trade-off between gradient vanishing and representation collapse in deep neural networks but incur prohibitive GPU memory overhead by expanding hidden-state dimensionality. This work proposes Frac-Connections: a lightweight multi-depth connectivity architecture that, for the first time, realizes the multi-scale strength principle of hyper-connections via *fractional blocking*βdynamically partitioning fixed-width hidden states into blocks and assigning each block a learnable, depth-dependent connection weight. Crucially, this design modulates gradient flow and information pathways across depths without increasing parameter count or memory footprint. Evaluated on a 7B MoE model trained on 3T tokens, Frac-Connections significantly outperforms standard residual connections, accelerating convergence and improving final performance on both language modeling and downstream tasks.
π Abstract
Residual connections are central to modern deep learning architectures, enabling the training of very deep networks by mitigating gradient vanishing. Hyper-Connections recently generalized residual connections by introducing multiple connection strengths at different depths, thereby addressing the seesaw effect between gradient vanishing and representation collapse. However, Hyper-Connections increase memory access costs by expanding the width of hidden states. In this paper, we propose Frac-Connections, a novel approach that divides hidden states into multiple parts rather than expanding their width. Frac-Connections retain partial benefits of Hyper-Connections while reducing memory consumption. To validate their effectiveness, we conduct large-scale experiments on language tasks, with the largest being a 7B MoE model trained on up to 3T tokens, demonstrating that Frac-Connections significantly outperform residual connections.