Stability as a Liability:Systematic Breakdown of Linguistic Structure in LLMs

πŸ“… 2026-01-26
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work uncovers an intrinsic tension between training stability and generation diversity in large language models: overly stable training dynamics implicitly minimize the forward KL divergence, leading to reduced output entropy and degradation of linguistic structure. To address this, we propose a feedback-driven controlled training framework that integrates maximum likelihood objectives with real-time generative statistical analysis. Through systematic experiments across diverse architectures and random seeds, we demonstrate that stable training often yields low-entropy, repetitive outputs. Our findings challenge the prevailing assumption that training stability is a sufficient proxy for generation quality, offering a novel perspective on how optimization dynamics shape a model’s expressive capacity.

Technology Category

Application Category

πŸ“ Abstract
Training stability is typically regarded as a prerequisite for reliable optimization in large language models. In this work, we analyze how stabilizing training dynamics affects the induced generation distribution. We show that under standard maximum likelihood training, stable parameter trajectories lead stationary solutions to approximately minimize the forward KL divergence to the empirical distribution, while implicitly reducing generative entropy. As a consequence, the learned model can concentrate probability mass on a limited subset of empirical modes, exhibiting systematic degeneration despite smooth loss convergence. We empirically validate this effect using a controlled feedback-based training framework that stabilizes internal generation statistics, observing consistent low-entropy outputs and repetitive behavior across architectures and random seeds. It indicates that optimization stability and generative expressivity are not inherently aligned, and that stability alone is an insufficient indicator of generative quality.
Problem

Research questions and friction points this paper is trying to address.

stability
generative degeneration
entropy
large language models
training dynamics
Innovation

Methods, ideas, or system contributions that make the work stand out.

training stability
generative entropy
forward KL divergence
degenerate generation
feedback-based training
πŸ”Ž Similar Papers
No similar papers found.
X
Xianzhe Meng
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China
Q
Qiangsheng Zeng
School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
Ling Luo
Ling Luo
Professor, School of Computer Science and Technology, Dalian University of Technology
Biomedical Text MiningBioNLPNatural Language ProcessingMachine Learning
Q
Qinghan Yang
School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China
J
Jiarui Hao
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China
W
Wenbo Wu
School of Electrical and Electronic Engineering, Huazhong University of Science and Technology, Wuhan, China
Q
Qinyu Wang
School of Mathematics and Statistics, Huazhong University of Science and Technology, Wuhan, China
Rui Yin
Rui Yin
Assistant Professor, University of Florida
Machine learningBioinformaticsProtein language modelRNA virus and cancerAD and Rare Disease
Lin Qi
Lin Qi
Ocean University of China
Computer VisionAI for Oceanography
R
Renzhi Lu
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China