🤖 AI Summary
This study challenges the prevailing hypothesis that improving text readability enhances the coherent generation capability of small language models (SLMs). Method: We construct a synthetically generated dataset with controlled readability and consistent structural properties, and conduct systematic ablation experiments using controlled-variable training to investigate the relationship between linguistic complexity and model learnability. Contribution/Results: (1) Statistical concision—measured by n-gram diversity—proves more predictive of SLM learning efficiency than conventional readability metrics; (2) SLMs trained on complex, adult-targeted corpora achieve coherence development rates and final performance comparable to—or even surpassing—those trained on simplified corpora. This work provides the first empirical challenge to the paradigm that directly analogizes child language acquisition mechanisms to SLM training, demonstrating that lexical and syntactic simplification is not a necessary condition for efficient learning in small models.
📝 Abstract
Recent studies suggest that very small language models (SLMs) can generate surprisingly coherent text when trained on simplified, child-directed corpora such as TinyStories. These findings have been interpreted as evidence that readability -- characterized by accessible vocabulary, familiar narrative structure, and simple syntax -- plays a key role in enabling such capabilities to emerge. In this paper, we challenge that interpretation. We construct synthetic datasets with matched structure but varied readability, and find that readability alone does not predict coherence or learning efficiency in SLMs. Models trained on complex, adult-level text perform comparably to those trained on simplified language, and even exhibit faster development of coherence during training. Instead, we show that statistical simplicity, as measured by n-gram diversity, is a stronger predictor of learnability. Our findings caution against the growing trend of anthropomorphizing language model training -- drawing parallels to human cognitive development without empirical basis -- and argue for more precise reasoning about what properties actually support capability emergence in small models.