🤖 AI Summary
Current evaluations of diffusion language models rely excessively on generation perplexity, often leading to invalid comparisons. This work identifies the limitations of this metric and, for the first time, formally decomposes generation perplexity and entropy as two components of the Kullback–Leibler (KL) divergence. Building on this insight, the paper introduces “generation frontier,” a novel information-theoretic evaluation framework. Experimental results on benchmarks such as OpenWebText demonstrate that this approach more accurately and reliably reflects generation quality in small-scale diffusion language models—comparable to GPT-2 small—and substantially improves comparability across models.
📝 Abstract
Diffusion language models have seen exciting recent progress, offering far more flexibility in generative trajectories than autoregressive models. This flexibility has motivated a growing body of research into new approaches to diffusion language modeling, which typically begins at the scale of GPT-2 small (150 million parameters). However, these advances introduce new issues with evaluation methodology. In this technical note, we discuss the limitations of current methodology and propose principled augmentations to ensure reliable comparisons. We first discuss why OpenWebText has become the standard benchmark, and why alternatives such as LM1B are inherently less meaningful. We then discuss the limitations of likelihood evaluations for diffusion models, and explain why relying on generative perplexity alone as a metric can lead to uninformative results. To address this, we show that generative perplexity and entropy are two components of the KL divergence to a reference distribution. This decomposition explains generative perplexity's sensitivity to entropy, and naturally suggests generative frontiers as a principled method for evaluating model generative quality. We conclude with empirical observations on model quality at this scale. We include a blog post with interactive content to illustrate the argument at https://patrickpynadath1.github.io/blog/eval_methodology/.