🤖 AI Summary
This work challenges the assumption that classifier-guided quality filtering (CQF) genuinely captures intrinsic data quality. We observe that while CQF improves downstream task performance, it fails to enhance language modeling capability on high-quality data—and even implicitly discards high-quality samples—revealing a “data quality illusion.” Methodologically, we propose a binary-classifier-based quality scoring framework and construct a synthetic dataset with progressively degraded quality via random token shuffling, enabling systematic analysis of training dynamics relative to ground-truth quality. Our key contribution is the first empirical demonstration that CQF’s efficacy stems from biased distributional shifts—not genuine quality awareness. Through controlled synthetic-data ablations and counterfactual experiments, we critically undermine CQF’s validity as a proxy for data quality, establishing new benchmarks and prompting fundamental reconsideration of quality evaluation in data curation.
📝 Abstract
Large-scale models are pretrained on massive web-crawled datasets containing documents of mixed quality, making data filtering essential. A popular method is Classifier-based Quality Filtering (CQF), which trains a binary classifier to distinguish between pretraining data and a small, high-quality set. It assigns each pretraining document a quality score defined as the classifier's score and retains only the top-scoring ones. We provide an in-depth analysis of CQF. We show that while CQF improves downstream task performance, it does not necessarily enhance language modeling on the high-quality dataset. We explain this paradox by the fact that CQF implicitly filters the high-quality dataset as well. We further compare the behavior of models trained with CQF to those trained on synthetic data of increasing quality, obtained via random token permutations, and find starkly different trends. Our results challenge the view that CQF captures a meaningful notion of data quality.