π€ AI Summary
This work addresses the challenge of incomparable evaluation results in extremely low-resource machine translation, where performance gains are often confounded by dataset-specific biases rather than genuine model improvements. To this end, the authors propose FREDβa novel difficulty metric framework that quantifies intrinsic dataset complexity along four dimensions: token fertility ratio, retrieval proxy, pre-training exposure, and corpus diversity. The framework reveals that train-test overlap and pre-training coverage are primary drivers of performance disparities, while also identifying high fertility as a fundamental bottleneck for translating non-Latin and extinct languages. By providing a transparent and standardized benchmark, FRED substantially enhances the interpretability and comparability of evaluations in extremely low-resource MT research.
π Abstract
The landscape of extremely low-resource machine translation (MT) is characterized by perplexing variability in reported performance, often making results across different language pairs difficult to contextualize. For researchers focused on specific language groups -- such as ancient languages -- it is nearly impossible to determine if breakthroughs reported in other contexts (e.g., native African or American languages) result from superior methodologies or are merely artifacts of benchmark collection. To address this problem, we introduce the FRED Difficulty Metrics, which include the Fertility Ratio (F), Retrieval Proxy (R), Pre-training Exposure (E), and Corpus Diversity (D) and serve as dataset-intrinsic metrics to contextualize reported scores. These metrics reveal that a significant portion of result variability is explained by train-test overlap and pre-training exposure rather than model capability. Additionally, we identify that some languages -- particularly extinct and non-Latin indigenous languages -- suffer from poor tokenization coverage (high token fertility), highlighting a fundamental limitation of transferring models from high-resource languages that lack a shared vocabulary. By providing these indices alongside performance scores, we enable more transparent evaluation of cross-lingual transfer and provide a more reliable foundation for the XLR MT community.