🤖 AI Summary
Low-resource language benchmarks—exemplified by Turkish—frequently suffer from linguistic inaccuracy and cultural misalignment, undermining NLP evaluation validity. Method: We propose the first six-dimensional dataset quality framework for low-resource languages, assessing linguistic correctness, cultural appropriateness, terminological accuracy, among others, via expert human annotation and multi-model collaborative evaluation (GPT-4o, Llama3.3-70B). Contribution/Results: Systematic evaluation of 17 mainstream Turkish benchmarks reveals that 70% fail to meet baseline quality thresholds, and 85% of assessed dimensions exhibit significant deficiencies. LLMs substantially underperform humans on cultural commonsense reasoning, yet demonstrate complementary strengths: GPT-4o excels at syntactic and terminological judgment, while Llama3.3-70B outperforms on cultural knowledge inference. This work establishes a reproducible methodology and empirically grounded benchmark for rigorous low-resource language NLP evaluation.
📝 Abstract
The reliance on translated or adapted datasets from English or multilingual resources introduces challenges regarding linguistic and cultural suitability. This study addresses the need for robust and culturally appropriate benchmarks by evaluating the quality of 17 commonly used Turkish benchmark datasets. Using a comprehensive framework that assesses six criteria, both human and LLM-judge annotators provide detailed evaluations to identify dataset strengths and shortcomings. Our results reveal that 70% of the benchmark datasets fail to meet our heuristic quality standards. The correctness of the usage of technical terms is the strongest criterion, but 85% of the criteria are not satisfied in the examined datasets. Although LLM judges demonstrate potential, they are less effective than human annotators, particularly in understanding cultural common sense knowledge and interpreting fluent, unambiguous text. GPT-4o has stronger labeling capabilities for grammatical and technical tasks, while Llama3.3-70B excels at correctness and cultural knowledge evaluation. Our findings emphasize the urgent need for more rigorous quality control in creating and adapting datasets for low-resource languages.