The illusion of a perfect metric: Why evaluating AI's words is harder than it looks

📅 2025-08-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current automatic evaluation metrics (AEMs) for natural language generation (NLG) suffer from low correlation with human judgments and poor task generality—especially in LLM-as-a-Judge and RAG settings—due to systemic limitations in task adaptability, relevance stability, and validation rigor. Through a systematic literature review and methodological analysis, we comparatively evaluate three dominant AEM categories: lexical matching, semantic similarity models, and LLM-based judges, exposing pervasive issues including high cross-task correlation variance and the absence of standardized validation protocols. Our key contributions are threefold: (1) we establish the inevitability of “no silver bullet” metrics for NLG evaluation; (2) we propose a task-driven, complementary evaluation paradigm that integrates multiple AEMs according to specific functional requirements; and (3) we advocate for a structured, reproducible, multi-dimensional human-alignment validation framework. This work provides a principled methodological foundation for advancing the scientific rigor of NLG evaluation.

Technology Category

Application Category

📝 Abstract
Evaluating Natural Language Generation (NLG) is crucial for the practical adoption of AI, but has been a longstanding research challenge. While human evaluation is considered the de-facto standard, it is expensive and lacks scalability. Practical applications have driven the development of various automatic evaluation metrics (AEM), designed to compare the model output with human-written references, generating a score which approximates human judgment. Over time, AEMs have evolved from simple lexical comparisons, to semantic similarity models and, more recently, to LLM-based evaluators. However, it seems that no single metric has emerged as a definitive solution, resulting in studies using different ones without fully considering the implications. This paper aims to show this by conducting a thorough examination of the methodologies of existing metrics, their documented strengths and limitations, validation methods, and correlations with human judgment. We identify several key challenges: metrics often capture only specific aspects of text quality, their effectiveness varies by task and dataset, validation practices remain unstructured, and correlations with human judgment are inconsistent. Importantly, we find that these challenges persist in the most recent type of metric, LLM-as-a-Judge, as well as in the evaluation of Retrieval Augmented Generation (RAG), an increasingly relevant task in academia and industry. Our findings challenge the quest for the 'perfect metric'. We propose selecting metrics based on task-specific needs and leveraging complementary evaluations and advocate that new metrics should focus on enhanced validation methodologies.
Problem

Research questions and friction points this paper is trying to address.

Evaluating AI-generated text quality remains a challenging research problem
Automatic evaluation metrics fail to consistently correlate with human judgment
Existing metrics capture only specific aspects of text quality inadequately
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-based evaluators for human judgment approximation
Task-specific metric selection for varied effectiveness
Enhanced validation methodologies for new metrics
🔎 Similar Papers
No similar papers found.
M
Maria Paz Oliva
Iris AI, Borgen, Norway
Adriana Correia
Adriana Correia
Research Engineer, Iris.ai
I
Ivan Vankov
Iris AI, Neurobiology BAS, Sofia, Bulgaria
V
Viktor Botev
Iris AI, Sofia, Bulgaria