🤖 AI Summary
This work addresses the challenge of accurately discriminating relative translation quality in reference-free machine translation evaluation by proposing a novel pairwise comparison-based quality estimation (QE) paradigm. The approach reformulates the QE task as predicting both the direction and magnitude of quality differences between two candidate translations, using human rating deltas as supervision signals. To enhance discriminability and reduce redundancy, a sign regularization term is introduced under candidate order reversal. Employing a lightweight neural architecture and a pairwise training objective, the model outperforms larger single-candidate QE models and reference-based metrics on the WMT24 benchmark. Furthermore, it enables efficient minimum Bayes-risk (MBR) decoding with substantially reduced scoring cost while maintaining near-optimal performance.
📝 Abstract
We present PEAR (Pairwise Evaluation for Automatic Relative Scoring), a supervised Quality Estimation (QE) metric family that reframes reference-free Machine Translation (MT) evaluation as a graded pairwise comparison. Given a source segment and two candidate translations, PEAR predicts the direction and magnitude of their quality difference. The metrics are trained using pairwise supervision derived from differences in human judgments, with an additional regularization term that encourages sign inversion under candidate order reversal. On the WMT24 meta-evaluation benchmark, PEAR outperforms strictly matched single-candidate QE baselines trained with the same data and backbones, isolating the benefit of the proposed pairwise formulation. Despite using substantially fewer parameters than recent large metrics, PEAR surpasses far larger QE models and reference-based metrics. Our analysis further indicates that PEAR yields a less redundant evaluation signal relative to other top metrics. Finally, we show that PEAR is an effective utility function for Minimum Bayes Risk (MBR) decoding, reducing pairwise scoring cost at negligible impact.