ALOPE: Adaptive Layer Optimization for Translation Quality Estimation using Large Language Models

📅 2025-08-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address performance bottlenecks in Large Language Model (LLM)-based Quality Estimation (QE) for machine translation—stemming from pretraining objective mismatch and imbalanced cross-lingual data distributions—this paper proposes a layer-level adaptive optimization framework. Built upon Low-Rank Adaptation (LoRA), the method dynamically selects and weights intermediate Transformer layer representations, integrating them via multi-head regression loss aggregation to strengthen cross-lingual alignment and reference-free quality prediction. Empirical results demonstrate that intermediate-layer representations are inherently more suitable for QE tasks. The proposed approach consistently outperforms state-of-the-art QE models across diverse LLM backbones, with particularly substantial gains on low-resource language pairs. To foster reproducibility and further research, the source code and trained models are publicly released.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have shown remarkable performance across a wide range of natural language processing tasks. Quality Estimation (QE) for Machine Translation (MT), which assesses the quality of a source-target pair without relying on reference translations, remains a challenging cross-lingual task for LLMs. The challenges stem from the inherent limitations of existing LLM-based QE systems, which are pre-trained for causal language modelling rather than regression-specific tasks, further elevated by the presence of low-resource languages given pre-training data distribution. This paper introduces ALOPE, an adaptive layer-optimization framework designed to enhance LLM-based QE by restructuring Transformer representations through layer-wise adaptation for improved regression-based prediction. Our framework integrates low-rank adapters (LoRA) with regression task heads, leveraging selected pre-trained Transformer layers for improved cross-lingual alignment. In addition to the layer-specific adaptation, ALOPE introduces two strategies-dynamic weighting, which adaptively combines representations from multiple layers, and multi-head regression, which aggregates regression losses from multiple heads for QE. Our framework shows improvements over various existing LLM-based QE approaches. Empirical evidence suggests that intermediate Transformer layers in LLMs provide contextual representations that are more aligned with the cross-lingual nature of the QE task. We make resultant models and framework code publicly available for further research, also allowing existing LLM-based MT frameworks to be scaled with QE capabilities.
Problem

Research questions and friction points this paper is trying to address.

Enhance LLM-based Quality Estimation for Machine Translation
Optimize Transformer layers for cross-lingual regression tasks
Improve low-resource language handling in QE systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive layer-optimization for Transformer representations
Low-rank adapters with regression task heads
Dynamic weighting and multi-head regression strategies
🔎 Similar Papers
No similar papers found.
Archchana Sindhujan
Archchana Sindhujan
PhD Student, University of Surrey
Natural Language ProcessingDeep Learning
Shenbin Qian
Shenbin Qian
Postdoctoral researcher at University of Oslo
Natural Language Processing
C
Chan Chi Chun Matthew
Institute for People-Centred AI and Centre for Translation Studies, School of Computer Science and Electronic Engineering, University of Surrey, United Kingdom
C
Constantin Orasan
Institute for People-Centred AI and Centre for Translation Studies, School of Computer Science and Electronic Engineering, University of Surrey, United Kingdom
Diptesh Kanojia
Diptesh Kanojia
Senior Lecturer at University of Surrey | Institute for People-Centred AI
Natural Language ProcessingArtificial Intelligence