Domain-Specific Quality Estimation for Machine Translation in Low-Resource Scenarios

📅 2026-03-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of sentence-level quality estimation (QE) for English-to-Indic language translation in low-resource, reference-free scenarios, particularly within domain-specific contexts such as healthcare and legal texts. The authors propose integrating Low-Rank Adaptation (LoRA) into intermediate layers of large language models, coupled with a regression head, and further extend this approach to Low-Rank Multiplicative Adaptation (LoRMA). Their method demonstrates significantly improved QE stability across zero-shot, few-shot, and instruction-prompting settings, with especially strong performance on semantically complex domain-specific content. Experiments conducted under the ALOPE framework validate the effectiveness of the approach on both open-source and closed-source models. The authors also release their code and a new domain-specific QE dataset to support future research.

Technology Category

Application Category

📝 Abstract
Quality Estimation (QE) is essential for assessing machine translation quality in reference-less settings, particularly for domain-specific and low-resource language scenarios. In this paper, we investigate sentence-level QE for English to Indic machine translation across four domains (Healthcare, Legal, Tourism, and General) and five language pairs. We systematically compare zero-shot, few-shot, and guideline-anchored prompting across selected closed-weight and open-weight LLMs. Findings indicate that while closed-weight models achieve strong performance via prompting alone, prompt-only approaches remain fragile for open-weight models, especially in high-risk domains. To address this, we adopt ALOPE, a framework for LLM-based QE that uses Low-Rank Adaptation with regression heads attached to selected intermediate Transformer layers. We also extend ALOPE with recently proposed Low-Rank Multiplicative Adaptation (LoRMA). Our results show that intermediate-layer adaptation consistently improves QE performance, with gains in semantically complex domains, indicating a path toward more robust QE in practical scenarios. We release code and domain-specific QE datasets publicly to support further research.
Problem

Research questions and friction points this paper is trying to address.

Quality Estimation
Machine Translation
Low-Resource
Domain-Specific
Reference-less
Innovation

Methods, ideas, or system contributions that make the work stand out.

Quality Estimation
Low-Rank Adaptation
Domain-Specific MT
LLM Prompting
LoRMA
🔎 Similar Papers
No similar papers found.
N
Namrata Patil Gurav
Institute for People-Centred AI, University of Surrey, United Kingdom
A
Akashdeep Ranu
Institute for People-Centred AI, University of Surrey, United Kingdom
Archchana Sindhujan
Archchana Sindhujan
PhD Student, University of Surrey
Natural Language ProcessingDeep Learning
Diptesh Kanojia
Diptesh Kanojia
Senior Lecturer at University of Surrey | Institute for People-Centred AI
Natural Language ProcessingArtificial Intelligence