Rethinking Human-like Translation Strategy: Integrating Drift-Diffusion Model with Large Language Models for Machine Translation

📅 2024-02-16
🏛️ arXiv.org
📈 Citations: 5
Influential: 0
📄 PDF
🤖 AI Summary
To address the lack of human-like decision-making mechanisms in machine translation (MT), this paper proposes Thinker-DDM, the first framework to incorporate the cognitive science–inspired drift-diffusion model (DDM) into MT. Thinker-DDM explicitly models translators’ dynamic speed–accuracy trade-offs under resource constraints. It integrates large language models (LLMs) with a learnable, adaptive threshold-based decision mechanism, thereby redefining the MT generation paradigm. Experiments on WMT22 and CommonMT demonstrate significant improvements over state-of-the-art baselines across both high- and low-resource translation settings. Further commonsense translation analysis confirms enhanced semantic consistency and robustness. This work pioneers the integration of cognitive modeling with LLM-based MT, establishing a novel paradigm for building more interpretable and human-aligned translation systems.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated promising potential in various downstream tasks, including machine translation. However, prior work on LLM-based machine translation has mainly focused on better utilizing training data, demonstrations, or pre-defined and universal knowledge to improve performance, with a lack of consideration of decision-making like human translators. In this paper, we incorporate Thinker with the Drift-Diffusion Model (Thinker-DDM) to address this issue. We then redefine the Drift-Diffusion process to emulate human translators' dynamic decision-making under constrained resources. We conduct extensive experiments under the high-resource, low-resource, and commonsense translation settings using the WMT22 and CommonMT datasets, in which Thinker-DDM outperforms baselines in the first two scenarios. We also perform additional analysis and evaluation on commonsense translation to illustrate the high effectiveness and efficacy of the proposed method.
Problem

Research questions and friction points this paper is trying to address.

Modeling human-like deliberation in machine translation
Improving translation under resource constraints
Enhancing commonsense translation performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Incorporates Thinker with Drift-Diffusion Model
Redefines Drift-Diffusion for dynamic decision-making
Outperforms baselines in resource-constrained scenarios
🔎 Similar Papers
No similar papers found.
Hongbin Na
Hongbin Na
Australian AI Institute, University of Technology Sydney / Shanghai AI Laboratory
Computational Social ScienceNarrative UnderstandingAI for Healthcare
Zimu Wang
Zimu Wang
Tsinghua University
recommendation
M
M. Maimaiti
Chinese Academy of Sciences
T
Tong Chen
Xi’an Jiaotong-Liverpool University
W
Wei Wang
Xi’an Jiaotong-Liverpool University
T
Tao Shen
AAII, FEIT, University of Technology Sydney
L
Ling Chen
AAII, FEIT, University of Technology Sydney