🤖 AI Summary
To address the lack of human-like decision-making mechanisms in machine translation (MT), this paper proposes Thinker-DDM, the first framework to incorporate the cognitive science–inspired drift-diffusion model (DDM) into MT. Thinker-DDM explicitly models translators’ dynamic speed–accuracy trade-offs under resource constraints. It integrates large language models (LLMs) with a learnable, adaptive threshold-based decision mechanism, thereby redefining the MT generation paradigm. Experiments on WMT22 and CommonMT demonstrate significant improvements over state-of-the-art baselines across both high- and low-resource translation settings. Further commonsense translation analysis confirms enhanced semantic consistency and robustness. This work pioneers the integration of cognitive modeling with LLM-based MT, establishing a novel paradigm for building more interpretable and human-aligned translation systems.
📝 Abstract
Large language models (LLMs) have demonstrated promising potential in various downstream tasks, including machine translation. However, prior work on LLM-based machine translation has mainly focused on better utilizing training data, demonstrations, or pre-defined and universal knowledge to improve performance, with a lack of consideration of decision-making like human translators. In this paper, we incorporate Thinker with the Drift-Diffusion Model (Thinker-DDM) to address this issue. We then redefine the Drift-Diffusion process to emulate human translators' dynamic decision-making under constrained resources. We conduct extensive experiments under the high-resource, low-resource, and commonsense translation settings using the WMT22 and CommonMT datasets, in which Thinker-DDM outperforms baselines in the first two scenarios. We also perform additional analysis and evaluation on commonsense translation to illustrate the high effectiveness and efficacy of the proposed method.