A Systematic Literature Review on Neural Code Translation

📅 2025-05-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The neural code translation field lacks a systematic, up-to-date survey. Method: We conduct a systematic literature review (SLR) of 57 core studies published between 2020 and 2025, introducing the first seven-dimensional technical framework—covering task formulation, data preprocessing, code modeling, architecture design, training strategies, evaluation protocols, and application scenarios—and integrating thematic coding, trend clustering, and cross-study comparison for qualitative and semi-quantitative analysis. Results: We identify critical bottlenecks including weak model generalizability and difficulty in cross-lingual semantic alignment. We confirm the dominance of Transformer-based models, the rising adoption of AST-enhanced modeling, and broad consensus on BLEU’s inadequacy for code translation evaluation. Furthermore, we propose reorienting evaluation toward industrial deployment contexts and publicly release the field’s first comprehensive technology landscape map and a reusable analytical template.

Technology Category

Application Category

📝 Abstract
Code translation aims to convert code from one programming language to another automatically. It is motivated by the need for multi-language software development and legacy system migration. In recent years, neural code translation has gained significant attention, driven by rapid advancements in deep learning and large language models. Researchers have proposed various techniques to improve neural code translation quality. However, to the best of our knowledge, no comprehensive systematic literature review has been conducted to summarize the key techniques and challenges in this field. To fill this research gap, we collected 57 primary studies covering the period 2020~2025 on neural code translation. These studies are analyzed from seven key perspectives: task characteristics, data preprocessing, code modeling, model construction, post-processing, evaluation subjects, and evaluation metrics. Our analysis reveals current research trends, identifies unresolved challenges, and shows potential directions for future work. These findings can provide valuable insights for both researchers and practitioners in the field of neural code translation.
Problem

Research questions and friction points this paper is trying to address.

Summarize key techniques and challenges in neural code translation
Analyze 57 studies on neural code translation from 2020-2025
Identify research trends and future directions for code translation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Systematic review of 57 neural code translation studies
Analysis from seven key technical perspectives
Identifies trends and challenges in code translation
🔎 Similar Papers
No similar papers found.
X
Xiang Chen
School of Artificial Intelligence and Computer Science, Nantong University, China and State Key Lab. for Novel Software Technology, Nanjing University, China
J
Jiacheng Xue
School of Artificial Intelligence and Computer Science, Nantong University, China
Xiaofei Xie
Xiaofei Xie
Singapore Management University
Software EngineeringLoop AnalysisTestingDeep Learning
C
Caokai Liang
School of Artificial Intelligence and Computer Science, Nantong University, China
Xiaolin Ju
Xiaolin Ju
Associate Professor of Nantong University
Software EngineeringSoftware analysis and testingProgram debugging