π€ AI Summary
Existing code large language models employ character-level random masking in Fill-in-the-Middle (FIM) pretraining, ignoring syntactic structure and thus causing semantic fragmentation and misalignment with realistic code editing patterns. This work proposes an Abstract Syntax Tree (AST)-driven, structure-aware FIM pretraining paradigm: leveraging ASTs to identify complete syntactic units (e.g., functions, expressions) and enforce syntax-consistent context completion. Key contributions include: (1) the first AST-guided FIM pretraining method; (2) Real-FIM-Evalβthe first multilingual benchmark derived from real GitHub commits for evaluating practical code editing; and (3) substantial improvements in real-world code editing capability on 1B- and 8B-parameter models, with up to +5.0 points on standard FIM benchmarks. The implementation is open-sourced and supports 12 programming languages.
π Abstract
Fill-in-the-Middle (FIM) is a common pretraining method for code LLMs, where models complete code segments given surrounding context. However, existing LLMs treat code as plain text and mask random character spans. We propose and evaluate AST-FIM, a pretraining strategy that leverages Abstract Syntax Trees (ASTs) to mask complete syntactic structures at scale, ensuring coherent training examples better aligned with universal code structures and common code editing patterns such as blocks, expressions, or functions. To evaluate real-world fill-in-the-middle (FIM) programming tasks, we introduce Real-FIM-Eval, a benchmark derived from 30,000+ GitHub commits across 12 languages. On infilling tasks, experiments on 1B and 8B parameter models show that AST-FIM is particularly beneficial for real-world code editing as it outperforms standard random-character FIM by up to 5 pts on standard FIM benchmarks. Our code is publicly available at https://github.com/gonglinyuan/ast_fim.