🤖 AI Summary
This work addresses the lack of systematic evaluation of large language models (LLMs) in competitive programming debugging. To this end, we propose DePro, a test-case-driven debugging approach centered on patching rather than rewriting code, which constitutes the first systematic assessment of LLM-based debugging in the Codeforces setting. DePro integrates brute-force reference generation, stress testing, and iterative LLM-based refinement to efficiently localize and correct errors. Evaluated on 13 real-world incorrect submissions, DePro achieves a 100% repair success rate, reduces the number of debugging attempts by up to 64%, and saves an average of 7.6 minutes per problem, substantially improving debugging efficiency.
📝 Abstract
Debugging consumes a substantial portion of the software development lifecycle, yet the effectiveness of Large Language Models(LLMs) in this task is not well understood. Competitive programming offers a rich benchmark for such evaluation, given its diverse problem domains and strict efficiency requirements. We present an empirical study of LLM-based debugging on competitive programming problems and introduce DePro, a test-case driven approach that assists programmers by correcting existing code rather than generating new solutions. DePro combines brute-force reference generation, stress testing, and iterative LLM-guided refinement to identify and resolve errors efficiently.Experiments on 13 faulty user submissions from Codeforces demonstrate that DePro consistently produces correct solutions, reducing debugging attempts by up to 64% and debugging time by an average of 7.6 minutes per problem compared to human programmers and zero-shot LLM debugging.