π€ AI Summary
This work addresses the critical ethical challenge of detecting AI-generated code. We introduce AIGCodeSet, the first large-scale, multi-model-balanced Python code annotation dataset, comprising 28,280 AI-generated and 47,550 human-written samples. To establish a rigorous cross-model generalization benchmark, we systematically integrate code generated by CodeLlama, Codestral, and Gemini 1.5 Flashβmarking the first such comprehensive evaluation framework. Methodologically, we propose a Bayesian classifier leveraging statistical token-level features of source code, achieving significant improvements in both accuracy and cross-model robustness over conventional machine learning and LLM-based baselines. Our approach is computationally efficient, inherently interpretable, and exhibits strong generalization across diverse generative models. Collectively, this work delivers a reliable, scalable technical foundation for academic integrity assessment and ethical governance in software engineering.
π Abstract
While large language models provide significant convenience for software development, they can lead to ethical issues in job interviews and student assignments. Therefore, determining whether a piece of code is written by a human or generated by an artificial intelligence (AI) model is a critical issue. In this study, we present AIGCodeSet, which consists of 2.828 AI-generated and 4.755 human-written Python codes, created using CodeLlama 34B, Codestral 22B, and Gemini 1.5 Flash. In addition, we share the results of our experiments conducted with baseline detection methods. Our experiments show that a Bayesian classifier outperforms the other models.