AppForge: From Assistant to Independent Developer - Are GPTs Ready for Software Development?

📅 2025-10-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the capability of large language models (LLMs) to generate complete, functional Android applications from scratch, with particular emphasis on system-level challenges—including state coordination, activity lifecycle management, and asynchronous operations. To enable rigorous evaluation, we introduce AppForge, the first automated, end-to-end benchmark for Android application generation, comprising 101 development tasks derived from real-world applications. We propose a novel multi-agent framework that autonomously generates functional specifications and executable test cases, all validated by domain experts to ensure fidelity and reproducibility. Extensive evaluation across 12 state-of-the-art LLMs reveals severe limitations: even the strongest model, GPT-5, achieves only 18.8% task success rate. These results expose fundamental shortcomings of current LLMs in handling complex, structured software engineering tasks requiring deep platform-specific knowledge and precise compositional reasoning.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated remarkable capability in function-level code generation tasks. Unlike isolated functions, real-world applications demand reasoning over the entire software system: developers must orchestrate how different components interact, maintain consistency across states over time, and ensure the application behaves correctly within the lifecycle and framework constraints. Yet, no existing benchmark adequately evaluates whether LLMs can bridge this gap and construct entire software systems from scratch. To address this gap, we propose APPFORGE, a benchmark consisting of 101 software development problems drawn from real-world Android apps. Given a natural language specification detailing the app functionality, a language model is tasked with implementing the functionality into an Android app from scratch. Developing an Android app from scratch requires understanding and coordinating app states, lifecycle management, and asynchronous operations, calling for LLMs to generate context-aware, robust, and maintainable code. To construct APPFORGE, we design a multi-agent system to automatically summarize the main functionalities from app documents and navigate the app to synthesize test cases validating the functional correctness of app implementation. Following rigorous manual verification by Android development experts, APPFORGE incorporates the test cases within an automated evaluation framework that enables reproducible assessment without human intervention, making it easily adoptable for future research. Our evaluation on 12 flagship LLMs show that all evaluated models achieve low effectiveness, with the best-performing model (GPT-5) developing only 18.8% functionally correct applications, highlighting fundamental limitations in current models' ability to handle complex, multi-component software engineering challenges.
Problem

Research questions and friction points this paper is trying to address.

Evaluating LLMs' ability to build complete software systems from scratch
Addressing the gap in benchmarks for full application development
Testing model performance on real-world Android app creation challenges
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposed APPFORGE benchmark for full software development
Used multi-agent system to synthesize test cases
Created automated evaluation framework without human intervention
🔎 Similar Papers
No similar papers found.
Dezhi Ran
Dezhi Ran
School of Computer Science, Peking University
Short Video StreamingSoftware TestingProgram Analysis
Y
Yuan Cao
Key Lab of HCST (PKU), MOE; School of Computer Science, Peking University, China
Mengzhou Wu
Mengzhou Wu
Peking University
Software EngineeringLarge Language Model
Simin Chen
Simin Chen
Columbia University
Software EngineeringMachine Learning
Y
Yuzhe Guo
Beijing Jiaotong University, China
J
Jun Ren
University of Texas at Dallas, USA
Zihe Song
Zihe Song
University of Texas at Dallas
Computer ScienceSoftware TestingImitation Learning
H
Hao Yu
Hong Kong University of Science and Technology, Hong Kong SAR, China
J
Jialei Wei
Fudan University Institute of Systems for Advanced Computing, Shanghai, China
L
Linyi Li
Simon Fraser University, Canada
W
Wei Yang
University of Texas at Dallas, USA
Baishakhi Ray
Baishakhi Ray
Associate Professor, Columbia University
Software EngineeringMachine LearningAI4CodeAI4SESE4AI
T
Tao Xie
Fudan University Institute of Systems for Advanced Computing, Shanghai, China