🤖 AI Summary
Despite the rapid proliferation of AI-powered programming assistants (e.g., GitHub Copilot), empirical understanding of developer needs and pain points in real-world development remains limited. Method: We conducted the first large-scale, manually annotated, fine-grained theme–sentiment analysis of 12,000 user reviews from 32 mainstream AI extensions on the VS Code Marketplace—selected from 1,085 tools (90% released within the past two years). Integrating qualitative coding with marketplace statistics, we constructed a taxonomy capturing functional satisfaction, performance bottlenecks, and usage expectations. Contribution/Results: Our analysis identifies four critical dimensions prioritized by developers: intelligence, contextual awareness, customizability, and resource efficiency. We derive five evidence-based practical implications, revealing significant gaps in current tools’ ability to comprehend complex logic, adapt to personalized workflows, and operate with low computational overhead—thereby providing both empirical grounding and a theoretical framework for designing next-generation AI programming assistants.
📝 Abstract
This paper aims to explore fundamental questions in the era when AI coding assistants like GitHub Copilot are widely adopted: what do developers truly value and criticize in AI coding assistants, and what does this reveal about their needs and expectations in real-world software development? Unlike previous studies that conduct observational research in controlled and simulated environments, we analyze extensive, first-hand user reviews of AI coding assistants, which capture developers' authentic perspectives and experiences drawn directly from their actual day-to-day work contexts. We identify 1,085 AI coding assistants from the Visual Studio Code Marketplace. Although they only account for 1.64% of all extensions, we observe a surge in these assistants: over 90% of them are released within the past two years. We then manually analyze the user reviews sampled from 32 AI coding assistants that have sufficient installations and reviews to construct a comprehensive taxonomy of user concerns and feedback about these assistants. We manually annotate each review's attitude when mentioning certain aspects of coding assistants, yielding nuanced insights into user satisfaction and dissatisfaction regarding specific features, concerns, and overall tool performance. Built on top of the findings-including how users demand not just intelligent suggestions but also context-aware, customizable, and resource-efficient interactions-we propose five practical implications and suggestions to guide the enhancement of AI coding assistants that satisfy user needs.