PromptPilot: Improving Human-AI Collaboration Through LLM-Enhanced Prompt Engineering

📅 2025-10-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In knowledge-intensive tasks, non-expert users struggle to craft effective prompts, limiting the practical efficacy of large language models (LLMs). To address this, we introduce PromptGuide—the first interactive, human-in-the-loop prompt engineering system, grounded in four empirically derived design principles. PromptGuide integrates LLM-based reasoning with principled human-computer interaction techniques to form a prompt enhancement framework that provides real-time feedback, strategy recommendations, and closed-loop optimization of user intent. A double-blind randomized controlled study (N=80) demonstrates that users employing PromptGuide achieve a significantly higher median task accuracy (78.3 vs. 61.7 for controls; p<0.01), alongside substantial improvements in efficiency, usability, and operational autonomy (all p<0.01). This work establishes a validated, interactive paradigm for prompt engineering that lowers the barrier to LLM adoption and strengthens user agency.

Technology Category

Application Category

📝 Abstract
Effective prompt engineering is critical to realizing the promised productivity gains of large language models (LLMs) in knowledge-intensive tasks. Yet, many users struggle to craft prompts that yield high-quality outputs, limiting the practical benefits of LLMs. Existing approaches, such as prompt handbooks or automated optimization pipelines, either require substantial effort, expert knowledge, or lack interactive guidance. To address this gap, we design and evaluate PromptPilot, an interactive prompting assistant grounded in four empirically derived design objectives for LLM-enhanced prompt engineering. We conducted a randomized controlled experiment with 80 participants completing three realistic, work-related writing tasks. Participants supported by PromptPilot achieved significantly higher performance (median: 78.3 vs. 61.7; p = .045, d = 0.56), and reported enhanced efficiency, ease-of-use, and autonomy during interaction. These findings empirically validate the effectiveness of our proposed design objectives, establishing LLM-enhanced prompt engineering as a viable technique for improving human-AI collaboration.
Problem

Research questions and friction points this paper is trying to address.

Users struggle to craft effective prompts for LLMs
Existing prompt engineering methods lack interactive guidance
Need to improve human-AI collaboration through enhanced prompting
Innovation

Methods, ideas, or system contributions that make the work stand out.

Interactive prompting assistant for human-AI collaboration
Empirically derived design objectives for LLM enhancement
Randomized controlled experiment validating performance improvement
🔎 Similar Papers
No similar papers found.
N
Niklas Gutheil
University of Bayreuth
V
Valentin Mayer
University of Bayreuth
L
Leopold Müller
University of Bayreuth
J
Jörg Rommelt
FIM Research Center for Information Management
Niklas Kühl
Niklas Kühl
University of Bayreuth
Artificial IntelligenceHuman-AI-TeamsFairness in Machine LearningAppropriate Reliance