🤖 AI Summary
To address the challenges of labor-intensive instruction authoring, poor interpretability, and high human cost in AI behavioral specification, this paper proposes Visionary Tuning: a human-in-the-loop, closed-loop prompt optimization paradigm. Our method leverages LLM-driven self-play to simulate user interactions and uncover latent behavioral patterns, integrating iterative human feedback for automated prompt generation and refinement. Crucially, it is the first approach to formalize behavioral specifications as observable, iterative interactive prompt engineering—thereby revealing non-obvious cognitive effects in tool usage. A user study (N=12) demonstrates significant improvements in pattern recognition accuracy and prompt quality; crowdsourced evaluation (N=60) confirms high instruction adherence and consistent output quality; and cross-preference movie rating case studies validate robustness across diverse user preference spectra.
📝 Abstract
Training AI models is challenging, particularly when crafting behavior instructions. Traditional methods rely on machines (supervised learning) or manual pattern discovery, which results in not interpretable models or time sink. While Large Language Models (LLMs) simplify instruction writing through natural language, articulating intended model behavior still remains difficult. We introduce Visionary Tuning, a human-in-the-loop self-playing followed by automatic self-refinement to improve behavior specification. Our system helps users clarify desired behavior through self-playing and generates prompts through self-improving, Our first evaluation involves user study conducted on a system implementation of Visionary Tuning within the context of chatbot behavior. Our system self-play itself by simulating user interactions to identify patterns and create effective prompts based on the pattern. In a within-subject study (N=12), participants pinpointed more patterns through self-playing and crafted better prompts. Surprisingly, users felt more or less success level in specifying the model behavior. Follow-up crowd studies (N=60) confirmed that the chatbot adhered to instructions without sacrificing quality. Our second evaluation is a case study on a real-world implementation using a movie rating dataset with Visionary Tuning, demonstrating its effectiveness and robustness in modeling a critic's preferences across the spectrum of low to highly rated movies. Together, these results suggest how AI improves the design process of interactive AI systems. Furthermore, they suggest how the benefits of these tools may be non-obvious to end-users. We reflect on these findings and suggest future directions.