🤖 AI Summary
Large language models (LLMs) frequently generate factual hallucinations due to ambiguous or incomplete user prompts, undermining output reliability. This work is the first to systematically identify low-quality user-side prompting as a primary cause of hallucination. To address this, we propose a plug-and-play prompt optimization framework that employs lightweight fine-tuning of a small language model to perform prompt cleaning and intent alignment, while automatically generating structured, information-complete supplementary task descriptions—entirely without external knowledge bases. Our method significantly enhances semantic clarity and task solvability of input prompts. Extensive evaluation across multiple LLMs demonstrates over 90% win rate against baselines, substantially mitigating hallucination and improving output accuracy. The framework exhibits strong generalizability across diverse tasks and models, and enables seamless deployment with minimal computational overhead.
📝 Abstract
Recent advancements in large language models (LLMs) highlight their fluency in generating responses to diverse prompts. However, these models sometimes generate plausible yet incorrect ``hallucinated" facts, undermining trust. A frequent but often overlooked cause of such errors is the use of poorly structured or vague prompts by users, leading LLMs to base responses on assumed rather than actual intentions. To mitigate hallucinations induced by these ill-formed prompts, we introduce Curative Prompt Refinement (CPR), a plug-and-play framework for curative prompt refinement that 1) cleans ill-formed prompts, and 2) generates additional informative task descriptions to align the intention of the user and the prompt using a fine-tuned small language model. When applied to language models, we discover that CPR significantly increases the quality of generation while also mitigating hallucination. Empirical studies show that prompts with CPR applied achieves over a 90% win rate over the original prompts without any external knowledge.