๐ค AI Summary
Large language models (LLMs) often implicitly commit to a single interpretation of ambiguous user queries, leading to intent misidentification and safety risks.
Method: We propose a structured multi-intent response framework that generates, in a single forward pass, multiple semantically plausible and mutually exclusive interpretations of the userโs intentโeach paired with a corresponding answer. Our approach employs a reinforcement learning paradigm with a novel multi-answer consistency reward function, leveraging multiple valid outputs as weak supervision signals.
Contribution/Results: Unlike iterative clarification or single-answer generation, our method achieves high response efficiency, transparency, and coverage without requiring additional user interaction. Experiments on dialogue QA and semantic parsing tasks demonstrate a +18.7% improvement in effective answer coverage. Human evaluation confirms 92.4% consistency between inferred intents and their associated answers, validating both efficacy and practical utility.
๐ Abstract
Large language models often respond to ambiguous requests by implicitly committing to one interpretation. Intent misunderstandings can frustrate users and create safety risks. To address this, we propose generating multiple interpretation-answer pairs in a single structured response to ambiguous requests. Our models are trained with reinforcement learning and customized reward functions using multiple valid answers as supervision. Experiments on conversational question answering and semantic parsing demonstrate that our method achieves higher coverage of valid answers than baseline approaches. Human evaluation confirms that predicted interpretations are highly aligned with their answers. Our approach promotes transparency with explicit interpretations, achieves efficiency by requiring only one generation step, and supports downstream applications through its structured output format.