Defining Boundaries: A Spectrum of Task Feasibility for Large Language Models

πŸ“… 2024-08-11
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 3
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Large language models (LLMs) frequently generate hallucinations or erroneous responses when confronted with queries beyond their capability boundaries, necessitating robust task feasibility identification and proactive refusal mechanisms. Method: We systematically construct the first taxonomy of LLM task infeasibility, covering diverse hallucination scenarios, and introduce UNFEASIBLE-Benchβ€”the first benchmark dataset dedicated to infeasible task identification. We further propose a capability-boundary-aware refusal mechanism: a binary classification framework for task feasibility assessment, supported by high-quality refusal annotations and supervised fine-tuning (SFT) to optimize refusal policies. Contribution/Results: Experiments demonstrate substantial improvements across mainstream LLMs: +28.7% in infeasible-task recognition accuracy and +34.1% in refusal reasonableness. This work provides both theoretical foundations and empirical evidence for the safe, controllable deployment of LLMs in safety-critical applications.

Technology Category

Application Category

πŸ“ Abstract
Large language models (LLMs) have shown remarkable performance in various tasks but often fail to handle queries that exceed their knowledge and capabilities, leading to incorrect or fabricated responses. This paper addresses the need for LLMs to recognize and refuse infeasible tasks due to the required skills surpassing their capabilities. We first conceptualize infeasible tasks for LLMs and provide categorizations that cover a spectrum of related hallucinations over existing literature. We develop and benchmark a new dataset comprising diverse infeasible and feasible tasks to evaluate multiple LLMs' abilities to reject infeasible tasks. Furthermore, we explore the potential of increasing LLMs' refusal capabilities with fine-tuning. Experiments validate the effectiveness of our trained models, offering promising directions for refining the operational boundaries of LLMs in real applications.
Problem

Research questions and friction points this paper is trying to address.

LLMs fail to handle queries beyond their capabilities
Addressing refusal of infeasible tasks to prevent hallucinations
Evaluating and improving models' ability to decline impossible requests
Innovation

Methods, ideas, or system contributions that make the work stand out.

Developed dataset for infeasible task evaluation
Fine-tuned models to enhance refusal capabilities
Categorized four types of LLM infeasible tasks
πŸ”Ž Similar Papers
No similar papers found.
W
Wenbo Zhang
University of California Irvine
Z
Zihang Xu
University of California Irvine
H
Hengrui Cai
University of California Irvine