🤖 AI Summary
To address the poor generalization of few-shot spoken language understanding (SLU) models in low-resource settings—caused by severe scarcity of labeled data—this paper proposes a dynamic exemplar selection method leveraging lightweight lexical information retrieval (IR), specifically BM25 and TF-IDF, to construct high-quality few-shot prompts. This is the first work to integrate lexical IR into SLU prompt engineering, enabling semantic alignment and task-relevance optimization of in-context examples without increasing prompt length or computational overhead. Integrated with instruction-tuned large language models, our approach achieves substantial improvements across multiple SLU benchmarks: average relative error rates decrease by 18.7% for both intent classification and slot filling. It overcomes key limitations of conventional manual or random exemplar selection, offering an efficient, scalable, and resource-light paradigm for low-resource spoken language understanding.
📝 Abstract
Understanding user queries is fundamental in many applications, such as home assistants, booking systems, or recommendations. Accordingly, it is crucial to develop accurate Spoken Language Understanding (SLU) approaches to ensure the reliability of the considered system. Current State-of-the-Art SLU techniques rely on large amounts of training data; however, only limited annotated examples are available for specific tasks or languages. In the meantime, instruction-tuned large language models (LLMs) have shown exceptional performance on unseen tasks in a few-shot setting when provided with adequate prompts. In this work, we propose to explore example selection by leveraging Information retrieval (IR) approaches to build an enhanced prompt that is applied to an SLU task. We evaluate the effectiveness of the proposed method on several SLU benchmarks. Experimental results show that lexical IR methods significantly enhance performance without increasing prompt length.