🤖 AI Summary
Existing research on large language model (LLM)-driven multi-turn dialogue systems suffers from fragmented frameworks, heterogeneous technical approaches, and inconsistent evaluation protocols. Method: This paper introduces, for the first time, a unified taxonomy of four LLM adaptation paradigms for multi-turn dialogue—prompt engineering, instruction tuning, retrieval-augmented generation (RAG), and dialogue state modeling—while distinguishing open-domain and task-oriented dialogue along technical and bottleneck dimensions. It constructs a structured technical landscape covering architectures, benchmark datasets, and evaluation metrics, and synthesizes multi-dimensional automatic and human evaluation methodologies. Contribution/Results: The survey identifies three critical research gaps—interpretability, long-horizon consistency, and controllable interaction—and establishes a theoretically grounded, practice-oriented reference framework to guide future work in LLM-based dialogue systems.
📝 Abstract
This survey provides a comprehensive review of research on multi-turn dialogue systems, with a particular focus on multi-turn dialogue systems based on large language models (LLMs). This paper aims to (a) give a summary of existing LLMs and approaches for adapting LLMs to downstream tasks; (b) elaborate recent advances in multi-turn dialogue systems, covering both LLM-based open-domain dialogue (ODD) and task-oriented dialogue (TOD) systems, along with datasets and evaluation metrics; (c) discuss some future emphasis and recent research problems arising from the development of LLMs and the increasing demands on multi-turn dialogue systems.