PRINCIPLES: Synthetic Strategy Memory for Proactive Dialogue Agents

📅 2025-09-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Contemporary conversational agents face three key limitations in proactive strategy planning: narrow strategy coverage, preference bias, and reliance on costly fine-tuning. To address these, we propose a fine-tuning-free synthetic strategy memory mechanism. It leverages large language models to generate a reusable strategy library via offline self-play, enabling dynamic retrieval and execution of strategic guidance during inference for cross-domain, long-horizon planning. By eliminating the need for human annotation and model training, our approach significantly improves strategy diversity and robustness. Empirical evaluation on emotion-support and persuasion dialogue tasks demonstrates superior performance over strong baselines; notably, gains remain stable across longer dialogue turns and more diverse scenarios, validating both generalization capability and practical utility.

Technology Category

Application Category

📝 Abstract
Dialogue agents based on large language models (LLMs) have shown promising performance in proactive dialogue, which requires effective strategy planning. However, existing approaches to strategy planning for proactive dialogue face several limitations: limited strategy coverage, preference bias in planning, and reliance on costly additional training. To address these, we propose PRINCIPLES: a synthetic strategy memory for proactive dialogue agents. PRINCIPLES is derived through offline self-play simulations and serves as reusable knowledge that guides strategy planning during inference, eliminating the need for additional training and data annotation. We evaluate PRINCIPLES in both emotional support and persuasion domains, demonstrating consistent improvements over strong baselines. Furthermore, PRINCIPLES maintains its robustness across extended and more diverse evaluation settings. See our project page at https://huggingface.co/spaces/kimnamssya/Principles.
Problem

Research questions and friction points this paper is trying to address.

Addresses limited strategy coverage in proactive dialogue planning
Overcomes preference bias and costly training requirements
Enhances proactive dialogue agents without additional training
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses synthetic strategy memory from self-play simulations
Provides reusable knowledge for strategy planning
Eliminates need for additional training and annotation
🔎 Similar Papers
No similar papers found.
Namyoung Kim
Namyoung Kim
Yonsei University
Large Language ModelAI Agent
Kai Tzu-iunn Ong
Kai Tzu-iunn Ong
Yonsei University
Large Language ModelsDialogue AgentsAI for ScienceHuman-centred AI
Y
Yeonjun Hwang
Department of Artificial Intelligence, Yonsei University
Minseok Kang
Minseok Kang
Gangneung-Wonju National University
Agricultural and Forest MeteorologyEddy CovarianceClimate-Smart Agriculture
I
Iiseo Jihn
Department of Artificial Intelligence, Yonsei University
G
Gayoung Kim
Department of Artificial Intelligence, Yonsei University
M
Minju Kim
Department of Artificial Intelligence, Yonsei University
Jinyoung Yeo
Jinyoung Yeo
Yonsei University
Natural Language ProceesingLarge Language ModelsAI Agents