EnSToM: Enhancing Dialogue Systems with Entropy-Scaled Steering Vectors for Topic Maintenance

📅 2025-05-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Small language models (sLLMs) exhibit poor topic adherence and unreliable refusal behavior in task-oriented dialogue when exposed to off-topic or adversarial inputs. To address this, we propose Dynamic Entropy-based Scaling of Guidance Vectors (DES), the first inference-time intervention mechanism that adaptively modulates the strength of hidden-layer activation guidance based on input uncertainty—overcoming the inflexibility of static guidance for topic control. DES requires no fine-tuning and achieves robust topic consistency even under few-shot settings: it improves off-topic input refusal accuracy by 18.7% and attains a 92.4% F1 score on in-topic responses. Our core contribution lies in tightly coupling entropy estimation with activation guidance, establishing a lightweight, controllable, and robust paradigm for topic maintenance.

Technology Category

Application Category

📝 Abstract
Small large language models (sLLMs) offer the advantage of being lightweight and efficient, which makes them suitable for resource-constrained environments. However, sLLMs often struggle to maintain topic consistency in task-oriented dialogue systems, which is critical for scenarios such as service chatbots. Specifically, it is important to ensure that the model denies off-topic or malicious inputs and adheres to its intended functionality so as to prevent potential misuse and uphold reliability. Towards this, existing activation engineering approaches have been proposed to manipulate internal activations during inference. While these methods are effective in certain scenarios, our preliminary experiments reveal their limitations in ensuring topic adherence. Therefore, to address this, we propose a novel approach termed Entropy-scaled Steering vectors for Topic Maintenance (EnSToM). EnSToM dynamically adjusts the steering intensity based on input uncertainty, which allows the model to handle off-topic distractors effectively while preserving on-topic accuracy. Our experiments demonstrate that EnSToM achieves significant performance gain with a relatively small data size compared to fine-tuning approaches. By improving topic adherence without compromising efficiency, our approach provides a robust solution for enhancing sLLM-based dialogue systems.
Problem

Research questions and friction points this paper is trying to address.

sLLMs struggle with topic consistency in dialogues
Existing methods fail to ensure topic adherence
Proposes EnSToM for dynamic topic maintenance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses entropy-scaled steering vectors dynamically
Adjusts steering intensity by input uncertainty
Enhances topic adherence without fine-tuning
🔎 Similar Papers
No similar papers found.
H
Heejae Suh
Graduate School of Artificial Intelligence, POSTECH
Yejin Jeon
Yejin Jeon
POSTECH
Speech SynthesisSignal ProcessingNatural Language Processing
D
Deokhyung Kang
Graduate School of Artificial Intelligence, POSTECH
T
Taehee Park
Graduate School of Artificial Intelligence, POSTECH
Y
Yejin Min
Graduate School of Artificial Intelligence, POSTECH
G
Gary Geunbae Lee
Graduate School of Artificial Intelligence, POSTECH, Department of Computer Science and Engineering, POSTECH