Format Inertia: A Failure Mechanism of LLMs in Medical Pre-Consultation

📅 2025-10-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper identifies a novel failure mode—“format inertia”—in large language models (LLMs) deployed for multi-turn medical pre-consultation dialogues: due to severe skew in turn distribution within supervised fine-tuning (SFT) data, models over-prioritize syntactically well-formed yet semantically redundant and diagnostically uninformative questions, with degradation worsening in longer dialogues. To address this, we propose a turn-distribution rebalancing strategy for data reconstruction, systematically mitigating format inertia. Experiments demonstrate that training on the reconstructed dataset significantly reduces the rate of ineffective queries, improves information density and clinical relevance, and enhances diagnostic support capability—all while preserving formatting compliance. This work is the first to formally name, empirically characterize, and quantitatively analyze format inertia in medical dialogue systems. It provides both a novel conceptual framework for understanding LLM brittleness in clinical contexts and a reproducible, data-centric methodology for robustness optimization.

Technology Category

Application Category

📝 Abstract
Recent advances in Large Language Models (LLMs) have brought significant improvements to various service domains, including chatbots and medical pre-consultation applications. In the healthcare domain, the most common approach for adapting LLMs to multi-turn dialogue generation is Supervised Fine-Tuning (SFT). However, datasets for SFT in tasks like medical pre-consultation typically exhibit a skewed turn-count distribution. Training on such data induces a novel failure mechanism we term **Format Inertia**, where models tend to generate repetitive, format-correct, but diagnostically uninformative questions in long medical dialogues. To mitigate this observed failure mechanism, we adopt a simple, data-centric method that rebalances the turn-count distribution of the training dataset. Experimental results show that our approach substantially alleviates Format Inertia in medical pre-consultation.
Problem

Research questions and friction points this paper is trying to address.

LLMs generate repetitive uninformative medical questions
Supervised fine-tuning causes format inertia in dialogues
Imbalanced training data reduces diagnostic effectiveness
Innovation

Methods, ideas, or system contributions that make the work stand out.

Rebalanced training dataset turn-count distribution
Mitigated format inertia failure mechanism
Improved diagnostic question generation quality
🔎 Similar Papers
No similar papers found.
S
Seungseop Lim
AITRICS, KAIST
Gibaeg Kim
Gibaeg Kim
AITRICS
NLPLLM
W
Wooseok Han
AITRICS, KAIST
J
Jean Seo
AITRICS, KAIST
H
Hyunkyung Lee
AITRICS, KAIST
J
Jaehyo Yoo
AITRICS, KAIST
Eunho Yang
Eunho Yang
KAIST
Machine LearningStatistics