Improving Clinical Diagnosis with Counterfactual Multi-Agent Reasoning

📅 2026-03-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the lack of explicit hypothesis-testing mechanisms in current large language models (LLMs) for clinical diagnosis, which renders their reasoning processes opaque. To this end, the authors propose a counterfactual multi-agent diagnostic framework inspired by medical training: it generates counterfactual cases by editing clinical findings and orchestrates iterative, critical discussions among multiple LLM agents to refine differential diagnoses. The approach explicitly formalizes diagnostic reasoning through counterfactual case editing and introduces a counterfactual probability gap metric to quantify diagnostic confidence. Integrating LLMs, counterfactual reasoning, multi-agent deliberation, and evidence-based confidence assessment, the framework consistently improves diagnostic accuracy across three benchmark datasets and seven LLMs, with particularly strong performance on complex, ambiguous cases. Human evaluations further confirm that its reasoning is more clinically coherent and practically useful.
📝 Abstract
Clinical diagnosis is a complex reasoning process in which clinicians gather evidence, form hypotheses, and test them against alternative explanations. In medical training, this reasoning is explicitly developed through counterfactual questioning--e.g., asking how a diagnosis would change if a key symptom were absent or altered--to strengthen differential diagnosis skills. As large language model (LLM)-based systems are increasingly used for diagnostic support, ensuring the interpretability of their recommendations becomes critical. However, most existing LLM-based diagnostic agents reason over fixed clinical evidence without explicitly testing how individual findings support or weaken competing diagnoses. In this work, we propose a counterfactual multi-agent diagnostic framework inspired by clinician training that makes hypothesis testing explicit and evidence-grounded. Our framework introduces counterfactual case editing to modify clinical findings and evaluate how these changes affect competing diagnoses. We further define the Counterfactual Probability Gap, a method that quantifies how strongly individual findings support a diagnosis by measuring confidence shifts under these edits. These counterfactual signals guide multi-round specialist discussions, enabling agents to challenge unsupported hypotheses, refine differential diagnoses, and produce more interpretable reasoning trajectories. Across three diagnostic benchmarks and seven LLMs, our method consistently improves diagnostic accuracy over prompting and prior multi-agent baselines, with the largest gains observed in complex and ambiguous cases. Human evaluation further indicates that our framework produces more clinically useful, reliable, and coherent reasoning. These results suggest that incorporating counterfactual evidence verification is an important step toward building reliable AI systems for clinical decision support.
Problem

Research questions and friction points this paper is trying to address.

clinical diagnosis
counterfactual reasoning
interpretability
differential diagnosis
large language models
Innovation

Methods, ideas, or system contributions that make the work stand out.

counterfactual reasoning
multi-agent diagnostic framework
Counterfactual Probability Gap
interpretable AI
clinical decision support
🔎 Similar Papers
No similar papers found.