🤖 AI Summary
In clinical text classification, the lack of interpretability in large language model (LLM) predictions impedes their trustworthy adoption in research and clinical practice. To address this, we propose CALM—a framework for interpretable classification of semi-structured clinical texts. CALM first decomposes patient records into semantically coherent, human-understandable blocks; it then models LLM predictions as additive combinations of contributions from these blocks, embedding interpretability directly into the forward pass. Its key innovation is the explicit construction of LLM predictions as a generalized additive model (GA²M), enabling both instance-level feature attribution and population-level risk curve visualization. Experiments demonstrate that CALM preserves LLM-level classification accuracy while substantially enhancing model transparency, clinical auditability, and stakeholder trust—facilitating quality assurance and discovery of clinically meaningful patterns.
📝 Abstract
Large Language Models have advanced clinical text classification, but their opaque predictions remain a critical barrier to practical adoption in research and clinical settings where investigators and physicians need to understand which parts of a patient's record drive risk signals. To address this challenge, we introduce extbf{CALM}, short for extbf{Classification with Additive Large Language Models}, an interpretable framework for semi-structured text where inputs are composed of semantically meaningful components, such as sections of an admission note or question-answer fields from an intake form. CALM predicts outcomes as the additive sum of each component's contribution, making these contributions part of the forward computation itself and enabling faithful explanations at both the patient and population level. The additive structure also enables clear visualizations, such as component-level risk curves similar to those used in generalized additive models, making the learned relationships easier to inspect and communicate. Although CALM expects semi-structured inputs, many clinical documents already have this form, and similar structure can often be automatically extracted from free-text notes. CALM achieves performance comparable to conventional LLM classifiers while improving trust, supporting quality-assurance checks, and revealing clinically meaningful patterns during model development and auditing.