Training-Free Adaptation of New-Generation LLMs using Legacy Clinical Models

πŸ“… 2026-01-06
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of deploying next-generation general-purpose large language models in clinical settings, where prohibitive retraining costs often hinder effective integration with existing specialized models. To overcome this limitation, the authors propose Cross-architecture Proxy Tuning (CAPT), a novel method that, for the first time, enables seamless fusion of general and clinical language models across different generations and vocabularies without any training. CAPT leverages contrastive decoding to align vocabulary-agnostic signals, thereby injecting clinical knowledge while preserving the general model’s reasoning capabilities and linguistic fluency. Evaluated on six clinical classification and generation tasks, CAPT substantially outperforms UniTE by 17.6% and proxy tuning by 41.4% on average, while also enhancing the clinical actionability and terminological specificity of generated outputs.

Technology Category

Application Category

πŸ“ Abstract
Adapting language models to the clinical domain through continued pretraining and fine-tuning requires costly retraining for each new model generation. We propose Cross-Architecture Proxy Tuning (CAPT), a model-ensembling approach that enables training-free adaptation of state-of-the-art general-domain models using existing clinical models. CAPT supports models with disjoint vocabularies, leveraging contrastive decoding to selectively inject clinically relevant signals while preserving the general-domain model's reasoning and fluency. On six clinical classification and text-generation tasks, CAPT with a new-generation general-domain model and an older-generation clinical model consistently outperforms both models individually and state-of-the-art ensembling approaches (average +17.6% over UniTE, +41.4% over proxy tuning across tasks). Through token-level analysis and physician case studies, we demonstrate that CAPT amplifies clinically actionable language, reduces context errors, and increases clinical specificity.
Problem

Research questions and friction points this paper is trying to address.

clinical adaptation
large language models
training-free
model ensembling
domain adaptation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-Architecture Proxy Tuning
training-free adaptation
clinical language models
contrastive decoding
model ensembling
πŸ”Ž Similar Papers
No similar papers found.
S
Sasha Ronaghi
Department of Biomedical Data Science, Stanford School of Medicine, Stanford University
C
Chloe Stanwyck
Department of Biomedical Data Science, Stanford School of Medicine, Stanford University
A
Asad Aali
Department of Radiology, Stanford School of Medicine, Stanford University
A
Amir Ronaghi
Department of Radiology, MemorialCare
M
Miguel Fuentes
Department of Medicine, Division of Computational Medicine, Stanford School of Medicine, Stanford University
Tina Hernandez-Boussard
Tina Hernandez-Boussard
Associate Dean of Research, Professor of Medicine, Stanford University
Clinical InformaticsHealth Services Research
Emily Alsentzer
Emily Alsentzer
Assistant Professor, Stanford University
machine learning for healthcare