Generative Medical Event Models Improve with Scale

📅 2025-08-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of modeling large-scale longitudinal clinical event data by proposing CoMET, a generative foundation model. Trained on the Epic Cosmos dataset, CoMET employs a decoder-only Transformer architecture and performs autoregressive pretraining on 115 billion discrete clinical events—first empirically validating and leveraging power-law scaling laws for computationally optimal model scaling in healthcare sequences. Its key contribution is zero-shot generalization across 78 diverse clinical tasks—including diagnosis prediction, disease prognosis, and healthcare operations—without task-specific fine-tuning, matching or surpassing supervised task-specific models. Performance improves consistently and stably with increasing parameter count across all metrics. Results demonstrate that generative modeling effectively captures the complex temporal dynamics of patient clinical trajectories, establishing a scalable, general-purpose foundation model paradigm for personalized medicine.

Technology Category

Application Category

📝 Abstract
Realizing personalized medicine at scale calls for methods that distill insights from longitudinal patient journeys, which can be viewed as a sequence of medical events. Foundation models pretrained on large-scale medical event data represent a promising direction for scaling real-world evidence generation and generalizing to diverse downstream tasks. Using Epic Cosmos, a dataset with medical events from de-identified longitudinal health records for 16.3 billion encounters over 300 million unique patient records from 310 health systems, we introduce the Cosmos Medical Event Transformer ( CoMET) models, a family of decoder-only transformer models pretrained on 118 million patients representing 115 billion discrete medical events (151 billion tokens). We present the largest scaling-law study for medical event data, establishing a methodology for pretraining and revealing power-law scaling relationships for compute, tokens, and model size. Based on this, we pretrained a series of compute-optimal models with up to 1 billion parameters. Conditioned on a patient's real-world history, CoMET autoregressively generates the next medical event, simulating patient health timelines. We studied 78 real-world tasks, including diagnosis prediction, disease prognosis, and healthcare operations. Remarkably for a foundation model with generic pretraining and simulation-based inference, CoMET generally outperformed or matched task-specific supervised models on these tasks, without requiring task-specific fine-tuning or few-shot examples. CoMET's predictive power consistently improves as the model and pretraining scale. Our results show that CoMET, a generative medical event foundation model, can effectively capture complex clinical dynamics, providing an extensible and generalizable framework to support clinical decision-making, streamline healthcare operations, and improve patient outcomes.
Problem

Research questions and friction points this paper is trying to address.

Develop generative models for simulating patient medical event sequences
Scale pretraining for medical event data to improve predictive accuracy
Generalize foundation models to diverse clinical tasks without fine-tuning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decoder-only transformer models pretrained on medical events
Largest scaling-law study for medical event data
Generative model for simulating patient health timelines
🔎 Similar Papers
No similar papers found.
S
Shane Waxler
Epic Systems
P
Paul Blazek
Epic Systems
D
Davis White
Epic Systems
D
Daniel Sneider
Epic Systems
K
Kevin Chung
Epic Systems
M
Mani Nagarathnam
Epic Systems
P
Patrick Williams
Epic Systems
H
Hank Voeller
Epic Systems
K
Karen Wong
Epic Systems
M
Matthew Swanhorst
Epic Systems
S
Sheng Zhang
Microsoft Research
Naoto Usuyama
Naoto Usuyama
Principal Researcher, Microsoft Research
Artificial IntelligencePrecision MedicineComputer VisionNatural Language Processing
C
Cliff Wong
Microsoft Research
Tristan Naumann
Tristan Naumann
Principal Researcher, Microsoft Research Health Futures
Artificial IntelligenceMachine LearningNatural Language ProcessingClinical Inference
Hoifung Poon
Hoifung Poon
General Manager, Microsoft Health Futures
precision healthreal-world evidencelarge language modelsmultimodal GenAI
A
Andrew Loza
Yale School of Medicine
Daniella Meeker
Daniella Meeker
Associate Professor, Biomedical Informatics and Data Science, Yale School of Medicine
S
Seth Hain
Epic Systems
R
Rahul Shah
Epic Systems