π€ AI Summary
This work proposes a supervised deep learning framework that abandons the conventional self-supervised pretraining paradigm in favor of direct supervision-driven representation learning, explicitly optimizing the geometric structure of learned representations to maximize the ratio of inter-class separation to intra-class variance. By doing so, the model focuses on clinically relevant dimensions, effectively incorporating inductive biases suited for high-quality, outcome-oriented clinical prediction tasks. Trained end-to-end in a single stage, the approach significantly outperforms multiple self-supervised baselines on electronic health recordβbased prediction tasks such as mortality and readmission risk. The method demonstrates consistent improvements in discriminative performance, calibration, and sample efficiency, highlighting the advantages of supervision-guided representation learning in clinical settings where labeled outcomes are reliable and task-specific.
π Abstract
Foundation models in healthcare have largely adopted self supervised pretraining objectives inherited from natural language processing and computer vision, emphasizing reconstruction and large scale representation learning prior to downstream adaptation. We revisit this paradigm in outcome centric clinical prediction settings and argue that, when high quality supervision is available, direct outcome alignment may provide a stronger inductive bias than generative pretraining. We propose a supervised deep learning framework that explicitly shapes representation geometry by maximizing inter class separation relative to within class variance, thereby concentrating model capacity along clinically meaningful axes. Across multiple longitudinal electronic health record tasks, including mortality and readmission prediction, our approach consistently outperforms masked, autoregressive, and contrastive pretraining baselines under matched model capacity. The proposed method improves discrimination, calibration, and sample efficiency, while simplifying the training pipeline to a single stage optimization. These findings suggest that in low entropy, outcome driven healthcare domains, supervision can act as the statistically optimal driver of representation learning, challenging the assumption that large scale self supervised pretraining is a prerequisite for strong clinical performance.