Toward Better EHR Reasoning in LLMs: Reinforcement Learning with Expert Attention Guidance

๐Ÿ“… 2025-08-19
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the poor generalization and difficulty in modeling high-dimensional structured data of large language models (LLMs) for temporal reasoning on electronic health records (EHR), this paper proposes EAG-RL, a two-stage training framework. In Stage I, an expert model guides Monte Carlo Tree Search to generate high-quality clinical reasoning paths. In Stage II, attention mechanisms align the expertโ€™s decision logic with the LLMโ€™s internal representations, enabling interpretable end-to-end reinforcement learning. Unlike conventional hybrid paradigms that treat LLMs merely as feature retrievers, EAG-RL enhances the LLMโ€™s intrinsic reasoning capability. Evaluated on two real-world EHR datasets, EAG-RL achieves an average 14.62% improvement in predictive performance, demonstrates enhanced robustness to feature perturbations, and exhibits strong cross-clinical-domain generalization.

Technology Category

Application Category

๐Ÿ“ Abstract
Improving large language models (LLMs) for electronic health record (EHR) reasoning is essential for enabling accurate and generalizable clinical predictions. While LLMs excel at medical text understanding, they underperform on EHR-based prediction tasks due to challenges in modeling temporally structured, high-dimensional data. Existing approaches often rely on hybrid paradigms, where LLMs serve merely as frozen prior retrievers while downstream deep learning (DL) models handle prediction, failing to improve the LLM's intrinsic reasoning capacity and inheriting the generalization limitations of DL models. To this end, we propose EAG-RL, a novel two-stage training framework designed to intrinsically enhance LLMs' EHR reasoning ability through expert attention guidance, where expert EHR models refer to task-specific DL models trained on EHR data. Concretely, EAG-RL first constructs high-quality, stepwise reasoning trajectories using expert-guided Monte Carlo Tree Search to effectively initialize the LLM's policy. Then, EAG-RL further optimizes the policy via reinforcement learning by aligning the LLM's attention with clinically salient features identified by expert EHR models. Extensive experiments on two real-world EHR datasets show that EAG-RL improves the intrinsic EHR reasoning ability of LLMs by an average of 14.62%, while also enhancing robustness to feature perturbations and generalization to unseen clinical domains. These results demonstrate the practical potential of EAG-RL for real-world deployment in clinical prediction tasks. Our code have been available at https://github.com/devilran6/EAG-RL.
Problem

Research questions and friction points this paper is trying to address.

Improving LLMs' EHR reasoning for clinical predictions
Addressing underperformance on EHR-based prediction tasks
Enhancing intrinsic reasoning capacity beyond hybrid paradigms
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reinforcement learning with expert attention guidance
Monte Carlo Tree Search for stepwise reasoning
Aligning LLM attention with clinical features
๐Ÿ”Ž Similar Papers
No similar papers found.
Y
Yue Fang
School of Computer Science, Peking University, Beijing, China
Y
Yuxin Guo
School of Computer Science, Peking University, Beijing, China
J
Jiaran Gao
School of Computer Science, Peking University, Beijing, China
H
Hongxin Ding
School of Computer Science, Peking University, Beijing, China
X
Xinke Jiang
School of Computer Science, Peking University, Beijing, China
Weibin Liao
Weibin Liao
Peking University
Large Language ModelReinforcement LearningMedical Image Analysis
Yongxin Xu
Yongxin Xu
Peking University
Large Language ModelsKnowledge GraphsElectronic Medical Record Analysis
Yinghao Zhu
Yinghao Zhu
The University of Hong Kong
Data MiningAI for Healthcare
Z
Zhibang Yang
School of Computer Science, Peking University, Beijing, China
L
Liantao Ma
National Engineering Research Center For Software Engineering, Peking University, Beijing, China
Junfeng Zhao
Junfeng Zhao
Assistant Professor at Arizona State University, Director of BELIV Lab
Connected & Automated VehicleMotion Planning & ControlsElectric VehiclesAI/ML
Y
Yasha Wang
National Engineering Research Center For Software Engineering, Peking University, Beijing, China