GUIDE: Reinforcement Learning for Behavioral Action Support in Type 1 Diabetes

📅 2026-03-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current management of type 1 diabetes primarily relies on insulin dosing and lacks personalized behavioral interventions—particularly regarding carbohydrate intake—making sustained glycemic control challenging. This work proposes GUIDE, a reinforcement learning–based decision support framework that, for the first time in offline reinforcement learning, introduces a structured behavioral action space to jointly recommend both bolus insulin doses and the type, intensity, and timing of carbohydrate consumption. Integrating a personalized glucose prediction model and employing conservative off-policy algorithms such as CQL-BC, the approach achieves an average time-in-range of 85.49% across 25 patients, with low hypoglycemia risk and high alignment between recommended actions and patients’ historical behaviors (cosine similarity: 0.87 ± 0.09).
📝 Abstract
Type 1 Diabetes (T1D) management requires continuous adjustment of insulin and lifestyle behaviors to maintain blood glucose within a safe target range. Although automated insulin delivery (AID) systems have improved glycemic outcomes, many patients still fail to achieve recommended clinical targets, warranting new approaches to improve glucose control in patients with T1D. While reinforcement learning (RL) has been utilized as a promising approach, current RL-based methods focus primarily on insulin-only treatment and do not provide behavioral recommendations for glucose control. To address this gap, we propose GUIDE, an RL-based decision-support framework designed to complement AID technologies by providing behavioral recommendations to prevent abnormal glucose events. GUIDE generates structured actions defined by intervention type, magnitude, and timing, including bolus insulin administration and carbohydrate intake events. GUIDE integrates a patient-specific glucose level predictor trained on real-world continuous glucose monitoring data and supports both offline and online RL algorithms within a unified environment. We evaluate both off-policy and on-policy methods across 25 individuals with T1D using standardized glycemic metrics. Among the evaluated approaches, the CQL-BC algorithm demonstrates the highest average time-in-range, reaching 85.49% while maintaining low hypoglycemia exposures. Behavioral similarity analysis further indicates that the learned CQL-BC policy preserves key structural characteristics of patient action patterns, achieving a mean cosine similarity of 0.87 $\pm$ 0.09 across subjects. These findings suggest that conservative offline RL with a structured behavioral action space can provide clinically meaningful and behaviorally plausible decision support for personalized diabetes management.
Problem

Research questions and friction points this paper is trying to address.

Type 1 Diabetes
Reinforcement Learning
Behavioral Recommendations
Glucose Control
Decision Support
Innovation

Methods, ideas, or system contributions that make the work stand out.

structured behavioral action space
offline reinforcement learning
personalized glucose prediction
behavioral decision support
conservative Q-learning
🔎 Similar Papers
No similar papers found.
Saman Khamesian
Saman Khamesian
Arizona State University
eHealthMobile HealthWearable SensorsMachine LearningNeural Networks
S
Sri Harini Balaji
School for Engineering of Matter Transport and Energy, Arizona State University, Tempe, USA
D
Di Yang Shi
Institute for Foundation of Machine Learning, The University of Texas at Austin, Austin, USA
S
Stephanie M. Carpenter
College of Health Solutions, Arizona State University, Phoenix, USA
Daniel E. Rivera
Daniel E. Rivera
Professor of Chemical Engineering, Arizona State University
control systems engineeringprocess controlsystem identificationbehavioral interventionssupply chain management
W
W. Bradley Knox
Institute for Foundation of Machine Learning, The University of Texas at Austin, Austin, USA
Peter Stone
Peter Stone
Professor of Computer Science, The University of Texas at Austin
Artificial IntelligenceMachine LearningReinforcement LearningMultiagent SystemsRobotics
Hassan Ghasemzadeh
Hassan Ghasemzadeh
Arizona State University
Digital HealthMachine LearningEmbedded SystemsAlgorithms