🤖 AI Summary
DeFi lending faces challenges including rigid interest rate models, high default rates, and low capital efficiency. This paper proposes a dynamic interest rate governance framework based on offline reinforcement learning (RL), the first to apply TD3-BC, conservative Q-learning (CQL), and behavioral cloning to DeFi rate control. The model is trained and evaluated exclusively on historical on-chain data from Aave—including critical crisis events such as the 2021 market crash and the 2023 USDC depeg—without requiring online, on-chain training. The approach is inherently risk-aware and adaptive. Experimental results demonstrate that TD3-BC achieves the optimal trade-off among fund utilization, capital stability, and default mitigation, significantly outperforming conventional rule-based models. This validates the feasibility and effectiveness of offline RL–driven, automated, and shock-resilient interest rate governance in DeFi.
📝 Abstract
Decentralized Finance (DeFi) lending enables permissionless borrowing via smart contracts. However, it faces challenges in optimizing interest rates, mitigating bad debt, and improving capital efficiency. Rule-based interest-rate models struggle to adapt to dynamic market conditions, leading to inefficiencies. This work applies Offline Reinforcement Learning (RL) to optimize interest rate adjustments in DeFi lending protocols. Using historical data from Aave protocol, we evaluate three RL approaches: Conservative Q-Learning (CQL), Behavior Cloning (BC), and TD3 with Behavior Cloning (TD3-BC). TD3-BC demonstrates superior performance in balancing utilization, capital stability, and risk, outperforming existing models. It adapts effectively to historical stress events like the May 2021 crash and the March 2023 USDC depeg, showcasing potential for automated, real-time governance.