π€ AI Summary
This work addresses first-order hyperbolic PDE systems with spatially varying time delaysβwhere delays depend on spatial position and conventional backstepping control requires precise delay models. We propose a model-free adaptive control framework that integrates Deep Operator Networks (DeepONets) into the Soft Actor-Critic (SAC) reinforcement learning architecture. DeepONets learn the spatiotemporal delay mapping directly from data, effectively replacing the analytical delay-dependent assumptions inherent in traditional backstepping design. Consequently, the controller operates without prior knowledge of the delay function, achieving both generalization across heterogeneous delay profiles and real-time adaptability. Simulation results demonstrate that the proposed method significantly outperforms both model-free RL baselines and classical backstepping controllers in convergence speed, robustness to delay perturbations, and steady-state control accuracy.
π Abstract
Control of distributed parameter systems affected by delays is a challenging task, particularly when the delays depend on spatial variables. The idea of integrating analytical control theory with learning-based control within a unified control scheme is becoming increasingly promising and advantageous. In this paper, we address the problem of controlling an unstable first-order hyperbolic PDE with spatially-varying delays by combining PDE backstepping control strategies and deep reinforcement learning (RL). To eliminate the assumption on the delay function required for the backstepping design, we propose a soft actor-critic (SAC) architecture incorporating a DeepONet to approximate the backstepping controller. The DeepONet extracts features from the backstepping controller and feeds them into the policy network. In simulations, our algorithm outperforms the baseline SAC without prior backstepping knowledge and the analytical controller.