🤖 AI Summary
To address the trade-offs among reliability, real-time schedulability, thermal safety, and power consumption in task replication for multicore embedded systems, this paper proposes a dynamic task replication optimization method based on Proximal Policy Optimization (PPO), a deep reinforcement learning algorithm. It is the first work to employ reinforcement learning for runtime adaptive determination of replica count, jointly optimizing system reliability objectives and core-level Thermal-Safe Power (TSP) constraints. The approach integrates real-time schedulability analysis, physics-aware thermal modeling, and power-aware task mapping into a unified framework. Compared with state-of-the-art methods, the proposed solution reduces power consumption by 63%, improves schedulability by 53%, and increases TSP constraint satisfaction rate by 72%. It effectively mitigates resource underutilization and thermal violation risks inherent in static replication strategies.
📝 Abstract
Embedded systems power many modern applications and must often meet strict reliability, real-time, thermal, and power requirements. Task replication can improve reliability by duplicating a task's execution to handle transient and permanent faults, but blindly applying replication often leads to excessive overhead and higher temperatures. Existing design-time methods typically choose the number of replicas based on worst-case conditions, which can waste resources under normal operation. In this paper, we present RL-TIME, a reinforcement learning-based approach that dynamically decides the number of replicas according to actual system conditions. By considering both the reliability target and a core-level Thermal Safe Power (TSP) constraint at run-time, RL-TIME adapts the replication strategy to avoid unnecessary overhead and overheating. Experimental results show that, compared to state-of-the-art methods, RL-TIME reduces power consumption by 63%, increases schedulability by 53%, and respects TSP 72% more often.