π€ AI Summary
This work addresses the inefficiency and high cost of online trial-and-error adaptation in multi-agent systems following abrupt environmental changes. To mitigate excessive reliance on costly real-world interactions, the authors propose TwinLoop, a novel framework that integrates digital twins with online multi-agent reinforcement learning for the first time. TwinLoop embeds a simulation loop within the digital twin to reconstruct system states upon contextual shifts, enabling accelerated policy evaluation and optimization based on the latest strategies. The refined policy parameters are then synchronized back to the physical agents. By leveraging the digital twinβs βwhat-ifβ analysis capabilities, the approach substantially reduces dependence on expensive online exploration. Experimental results in a vehicular edge computing task offloading scenario demonstrate that TwinLoop significantly enhances adaptation efficiency in dynamic environments while markedly lowering online learning overhead.
π Abstract
Decentralised online learning enables runtime adaptation in cyber-physical multi-agent systems, but when operating conditions change, learned policies often require substantial trial-and-error interaction before recovering performance. To address this, we propose TwinLoop, a simulation-in-the-loop digital twin framework for online multi-agent reinforcement learning. When a context shift occurs, the digital twin is triggered to reconstruct the current system state, initialise from the latest agent policies, and perform accelerated policy improvement with simulation what-if analysis before synchronising updated parameters back to the agents in the physical system. We evaluate TwinLoop in a vehicular edge computing task-offloading scenario with changing workload and infrastructure conditions. The results suggest that digital twins can improve post-shift adaptation efficiency and reduce reliance on costly online trial-and-error.