Multi-agent Reinforcement Learning-based In-place Scaling Engine for Edge-cloud Systems

📅 2025-07-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the sluggish and inefficient resource scaling responses of edge-cloud systems under dynamic workloads, this paper proposes an in-situ scaling engine based on multi-agent reinforcement learning (MARL). It is the first to introduce a MARL framework for in-situ scheduling in edge-cloud environments, enabling distributed agents to collaboratively make real-time scaling decisions guided by microservice performance feedback. The engine employs a dual-algorithm architecture integrating Deep Q-Networks (DQN) and Proximal Policy Optimization (PPO), jointly optimizing for low latency and high resource efficiency. Experimental results demonstrate that, compared to conventional heuristic strategies, the proposed approach reduces average scaling response latency by 23.6% and improves CPU utilization by 31.4%, thereby significantly enhancing system adaptability and scalability.

Technology Category

Application Category

📝 Abstract
Modern edge-cloud systems face challenges in efficiently scaling resources to handle dynamic and unpredictable workloads. Traditional scaling approaches typically rely on static thresholds and predefined rules, which are often inadequate for optimizing resource utilization and maintaining performance in distributed and dynamic environments. This inefficiency hinders the adaptability and performance required in edge-cloud infrastructures, which can only be achieved through the newly proposed in-place scaling. To address this problem, we propose the Multi-Agent Reinforcement Learning-based In-place Scaling Engine (MARLISE) that enables seamless, dynamic, reactive control with in-place resource scaling. We develop our solution using two Deep Reinforcement Learning algorithms: Deep Q-Network (DQN), and Proximal Policy Optimization (PPO). We analyze each version of the proposed MARLISE solution using dynamic workloads, demonstrating their ability to ensure low response times of microservices and scalability. Our results show that MARLISE-based approaches outperform heuristic method in managing resource elasticity while maintaining microservice response times and achieving higher resource efficiency.
Problem

Research questions and friction points this paper is trying to address.

Dynamic workload scaling in edge-cloud systems
Inefficiency of traditional static scaling methods
Optimizing resource utilization and performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-agent Reinforcement Learning for dynamic scaling
In-place resource scaling with DQN and PPO
Low response times and high resource efficiency
🔎 Similar Papers
J
Jovan Prodanov
Jožef Stefan Institute, Ljubljana, Slovenia.
B
Blaž Bertalanič
Jožef Stefan Institute, Ljubljana, Slovenia.
Carolina Fortuna
Carolina Fortuna
Jozef Stefan Institute
artificial intelligencecyber-physical systems
S
Shih-Kai Chou
Jožef Stefan Institute, Ljubljana, Slovenia.
M
Matjaž Branko Jurič
Faculty of Computer and Information Science, University of Ljubljana, Ljubljana, Slovenia
R
Ramon Sanchez-Iborra
Department of Information and Communication Engineering, University of Murcia, Murcia, Spain
J
Jernej Hribar
Jožef Stefan Institute, Ljubljana, Slovenia.