🤖 AI Summary
To address the joint optimization of cost and timeliness in dynamic cloud workflow scheduling, this paper proposes an online joint decision-making framework integrating self-attention mechanisms with evolutionary reinforcement learning. The method uniquely combines self-attention modeling of task dependency structures with genetic algorithm-optimized Deep Q-Network (DQN) policy training, enabling simultaneous minimization of virtual machine rental costs and deadline violation penalties. Extensive experiments conducted in CloudSim using real-world workflow traces demonstrate that, compared to state-of-the-art baselines, the proposed approach achieves an average 18.7% reduction in scheduling cost and a 32.4% decrease in deadline violation rate. These results substantiate significant improvements in both adaptability and cost-efficiency under dynamic cloud environments.