🤖 AI Summary
To address the challenge of multi-objective, multi-tier supply chain coordination under non-stationary market conditions, this paper proposes a Markov decision process–based multi-objective reinforcement learning (MORL) framework. Methodologically, it introduces a shared experience replay buffer to enable cross-objective policy knowledge transfer and incorporates a Pareto front approximation mechanism to enhance both the quality and density of the solution set. Evaluated across multiple scenarios in a custom simulation environment, the framework outperforms weighted single-objective RL and multi-objective evolutionary algorithms (MOEAs): it achieves a 75% improvement in hypervolume metric and yields solution-set density eleven times higher than that of single-objective RL. Furthermore, it significantly reduces demand loss and improves inventory stability. The proposed approach establishes a scalable, intelligent decision-making paradigm for triple-bottom-line optimization—integrating economic, environmental, and social objectives—in sustainable supply chains.
📝 Abstract
This study develops a generalised multi-objective, multi-echelon supply chain optimisation model with non-stationary markets based on a Markov decision process, incorporating economic, environmental, and social considerations. The model is evaluated using a multi-objective reinforcement learning (RL) method, benchmarked against an originally single-objective RL algorithm modified with weighted sum using predefined weights, and a multi-objective evolutionary algorithm (MOEA)-based approach. We conduct experiments on varying network complexities, mimicking typical real-world challenges using a customisable simulator. The model determines production and delivery quantities across supply chain routes to achieve near-optimal trade-offs between competing objectives, approximating Pareto front sets. The results demonstrate that the primary approach provides the most balanced trade-off between optimality, diversity, and density, further enhanced with a shared experience buffer that allows knowledge transfer among policies. In complex settings, it achieves up to 75% higher hypervolume than the MOEA-based method and generates solutions that are approximately eleven times denser, signifying better robustness, than those produced by the modified single-objective RL method. Moreover, it ensures stable production and inventory levels while minimising demand loss.