🤖 AI Summary
In multi-agent multi-objective systems (MAMOS), heterogeneous utility functions induce non-stationary learning dynamics and hinder convergence to Bayesian Nash equilibria (BNE). Method: We propose the first theoretically grounded framework for BNE approximation under decentralized execution, establishing that global utility awareness is a necessary condition for BNE—thereby designing a centralized-training-with-decentralized-execution paradigm. Our approach implicitly models joint beliefs over other agents’ utilities and policies via attention-based mechanisms (Agent-Attention), integrates multi-agent deep reinforcement learning, and explicitly models the utility space to optimize local policies from partial observations. Contribution/Results: Evaluated on MAMO Particle and MOMALand benchmarks, our method significantly outperforms state-of-the-art approaches, demonstrating superior effectiveness, robustness to utility heterogeneity, and theoretical rigor in approximating BNE under decentralized execution.
📝 Abstract
Multi-agent multi-objective systems (MAMOS) have emerged as powerful frameworks for modelling complex decision-making problems across various real-world domains, such as robotic exploration, autonomous traffic management, and sensor network optimisation. MAMOS offers enhanced scalability and robustness through decentralised control and more accurately reflects inherent trade-offs between conflicting objectives. In MAMOS, each agent uses utility functions that map return vectors to scalar values. Existing MAMOS optimisation methods face challenges in handling heterogeneous objective and utility function settings, where training non-stationarity is intensified due to private utility functions and the associated policies. In this paper, we first theoretically prove that direct access to, or structured modeling of, global utility functions is necessary for the Bayesian Nash Equilibrium under decentralised execution constraints. To access the global utility functions while preserving the decentralised execution, we propose an Agent-Attention Multi-Agent Multi-Objective Reinforcement Learning (AA-MAMORL) framework. Our approach implicitly learns a joint belief over other agents'utility functions and their associated policies during centralised training, effectively mapping global states and utilities to each agent's policy. In execution, each agent independently selects actions based on local observations and its private utility function to approximate a BNE, without relying on inter-agent communication. We conduct comprehensive experiments in both a custom-designed MAMO Particle environment and the standard MOMALand benchmark. The results demonstrate that access to global preferences and our proposed AA-MAMORL significantly improve performance and consistently outperform state-of-the-art methods.