π€ AI Summary
This work addresses the challenge of inefficient exploration in sparse-reward, highly cooperative open-domain multi-agent tasks, where independent exploration by individual agents often leads to redundancy and hinders the discovery of effective collaborative strategies. To overcome this, the paper introduces Counterfactual Conditional Likelihood (CCL), a novel reward mechanism that incorporates counterfactual reasoning into multi-agent exploration for the first time. CCL quantifies each agentβs unique contribution to the teamβs joint information gain, thereby guiding non-redundant exploration. Integrated within a multi-agent reinforcement learning framework suitable for continuous action spaces, the proposed method significantly accelerates learning under sparse team rewards and achieves superior performance on tasks requiring tight coordination among agents.
π Abstract
Efficient exploration is critical for multiagent systems to discover coordinated strategies, particularly in open-ended domains such as search and rescue or planetary surveying. However, when exploration is encouraged only at the individual agent level, it often leads to redundancy, as agents act without awareness of how their teammates are exploring. In this work, we introduce Counterfactual Conditional Likelihood (CCL) rewards, which score each agent's exploration by isolating its unique contribution to team exploration. Unlike prior methods that reward agents solely for the novelty of their individual observations, CCL emphasizes observations that are informative with respect to the joint exploration of the team. Experiments in continuous multiagent domains show that CCL rewards accelerate learning for domains with sparse team rewards, where most joint actions yield zero rewards, and are particularly effective in tasks that require tight coordination among agents.