๐ค AI Summary
In anonymous multi-matrix games, classical fictitious play (FP) suffers from inefficient exploration and poor scalability due to the exponential growth of the joint action space with the number of agents. To address this, we propose aggregated fictitious play (agg-FP), which abandons individual strategy modeling and instead tracks the global frequency of each actionโs selection, leveraging anonymity to compress the strategy space. Theoretically, agg-FP retains the same Nash equilibrium convergence guarantees as standard FP while reducing state-space complexity from exponential to polynomial in the number of agents and actions. Empirical evaluations demonstrate significantly accelerated convergence; notably, agg-FP is the first algorithm to achieve scalable, distributed, and theoretically grounded equilibrium learning in anonymous multi-matrix games.
๐ Abstract
Fictitious play (FP) is a well-studied algorithm that enables agents to learn Nash equilibrium in games with certain reward structures. However, when agents have no prior knowledge of the reward functions, FP faces a major challenge: the joint action space grows exponentially with the number of agents, which slows down reward exploration. Anonymous games offer a structure that mitigates this issue. In these games, the rewards depend only on the actions taken; not on who is taking which action. Under such a structure, we introduce aggregate fictitious play (agg-FP), a variant of FP where each agent tracks the frequency of the number of other agents playing each action, rather than these agents' individual actions. We show that in anonymous polymatrix games, agg-FP converges to a Nash equilibrium under the same conditions as classical FP. In essence, by aggregating the agents' actions, we reduce the action space without losing the convergence guarantees. Using simulations, we provide empirical evidence on how this reduction accelerates convergence.