Understanding LLM Agent Behaviours via Game Theory: Strategy Recognition, Biases and Multi-Agent Dynamics

๐Ÿ“… 2025-12-08
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This study investigates strategic intent recognition in large language models (LLMs) within multi-agent interactions. To address this, we extend the FAIRGAME framework by designing payoff-scaled Prisonerโ€™s Dilemma and dynamic multi-agent Public Goods Game environments. Leveraging strategy classification models and behavioral trajectory analysis, we systematically evaluate LLMsโ€™ cooperation versus defection tendencies in repeated social dilemmas. Our key findings reveal that linguistic formulation exerts influence on strategic choice comparable to architectural differences across models; LLMs exhibit pronounced incentive sensitivity, cross-lingual behavioral variation, endgame defection bias, and a systematic cooperation bias. These results provide empirical grounding and methodological tools for AI safety governance, design of multi-agent coordination mechanisms, and development of AI-driven social infrastructure.

Technology Category

Application Category

๐Ÿ“ Abstract
As Large Language Models (LLMs) increasingly operate as autonomous decision-makers in interactive and multi-agent systems and human societies, understanding their strategic behaviour has profound implications for safety, coordination, and the design of AI-driven social and economic infrastructures. Assessing such behaviour requires methods that capture not only what LLMs output, but the underlying intentions that guide their decisions. In this work, we extend the FAIRGAME framework to systematically evaluate LLM behaviour in repeated social dilemmas through two complementary advances: a payoff-scaled Prisoners Dilemma isolating sensitivity to incentive magnitude, and an integrated multi-agent Public Goods Game with dynamic payoffs and multi-agent histories. These environments reveal consistent behavioural signatures across models and languages, including incentive-sensitive cooperation, cross-linguistic divergence and end-game alignment toward defection. To interpret these patterns, we train traditional supervised classification models on canonical repeated-game strategies and apply them to FAIRGAME trajectories, showing that LLMs exhibit systematic, model- and language-dependent behavioural intentions, with linguistic framing at times exerting effects as strong as architectural differences. Together, these findings provide a unified methodological foundation for auditing LLMs as strategic agents and reveal systematic cooperation biases with direct implications for AI governance, collective decision-making, and the design of safe multi-agent systems.
Problem

Research questions and friction points this paper is trying to address.

Analyze LLM strategic behavior in repeated social dilemmas
Identify systematic cooperation biases across models and languages
Provide a foundation for auditing LLMs as strategic agents
Innovation

Methods, ideas, or system contributions that make the work stand out.

Extending FAIRGAME framework for LLM evaluation
Using payoff-scaled Prisoners Dilemma and Public Goods Game
Training classifiers to interpret LLM strategic intentions
๐Ÿ”Ž Similar Papers
No similar papers found.
T
Trung-Kiet Huynh
Faculty of Information and Technology, Ho Chi Minh City University of Science (HCMUS), Vietnam
D
Duy-Minh Dao-Sy
Faculty of Information and Technology, Ho Chi Minh City University of Science (HCMUS), Vietnam
T
Thanh-Bang Cao
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
P
Phong-Hao Le
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
H
Hong-Dan Nguyen
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
P
Phu-Quy Nguyen-Lam
Faculty of Information and Technology, Ho Chi Minh City University of Science (HCMUS), Vietnam
M
Minh-Luan Nguyen-Vo
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
H
Hong-Phat Pham
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
P
Phu-Hoa Pham
Faculty of Information and Technology, Ho Chi Minh City University of Science (HCMUS), Vietnam
T
Thien-Kim Than
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
C
Chi-Nguyen Tran
Faculty of Information and Technology, Ho Chi Minh City University of Science (HCMUS), Vietnam
H
Huy Tran
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
G
Gia-Thoai Tran-Le
Faculty of Computer Science and Engineering, Ho Chi Minh City University of Technology (HCMUT), Vietnam
Alessio Buscemi
Alessio Buscemi
Luxembourg Institute of Science and Technology
Large Language ModelsAIMachine LearningAutomotive networks
Le Hong Trang
Le Hong Trang
Ho Chi Minh City University of Technology
OptimizationStable MatchingData MiningApplied Machine Learning
The Anh Han
The Anh Han
Professor of Computer Science, Teesside University
Evolutionary Game TheoryArtificial IntelligenceEvolution of CooperationMulti-agent Systems