Trust as Monitoring: Evolutionary Dynamics of User Trust and AI Developer Behaviour

📅 2026-03-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses a critical limitation in existing AI governance models, which treat user trust as a one-time adoption decision and overlook its dynamic evolution through repeated interactions, thereby failing to ensure system safety. The work innovatively conceptualizes trust as users’ reduced monitoring behavior within asymmetric, repeated interactions with developers and, drawing on evolutionary game theory, reveals for the first time the co-evolutionary dynamics between trust and security strategies. Through analyses of infinite-population replicator dynamics, finite-population stochastic evolution, and Q-learning simulations, three long-term equilibria emerge: no adoption with insecurity, widespread adoption yet insecurity, and widespread adoption with security—the latter achievable only when penalties are sufficiently stringent and users can sustain low-cost monitoring. The findings underscore that effective governance requires the joint design of affordable monitoring mechanisms and credible punishment schemes, rather than reliance on regulation alone or uncritical trust.

Technology Category

Application Category

📝 Abstract
AI safety is an increasingly urgent concern as the capabilities and adoption of AI systems grow. Existing evolutionary models of AI governance have primarily examined incentives for safe development and effective regulation, typically representing users' trust as a one-shot adoption choice rather than as a dynamic, evolving process shaped by repeated interactions. We instead model trust as reduced monitoring in a repeated, asymmetric interaction between users and AI developers, where checking AI behaviour is costly. Using evolutionary game theory, we study how user trust strategies and developer choices between safe (compliant) and unsafe (non-compliant) AI co-evolve under different levels of monitoring cost and institutional regimes. We complement the infinite-population replicator analysis with stochastic finite-population dynamics and reinforcement learning (Q-learning) simulations. Across these approaches, we find three robust long-run regimes: no adoption with unsafe development, unsafe but widely adopted systems, and safe systems that are widely adopted. Only the last is desirable, and it arises when penalties for unsafe behaviour exceed the extra cost of safety and users can still afford to monitor at least occasionally. Our results formally support governance proposals that emphasise transparency, low-cost monitoring, and meaningful sanctions, and they show that neither regulation alone nor blind user trust is sufficient to prevent evolutionary drift towards unsafe or low-adoption outcomes.
Problem

Research questions and friction points this paper is trying to address.

trust
AI safety
evolutionary dynamics
monitoring
user-developer interaction
Innovation

Methods, ideas, or system contributions that make the work stand out.

evolutionary game theory
trust as monitoring
AI safety governance
co-evolutionary dynamics
reinforcement learning simulation
🔎 Similar Papers
No similar papers found.
A
Adeela Bashir
School Computing, Engineering and Digital Technologies, Teesside University
Zhao Song
Zhao Song
Teesside University
game theoryreinforcement learningbehavioral science
N
Ndidi Bianca Ogbo
School Computing, Engineering and Digital Technologies, Teesside University
N
Nataliya Balabanova
School of Mathematics, University of Birmingham
Martin Smit
Martin Smit
PhD in Artificial Intelligence, University of Amsterdam
multi-agent systemsreinforcement learning
C
Chin-wing Leung
Department of Computer Science, University of Warwick
P
Paolo Bova
School Computing, Engineering and Digital Technologies, Teesside University
Manuel Chica Serrano
Manuel Chica Serrano
University of Granada
Multiobjective OptimizationAgent Based ModelingEvolutionary Game TheoryData ScienceMarketing
D
Dhanushka Dissanayake
School Computing, Engineering and Digital Technologies, Teesside University
M
Manh Hong Duong
School of Mathematics, University of Birmingham
E
Elias Fernandez Domingos
Machine Learning Group, Université libre de Bruxelles; AI Lab, Vrije Universiteit Brussel
N
Nikita Huber-Kralj
School Computing, Engineering and Digital Technologies, Teesside University
M
Marcus Krellner
University of Technology Dresden
Andrew Powell
Andrew Powell
Principal Advisor, Research Department, Inter American Development Bank
International EconomicsFinanceBanking
S
Stefan Sarkadi
School of Engineering and Physical Sciences, University of Lincoln
Fernando P. Santos
Fernando P. Santos
Informatics Institute (IvI), University of Amsterdam
multiagent systemscomplex systemsevolutionary game theorynetwork sciencealgorithmic fairness
Z
Zia Ush Shamszaman
School Computing, Engineering and Digital Technologies, Teesside University
C
Chaimaa Tarzi
School Computing, Engineering and Digital Technologies, Teesside University
Paolo Turrini
Paolo Turrini
Department of Computer Science, University of Warwick
Artificial IntelligenceGame TheoryReinforcement LearningSocial Networks
G
Grace Ibukunoluwa Ufeoshi
School Computing, Engineering and Digital Technologies, Teesside University
V
Victor A. Vargas-Perez
Department of Computer Science and Artificial Intelligence (DECSAI) and Andalusian Research Institute DaSCI “Data Science and Computational Intelligence”, University of Granada, Spain
Alessandro Di Stefano
Alessandro Di Stefano
Senior Lecturer in Computer Science, Teesside University, SCEDT, UK
Game TheoryNetwork ScienceMachine LearningSocial DynamicsEpidemic Spreading
Simon T. Powers
Simon T. Powers
Division of Computing Science and Mathematics, University of Stirling
Multi-Agent SystemsSocio-Technical SystemsInstitutionsTrustGame Theory
The Anh Han
The Anh Han
Professor of Computer Science, Teesside University
Evolutionary Game TheoryArtificial IntelligenceEvolution of CooperationMulti-agent Systems