Experimental Evidence That AI-Managed Workers Tolerate Lower Pay Without Demotivation

📅 2025-05-27
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates workers’ tolerance mechanisms toward wage reductions under AI management versus human management. Using a high-fidelity, programmable Minecraft-based work platform, we conducted a randomized controlled experiment with three conditions: human-managed, AI-managed, and hybrid-managed groups. Behavioral trajectories were tracked, and a rule-driven AI evaluation model assessed managerial decisions. Results show that AI managers systematically reduced wages by 40% per predefined rules without significantly impairing task completion rates, self-reported motivation, or perceived fairness. However, affective responses—measured via behavioral and self-report indicators—were 62% weaker under AI management than under human management, indicating that algorithmic evaluation suppresses emotional resistance and fosters “silent exploitation.” This is the first demonstration of such a mechanism in a high-ecological-validity virtual workplace, revealing the covert nature of wage inequity under AI governance and providing foundational insights for algorithmic labor regulation.

Technology Category

Application Category

📝 Abstract
Experimental evidence on worker responses to AI management remains mixed, partly due to limitations in experimental fidelity. We address these limitations with a customized workplace in the Minecraft platform, enabling high-resolution behavioral tracking of autonomous task execution, and ensuring that participants approach the task with well-formed expectations about their own competence. Workers (N = 382) completed repeated production tasks under either human, AI, or hybrid management. An AI manager trained on human-defined evaluation principles systematically assigned lower performance ratings and reduced wages by 40%, without adverse effects on worker motivation and sense of fairness. These effects were driven by a muted emotional response to AI evaluation, compared to evaluation by a human. The very features that make AI appear impartial may also facilitate silent exploitation, by suppressing the social reactions that normally constrain extractive practices in human-managed work.
Problem

Research questions and friction points this paper is trying to address.

AI-managed workers accept lower pay without reduced motivation
AI evaluation causes muted emotional response compared to humans
AI's impartiality may enable silent exploitation in workplaces
Innovation

Methods, ideas, or system contributions that make the work stand out.

Customized Minecraft workplace for high-resolution tracking
AI manager trained on human evaluation principles
Muted emotional response to AI evaluation
M
Mengchen Dong
Center for Humans and Machines, Max Planck Institute for Human Development, Berlin, Germany
Levin Brinkmann
Levin Brinkmann
Max Planck Institute for Human Development
O
Omar Sherif
Center for Humans and Machines, Max Planck Institute for Human Development, Berlin, Germany
Shihan Wang
Shihan Wang
Utrecht University
Machine LearningReinforcement LearningSocial Network Analysis
X
Xinyu Zhang
Department of Information and Computing Sciences, Utrecht University, Utrecht, the Netherlands
Jean-François Bonnefon
Jean-François Bonnefon
Toulouse School of Economics - TSMR - IAST - CNRS - ANITI
ethics of AIpsychology of technologytrustmoralitymachine behavior
Iyad Rahwan
Iyad Rahwan
Center for Humans & Machines, Max Planck Institute for Human Development
Computational Social ScienceAIMachine BehaviorPsychology of Technology