LLMs as Workers in Human-Computational Algorithms? Replicating Crowdsourcing Pipelines with LLMs

📅 2023-07-19
🏛️ arXiv.org
📈 Citations: 23
Influential: 1
📄 PDF
🤖 AI Summary
This study investigates the adaptability and task-allocation potential of large language models (LLMs) in complex human-AI collaborative crowdsourcing pipelines, moving beyond atomic tasks to examine end-to-end, multi-stage computational workflows. Method: We propose a “human-AI complementary skill training” framework—the first systematic evaluation of LLM capabilities within real-world crowdsourcing pipelines—employing prompt engineering, multi-stage task decomposition, comparative experiments across interaction modalities (e.g., text-only vs. structured feedback), and cross-analysis of human and LLM performance. Contribution/Results: We find that LLM subtask success rates critically depend on skill type, instruction comprehension fidelity, and interaction design. LLMs robustly perform specific high-fit subtasks (e.g., information extraction, preliminary filtering), yet human oversight remains essential for logical closure, accountability attribution, and dynamic coordination. Crucially, LLM outputs exhibit high sensitivity to minor instruction perturbations, underscoring the necessity of safety-aware, co-design principles for reliable human-AI collaboration.
📝 Abstract
LLMs have shown promise in replicating human-like behavior in crowdsourcing tasks that were previously thought to be exclusive to human abilities. However, current efforts focus mainly on simple atomic tasks. We explore whether LLMs can replicate more complex crowdsourcing pipelines. We find that modern LLMs can simulate some of crowdworkers' abilities in these ``human computation algorithms,'' but the level of success is variable and influenced by requesters' understanding of LLM capabilities, the specific skills required for sub-tasks, and the optimal interaction modality for performing these sub-tasks. We reflect on human and LLMs' different sensitivities to instructions, stress the importance of enabling human-facing safeguards for LLMs, and discuss the potential of training humans and LLMs with complementary skill sets. Crucially, we show that replicating crowdsourcing pipelines offers a valuable platform to investigate 1) the relative LLM strengths on different tasks (by cross-comparing their performances on sub-tasks) and 2) LLMs' potential in complex tasks, where they can complete part of the tasks while leaving others to humans.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Complex Computational Tasks
Division of Labor
Innovation

Methods, ideas, or system contributions that make the work stand out.

Large Language Models
Complex Human Collaborative Tasks
Human-LLM Collaboration
🔎 Similar Papers
No similar papers found.
T
Tongshuang Sherry Wu
Carnegie Mellon University, Pittsburgh, PA, USA
Haiyi Zhu
Haiyi Zhu
Associate Professor, Carnegie Mellon University
Social ComputingHuman-Computer InteractionHuman-AI Interaction
M
Maya Albayrak
Carnegie Mellon University, Pittsburgh, PA, USA
A
Alexis Axon
Carnegie Mellon University, Pittsburgh, PA, USA
Amanda Bertsch
Amanda Bertsch
PhD student, Language Technologies Institute, Carnegie Mellon University
summarizationlong-context NLUconditional generationNLP
W
Wenxing Deng
Carnegie Mellon University, Pittsburgh, PA, USA
Ziqi Ding
Ziqi Ding
UNSW Sydney
CAPTCHAUsabilityCognitive Science
B
B. Guo
Carnegie Mellon University, Pittsburgh, PA, USA
Sireesh Gururaja
Sireesh Gururaja
PhD Student, Carnegie Mellon University
Natural Language ProcessingPretraining DynamicsIncentives that shape NLP
Tzu-Sheng Kuo
Tzu-Sheng Kuo
HCII, Carnegie Mellon University
human-computer interactionresponsible AIsocial computingparticipatory design
J
Jenny T Liang
Carnegie Mellon University, Pittsburgh, PA, USA
Ryan Liu
Ryan Liu
PhD Student in Computer Science, Princeton University
Large Language ModelsComputational Cognitive ScienceNLP Applications
I
Ihita Mandal
Carnegie Mellon University, Pittsburgh, PA, USA
Jeremiah Milbauer
Jeremiah Milbauer
Carnegie Mellon University
natural language processingcomputational social sciencehuman-AI interaction
Xiaolin Ni
Xiaolin Ni
Carnegie Mellon University, Pittsburgh, PA, USA
N
N. Padmanabhan
Carnegie Mellon University, Pittsburgh, PA, USA
S
Subhashini Ramkumar
Carnegie Mellon University, Pittsburgh, PA, USA
A
A. Sudjianto
Carnegie Mellon University, Pittsburgh, PA, USA
Jordan Taylor
Jordan Taylor
Carnegie Mellon University
Critical ComputingHCIMarginalized Communities
Y
Ying-Jui Tseng
Carnegie Mellon University, Pittsburgh, PA, USA
P
Patricia Vaidos
Carnegie Mellon University, Pittsburgh, PA, USA
Z
Zhijin Wu
Carnegie Mellon University, Pittsburgh, PA, USA
W
Wei Wu
Carnegie Mellon University, Pittsburgh, PA, USA
Chenyang Yang
Chenyang Yang
Carnegie Mellon University
Software EngineeringSE4AIHuman-AI Interaction