Algorithmic Prompt Generation for Diverse Human-like Teaming and Communication with Large Language Models

📅 2025-04-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Scaling human team behavioral data for human-AI collaborative decision-making remains challenging due to the difficulty of collecting diverse, high-quality human interaction traces. Method: This paper introduces the first algorithmic prompt-generation framework that integrates Quality-Diversity (QD) optimization with large language model (LLM) agents—requiring neither handcrafted prompts nor large-scale user studies. It automatically discovers prompt strategies that elicit multidimensional, human-like communication and coordination behaviors from LLMs in multi-step collaborative settings, synthesizing a broad spectrum of team behavioral patterns. Contribution/Results: Evaluated in a 54-participant user study, the generated behaviors accurately reproduce key human collaboration trends and uncover novel coordination patterns otherwise obscured by data sparsity. The approach significantly outperforms baseline methods in both behavioral diversity and fidelity to human behavior.

Technology Category

Application Category

📝 Abstract
Understanding how humans collaborate and communicate in teams is essential for improving human-agent teaming and AI-assisted decision-making. However, relying solely on data from large-scale user studies is impractical due to logistical, ethical, and practical constraints, necessitating synthetic models of multiple diverse human behaviors. Recently, agents powered by Large Language Models (LLMs) have been shown to emulate human-like behavior in social settings. But, obtaining a large set of diverse behaviors requires manual effort in the form of designing prompts. On the other hand, Quality Diversity (QD) optimization has been shown to be capable of generating diverse Reinforcement Learning (RL) agent behavior. In this work, we combine QD optimization with LLM-powered agents to iteratively search for prompts that generate diverse team behavior in a long-horizon, multi-step collaborative environment. We first show, through a human-subjects experiment (n=54 participants), that humans exhibit diverse coordination and communication behavior in this domain. We then show that our approach can effectively replicate trends from human teaming data and also capture behaviors that are not easily observed without collecting large amounts of data. Our findings highlight the combination of QD and LLM-powered agents as an effective tool for studying teaming and communication strategies in multi-agent collaboration.
Problem

Research questions and friction points this paper is trying to address.

Generating diverse human-like teaming behaviors using LLMs
Overcoming constraints of manual prompt design for diversity
Replicating human communication trends in multi-agent collaboration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines QD optimization with LLM-powered agents
Generates diverse prompts for human-like teaming
Replicates human trends with synthetic behaviors
🔎 Similar Papers
No similar papers found.
Siddharth Srikanth
Siddharth Srikanth
University of Southern California
RoboticsOpen Ended LearningReinforcement Learning
Varun Bhatt
Varun Bhatt
PhD Student, University of Southern California
B
Boshen Zhang
Thomas Lord Department of Computer Science, University of Southern California
W
Werner Hager
School of Computing and Information, University of Pittsburgh
C
Charles Michael Lewis
School of Computing and Information, University of Pittsburgh
K
Katia P. Sycara
Robotics Institute, Carnegie Mellon University
Aaquib Tabrez
Aaquib Tabrez
Postdoctoral Associate, Cornell University
Explainable AIHuman-Robot InteractionReinforcement LearningRoboticsAugmented Reality
Stefanos Nikolaidis
Stefanos Nikolaidis
Associate Professor of Computer Science, University of Southern California
roboticsartificial intelligencemachine learning