HCAST: Human-Calibrated Autonomy Software Tasks

📅 2025-03-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Quantifying the societal trustworthiness of highly autonomous AI systems remains an open challenge, with no established framework for assessing their real-world reliability and human-aligned performance. Method: This paper introduces “X-hour trustworthiness”—a novel paradigm centered on human completion time (X hours)—and proposes HCAST, the first benchmark for engineering and reasoning tasks comprising 189 realistic problems. HCAST establishes precise human baselines via 563 expert annotations (exceeding 1,500 human hours), covering task durations from 1 minute to over 8 hours. Evaluation employs standardized execution protocols, human–AI isomorphic environments, and multi-domain expert baselines. Contribution/Results: HCAST enables direct mapping between AI capability and human-trustable task duration. Experiments reveal that state-of-the-art AI achieves 70–80% success on tasks requiring <1 hour for humans, but drops below 20% on tasks requiring >4 hours—clearly delineating current capability boundaries and providing a foundational metric for sociotechnical AI trust assessment.

Technology Category

Application Category

📝 Abstract
To understand and predict the societal impacts of highly autonomous AI systems, we need benchmarks with grounding, i.e., metrics that directly connect AI performance to real-world effects we care about. We present HCAST (Human-Calibrated Autonomy Software Tasks), a benchmark of 189 machine learning engineering, cybersecurity, software engineering, and general reasoning tasks. We collect 563 human baselines (totaling over 1500 hours) from people skilled in these domains, working under identical conditions as AI agents, which lets us estimate that HCAST tasks take humans between one minute and 8+ hours. Measuring the time tasks take for humans provides an intuitive metric for evaluating AI capabilities, helping answer the question"can an agent be trusted to complete a task that would take a human X hours?"We evaluate the success rates of AI agents built on frontier foundation models, and we find that current agents succeed 70-80% of the time on tasks that take humans less than one hour, and less than 20% of the time on tasks that take humans more than 4 hours.
Problem

Research questions and friction points this paper is trying to address.

Develop benchmarks linking AI performance to real-world societal impacts
Measure human task completion times to calibrate AI trustworthiness
Evaluate AI success rates on tasks of varying human difficulty levels
Innovation

Methods, ideas, or system contributions that make the work stand out.

Human-calibrated benchmark for AI tasks
563 human baselines under identical conditions
Time-based metric for AI trust evaluation
🔎 Similar Papers
No similar papers found.
David Rein
David Rein
Member of Technical Staff, METR
machine learning
Joel Becker
Joel Becker
Member of Technical Staff, METR; CEO, Qally's
Machine Learning
A
Amy Deng
Model Evaluation & Threat Research (METR)
S
Seraphina Nix
Model Evaluation & Threat Research (METR)
C
Chris Canal
Equistamp
D
Daniel O'Connel
Equistamp
P
Pip Arnott
Independent Contributor
R
Ryan Bloom
Model Evaluation & Threat Research (METR)
T
Thomas Broadley
Model Evaluation & Threat Research (METR)
K
Katharyn Garcia
Model Evaluation & Threat Research (METR)
B
Brian Goodrich
Model Evaluation & Threat Research (METR)
M
Max Hasin
Model Evaluation & Threat Research (METR)
S
Sami Jawhar
Model Evaluation & Threat Research (METR)
M
Megan Kinniment
Model Evaluation & Threat Research (METR)
Thomas Kwa
Thomas Kwa
Unknown affiliation
AI safetymechanistic interpretability
A
Aron Lajko
Independent Contributor
N
Nate Rush
Model Evaluation & Threat Research (METR)
L
L. Sato
Model Evaluation & Threat Research (METR)
S
Sydney von Arx
Model Evaluation & Threat Research (METR)
B
Ben West
Model Evaluation & Threat Research (METR)
L
Lawrence Chan
Model Evaluation & Threat Research (METR)
Elizabeth Barnes
Elizabeth Barnes
Model Evaluation & Threat Research (METR)