🤖 AI Summary
This paper addresses the fundamental tension among legal compliance, accountability attribution, and technical explainability in human-in-the-loop (HITL) systems, exposing structural gaps in current UK and EU regulatory frameworks for responsibility assignment and the attendant “human scapegoating” risk. Methodologically, it innovatively integrates computability theory—specifically oracle machines, Turing and many-one reductions, and total functions—to formally model three HITL interaction paradigms: monitoring, single-point intervention, and deep interaction. It establishes a law–technology co-assessment framework and proposes a dynamic accountability allocation principle grounded in AI systems’ technical capability boundaries. Furthermore, it systematically constructs a taxonomy of HITL failure modes. The contributions provide a rigorous theoretical foundation and actionable design guidelines for HITL system development, regulatory refinement, and accountability mechanism reform.
📝 Abstract
The legal compliance and safety of different Human-in-the-loop (HITL) setups for AI can vary greatly. This manuscript aims to identify new ways of choosing between such setups, and shows that there is an unavoidable trade-off between the attribution of legal responsibility and the technical explainability of AI. We begin by using the notion of oracle machines from computability theory to formalise different HITL setups, distinguishing between trivial human monitoring, single endpoint human action, and highly involved interaction between the human(s) and the AI. These correspond to total functions, many-one reductions, and Turing reductions respectively. A taxonomy categorising HITL failure modes is then presented, highlighting the limitations on what any HITL setup can actually achieve. Our approach then identifies oversights from UK and EU legal frameworks, which focus on certain HITL setups which may not always achieve the desired ethical, legal, and sociotechnical outcomes. We suggest areas where the law should recognise the effectiveness of different HITL setups and assign responsibility in these contexts, avoiding unnecessary and unproductive human"scapegoating". Overall, we show how HITL setups involve many technical design decisions, and can be prone to failures which are often out of the humans' control. This opens up a new analytic perspective on the challenges arising in the creation of HITL setups, helping inform AI developers and lawmakers on designing HITL to better achieve their desired outcomes.