🤖 AI Summary
Small UAV swarms deployed in open-world missions—such as post-disaster search and rescue—suffer from limited unknown-object recognition and semantic reasoning capabilities of conventional vision systems; direct integration of large language models (LLMs) risks hallucination, compromising decision safety. To address this, we propose a “Cognitive Guardrail” mechanism: the first framework to tightly integrate LLM-based semantic reasoning with formal safety constraints, task-relevance verification, and uncertainty quantification—yielding a verifiable, controllable high-level autonomous decision-making architecture. Our approach synergistically combines computer vision, LLMs, multi-agent coordination, and probabilistic uncertainty modeling. Evaluated in both simulation and real-world deployments, it reduces mission failure rate by 76% and achieves over 92% accuracy in critical target response, significantly enhancing trustworthiness of semantic understanding and adaptive decision-making for UAV swarms in dynamic, open environments.
📝 Abstract
Small Uncrewed Aerial Systems (sUAS) are increasingly deployed as autonomous swarms in search-and-rescue and other disaster-response scenarios. In these settings, they use computer vision (CV) to detect objects of interest and autonomously adapt their missions. However, traditional CV systems often struggle to recognize unfamiliar objects in open-world environments or to infer their relevance for mission planning. To address this, we incorporate large language models (LLMs) to reason about detected objects and their implications. While LLMs can offer valuable insights, they are also prone to hallucinations and may produce incorrect, misleading, or unsafe recommendations. To ensure safe and sensible decision-making under uncertainty, high-level decisions must be governed by cognitive guardrails. This article presents the design, simulation, and real-world integration of these guardrails for sUAS swarms in search-and-rescue missions.