🤖 AI Summary
To address security risks—specifically, data poisoning or adversarial manipulation of large language models (LLMs) leading to spurious outputs in space-domain AI applications—this paper proposes the first LLM output authenticity verification framework tailored to aerospace mission scenarios. Methodologically, it integrates textual semantic features, statistical anomaly patterns, and adversarial robustness modeling to construct a supervised binary classifier capable of distinguishing legitimate outputs from manipulated ones with high accuracy. Key contributions include: (1) formal definition and modeling of space-mission-specific textual credibility criteria; (2) open-sourcing a benchmark dataset and evaluation protocol covering diverse aerospace text genres; and (3) empirical validation on a Kaggle competition, delivering a reusable, end-to-end text forensics toolkit. The framework significantly enhances the poisoning resilience and operational robustness of safety-critical AI systems in space applications.
📝 Abstract
The "Fake or Real" competition hosted on Kaggle (href{https://www.kaggle.com/competitions/fake-or-real-the-impostor-hunt}{https://www.kaggle.com/competitions/fake-or-real-the-impostor-hunt}) is the second part of a series of follow-up competitions and hackathons related to the "Assurance for Space Domain AI Applications" project funded by the European Space Agency (href{https://assurance-ai.space-codev.org/}{https://assurance-ai.space-codev.org/}). The competition idea is based on two real-life AI security threats identified within the project -- data poisoning and overreliance in Large Language Models. The task is to distinguish between the proper output from LLM and the output generated under malicious modification of the LLM. As this problem was not extensively researched, participants are required to develop new techniques to address this issue or adjust already existing ones to this problem's statement.