🤖 AI Summary
This study addresses the challenge posed by Article 50(2) of the EU Artificial Intelligence Act, which mandates that AI-generated content bear dual transparency markers—both human-readable and machine-verifiable—a requirement current generative AI architectures struggle to fulfill. Focusing on two representative scenarios, synthetic data generation and automated fact-checking, the work identifies three structural gaps between regulatory demands and AI system capabilities: the absence of cross-platform marking standards, misalignment in reliability benchmarks, and insufficient adaptation to user heterogeneity. Through methods including large language model output tracing, watermark robustness analysis, and marker format evaluation, the paper advocates embedding transparency as a core design principle. It proposes an integrative framework harmonizing legal semantics, AI engineering, and human-centered design to chart viable theoretical and practical pathways toward regulatory compliance and co-evolution of AI systems.
📝 Abstract
Art. 50 II of the EU Artificial Intelligence Act mandates dual transparency for AI-generated content: outputs must be labeled in both human-understandable and machine-readable form for automated verification. This requirement, entering into force in August 2026, collides with fundamental constraints of current generative AI systems. Using synthetic data generation and automated fact-checking as diagnostic use cases, we show that compliance cannot be reduced to post-hoc labeling. In fact-checking pipelines, provenance tracking is not feasible under iterative editorial workflows and non-deterministic LLM outputs; moreover, the assistive-function exemption does not apply, as such systems actively assign truth values rather than supporting editorial presentation. In synthetic data generation, persistent dual-mode marking is paradoxical: watermarks surviving human inspection risk being learned as spurious features during training, while marks suited for machine verification are fragile under standard data processing. Across both domains, three structural gaps obstruct compliance: (a) absent cross-platform marking formats for interleaved human-AI outputs; (b) misalignment between the regulation's 'reliability' criterion and probabilistic model behavior; and (c) missing guidance for adapting disclosures to heterogeneous user expertise. Closing these gaps requires transparency to be treated as an architectural design requirement, demanding interdisciplinary research across legal semantics, AI engineering, and human-centered desi