๐ค AI Summary
Existing generative model watermarking methods lack a theoretical characterization of the trade-off between detection capability and generation fidelity.
Method: We formulate watermark embedding as a statistical hypothesis test between the original and watermarked distributions, and leverage $f$-divergence to derive, for the first time, a tight lower bound on fidelity loss under prescribed Type-I and Type-II error constraints. We further integrate hypothesis testing theory, information divergence analysis, and optimal sampling strategies to design a practical watermark embedding mechanism that minimizes fidelity degradation.
Contribution/Results: Our work establishes a rigorous theoretical link between detection performance and generation quality. It yields the explicit form of the optimal watermarked distribution achieving the bound and provides a general design principle for high-fidelity watermarking with controllable error ratesโenabling principled, provably optimal watermark embedding in generative models.
๐ Abstract
Watermarking has recently emerged as a crucial tool for protecting the intellectual property of generative models and for distinguishing AI-generated content from human-generated data. Despite its practical success, most existing watermarking schemes are empirically driven and lack a theoretical understanding of the fundamental trade-off between detection power and generation fidelity. To address this gap, we formulate watermarking as a statistical hypothesis testing problem between a null distribution and its watermarked counterpart. Under explicit constraints on false-positive and false-negative rates, we derive a tight lower bound on the achievable fidelity loss, measured by a general $f$-divergence, and characterize the optimal watermarked distribution that attains this bound. We further develop a corresponding sampling rule that provides an optimal mechanism for inserting watermarks with minimal fidelity distortion. Our result establishes a simple yet broadly applicable principle linking hypothesis testing, information divergence, and watermark generation.