D2RA: Dual Domain Regeneration Attack

📅 2025-10-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing semantic watermarking schemes for generative models remain vulnerable under adversarial conditions—particularly against resource-constrained, model-agnostic attacks with zero access to the watermarked model. To address this, we propose a training-free, single-image attack method that jointly optimizes representations in both latent and frequency domains. Our approach projects watermarked images onto natural image priors while simultaneously suppressing watermark signals and preserving visual fidelity. This constitutes the first dual-domain joint optimization framework compatible with diverse watermark encoding schemes, enabling efficient watermark removal under zero-query and low-compute constraints. Experiments demonstrate that our method significantly reduces watermark detectability (average reduction >70%) and achieves superior PSNR and SSIM scores compared to state-of-the-art attacks. These results expose a fundamental security deficiency in current semantic watermarking mechanisms.

Technology Category

Application Category

📝 Abstract
The growing use of generative models has intensified the need for watermarking methods that ensure content attribution and provenance. While recent semantic watermarking schemes improve robustness by embedding signals in latent or frequency representations, we show they remain vulnerable even under resource-constrained adversarial settings. We present D2RA, a training-free, single-image attack that removes or weakens watermarks without access to the underlying model. By projecting watermarked images onto natural priors across complementary representations, D2RA suppresses watermark signals while preserving visual fidelity. Experiments across diverse watermarking schemes demonstrate that our approach consistently reduces watermark detectability, revealing fundamental weaknesses in current designs. Our code is available at https://github.com/Pragati-Meshram/DAWN.
Problem

Research questions and friction points this paper is trying to address.

Exposing vulnerabilities in semantic watermarking schemes for generative models
Removing watermarks without model access while preserving visual quality
Demonstrating fundamental weaknesses in current watermarking method designs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Training-free single-image attack removes watermarks
Projects images onto natural priors across representations
Suppresses watermark signals while preserving visual fidelity
🔎 Similar Papers
No similar papers found.
P
Pragati Shuddhodhan Meshram
University of Illinois Urbana-Champaign
Varun Chandrasekaran
Varun Chandrasekaran
University of Illinois Urbana-Champaign
SecurityPrivacyArtificial Intelligence