DMark: Order-Agnostic Watermarking for Diffusion Large Language Models

📅 2025-10-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional causal watermarks fail for diffusion-based large language models (dLLMs) due to their non-sequential, non-autoregressive decoding. To address this, we propose DMark—the first watermarking framework specifically designed for dLLMs. Instead of relying on sequential causality, DMark introduces predictive watermarking, bidirectional watermarking, and their joint strategy, leveraging the intrinsic dependencies between the forward and reverse processes of diffusion models to embed and detect robust watermark signals. Experiments across multiple dLLMs show that DMark achieves watermark detection rates of 92.0%–99.5% at a 1% false positive rate—substantially outperforming prior methods (49.6%–71.2%)—while preserving text quality and demonstrating resilience against textual tampering.

Technology Category

Application Category

📝 Abstract
Diffusion large language models (dLLMs) offer faster generation than autoregressive models while maintaining comparable quality, but existing watermarking methods fail on them due to their non-sequential decoding. Unlike autoregressive models that generate tokens left-to-right, dLLMs can finalize tokens in arbitrary order, breaking the causal design underlying traditional watermarks. We present DMark, the first watermarking framework designed specifically for dLLMs. DMark introduces three complementary strategies to restore watermark detectability: predictive watermarking uses model-predicted tokens when actual context is unavailable; bidirectional watermarking exploits both forward and backward dependencies unique to diffusion decoding; and predictive-bidirectional watermarking combines both approaches to maximize detection strength. Experiments across multiple dLLMs show that DMark achieves 92.0-99.5% detection rates at 1% false positive rate while maintaining text quality, compared to only 49.6-71.2% for naive adaptations of existing methods. DMark also demonstrates robustness against text manipulations, establishing that effective watermarking is feasible for non-autoregressive language models.
Problem

Research questions and friction points this paper is trying to address.

Develops watermarking for diffusion models with non-sequential decoding
Restores detectability via predictive and bidirectional dependency strategies
Ensures robustness against text manipulations while preserving quality
Innovation

Methods, ideas, or system contributions that make the work stand out.

Predictive watermarking uses model-predicted tokens for context
Bidirectional watermarking exploits forward and backward dependencies
Predictive-bidirectional combines both approaches to maximize detection
🔎 Similar Papers
No similar papers found.