CertDW: Towards Certified Dataset Ownership Verification via Conformal Prediction

📅 2025-06-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing dataset ownership verification methods rely on trusted verification assumptions and are vulnerable to perturbations or adaptive attacks. This paper proposes CertDW, the first provably robust certified watermarking framework for dataset ownership authentication. CertDW pioneers the integration of conformal prediction into dataset watermarking, introducing two key statistics: Principal Probability (PP) and Watermark Robustness (WR). It establishes a provable lower bound on WR in terms of PP, enabling pixel-level perturbation-robust certification. The framework comprises two phases—watermark embedding and statistical verification—and jointly incorporates hypothesis testing, prediction stability analysis, and adversarial modeling. Evaluated on benchmarks including ImageNet, CertDW achieves >98% authentication accuracy and demonstrates strong robustness against pruning, fine-tuning, and adversarial distillation. Reliable certification is achieved when WR significantly exceeds the PP threshold of benign models.

Technology Category

Application Category

📝 Abstract
Deep neural networks (DNNs) rely heavily on high-quality open-source datasets (e.g., ImageNet) for their success, making dataset ownership verification (DOV) crucial for protecting public dataset copyrights. In this paper, we find existing DOV methods (implicitly) assume that the verification process is faithful, where the suspicious model will directly verify ownership by using the verification samples as input and returning their results. However, this assumption may not necessarily hold in practice and their performance may degrade sharply when subjected to intentional or unintentional perturbations. To address this limitation, we propose the first certified dataset watermark (i.e., CertDW) and CertDW-based certified dataset ownership verification method that ensures reliable verification even under malicious attacks, under certain conditions (e.g., constrained pixel-level perturbation). Specifically, inspired by conformal prediction, we introduce two statistical measures, including principal probability (PP) and watermark robustness (WR), to assess model prediction stability on benign and watermarked samples under noise perturbations. We prove there exists a provable lower bound between PP and WR, enabling ownership verification when a suspicious model's WR value significantly exceeds the PP values of multiple benign models trained on watermark-free datasets. If the number of PP values smaller than WR exceeds a threshold, the suspicious model is regarded as having been trained on the protected dataset. Extensive experiments on benchmark datasets verify the effectiveness of our CertDW method and its resistance to potential adaptive attacks. Our codes are at href{https://github.com/NcepuQiaoTing/CertDW}{GitHub}.
Problem

Research questions and friction points this paper is trying to address.

Ensuring reliable dataset ownership verification under attacks
Certifying watermark robustness against malicious perturbations
Providing statistical measures for model prediction stability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Certified dataset watermark for ownership verification
Conformal prediction for statistical measures
Provable lower bound between PP and WR
🔎 Similar Papers
No similar papers found.
T
Ting Qiao
School of Control and Computer Engineering, North China Electric Power University, Beijing,102206, China
Y
Yiming Li
College of Computing and Data Science, Nanyang Technological University, Singapore, 639798, Singapore
J
Jianbin Li
School of Control and Computer Engineering, North China Electric Power University, Beijing,102206, China
Y
Yingjia Wang
School of Control and Computer Engineering, North China Electric Power University, Beijing,102206, China
L
Leyi Qi
School of Cybersecurity, Northwestern Polytechnical University, Xi’an, 710072, China
Junfeng Guo
Junfeng Guo
University of Maryland, College Park
Trustworthy Machine LearningComputer VisionNature Language Processing
R
Ruili Feng
Alibaba Group, Hangzhou, 311100, China
Dacheng Tao
Dacheng Tao
Nanyang Technological University
artificial intelligencemachine learningcomputer visionimage processingdata mining