π€ AI Summary
Existing large-area visible watermark removal methods suffer from poor background restoration and excessive reliance on high-precision watermark masks. To address these issues, this paper proposes a feature-adaptive framework built upon a pre-trained image inpainting model. Methodologically, we design a dual-branch residual background feature encoder coupled with a gated fusion module to decouple watermark and background feature representations; additionally, we introduce a coarse-grain watermark mask-guided weakly supervised training paradigm, substantially reducing dependence on accurate masks during inference. Extensive experiments on both synthetic and real-world datasets demonstrate that our approach consistently outperforms state-of-the-art methods, achieving superior and more robust performance in terms of watermark removal completeness and background reconstruction fidelity. The proposed framework establishes a new paradigm for large-area watermark removal without requiring precise mask guidance.
π Abstract
Visible watermark removal which involves watermark cleaning and background content restoration is pivotal to evaluate the resilience of watermarks. Existing deep neural network (DNN)-based models still struggle with large-area watermarks and are overly dependent on the quality of watermark mask prediction. To overcome these challenges, we introduce a novel feature adapting framework that leverages the representation modeling capacity of a pre-trained image inpainting model. Our approach bridges the knowledge gap between image inpainting and watermark removal by fusing information of the residual background content beneath watermarks into the inpainting backbone model. We establish a dual-branch system to capture and embed features from the residual background content, which are merged into intermediate features of the inpainting backbone model via gated feature fusion modules. Moreover, for relieving the dependence on high-quality watermark masks, we introduce a new training paradigm by utilizing coarse watermark masks to guide the inference process. This contributes to a visible image removal model which is insensitive to the quality of watermark mask during testing. Extensive experiments on both a large-scale synthesized dataset and a real-world dataset demonstrate that our approach significantly outperforms existing state-of-the-art methods. The source code is available in the supplementary materials.