AMCR: A Framework for Assessing and Mitigating Copyright Risks in Generative Models

📅 2025-08-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Text-to-image generative models often inadvertently reproduce copyrighted content, and existing prompt-filtering approaches fail to address latent copyright infringement. This paper proposes the first generation-process-oriented framework for proactive copyright-risk mitigation: (1) high-risk prompts are semantically reconstructed to avoid potential infringement; (2) an attention-guided cross-modal similarity analysis is introduced to precisely detect implicit copyright associations during image synthesis; and (3) an adaptive suppression module dynamically modulates copyright-sensitive features throughout the diffusion process. The framework preserves visual fidelity while substantially reducing infringement incidence—achieving an average 62.3% reduction in copyright leakage across multiple benchmarks compared to state-of-the-art methods. To our knowledge, this is the first approach enabling fine-grained, end-to-end controllability of copyright risk throughout the entire generative pipeline.

Technology Category

Application Category

📝 Abstract
Generative models have achieved impressive results in text to image tasks, significantly advancing visual content creation. However, this progress comes at a cost, as such models rely heavily on large-scale training data and may unintentionally replicate copyrighted elements, creating serious legal and ethical challenges for real-world deployment. To address these concerns, researchers have proposed various strategies to mitigate copyright risks, most of which are prompt based methods that filter or rewrite user inputs to prevent explicit infringement. While effective in handling obvious cases, these approaches often fall short in more subtle situations, where seemingly benign prompts can still lead to infringing outputs. To address these limitations, this paper introduces Assessing and Mitigating Copyright Risks (AMCR), a comprehensive framework which i) builds upon prompt-based strategies by systematically restructuring risky prompts into safe and non-sensitive forms, ii) detects partial infringements through attention-based similarity analysis, and iii) adaptively mitigates risks during generation to reduce copyright violations without compromising image quality. Extensive experiments validate the effectiveness of AMCR in revealing and mitigating latent copyright risks, offering practical insights and benchmarks for the safer deployment of generative models.
Problem

Research questions and friction points this paper is trying to address.

Mitigating copyright risks in generative models
Detecting partial infringements through similarity analysis
Reducing copyright violations without compromising image quality
Innovation

Methods, ideas, or system contributions that make the work stand out.

Systematically restructures risky prompts into safe forms
Detects partial infringements via attention-based similarity analysis
Adaptively mitigates risks during generation without quality loss
🔎 Similar Papers
No similar papers found.
Zhipeng Yin
Zhipeng Yin
Florida International University
Trustworthy AIAlgorithmic fairnessCopyrightMachine learning
Zichong Wang
Zichong Wang
Florida International University
Trustworthy MLCausal InferenceGraph MiningAlgorithmic Fairness
Avash Palikhe
Avash Palikhe
Florida International University
Algorithmic FairnessExplainabilityEthical AI
Z
Zhen Liu
Guangdong University of Foreign Studies, Guangzhou, Guangdong, China
J
Jun Liu
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
W
Wenbin Zhang
Florida International University, Miami, Florida, United States