Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Secure and Trustworthy Large Language Models

DUAW: Data-free Universal Adversarial Watermark against Stable Diffusion Customization

Xiaoyu Ye · Hao Huang · Jiaqi An · Yongtao Wang


Abstract:

Stable Diffusion (SD) customization approaches enable users to personalize SD model outputs, greatly enhancing the flexibility and diversity of AI art. However, they also allow individuals to plagiarize specific styles or subjects from copyrighted images, which raises significant concerns about potential copyright infringement. To address this issue, we propose an invisible data-free universal adversarial watermark (DUAW), aiming to protect copyrighted images from different customization approaches across various versions of SD models. First, DUAW is designed to disrupt the variational autoencoder during SD customization. Second, DUAW is trained on synthetic images produced by a Large Language Model (LLM) and a pretrained SD model, that is, it is generated in a data-free manner without the use of any copyrighted images. Once crafted, DUAW can be imperceptibly integrated into any copyrighted image, serving as a protective measure by inducing significant distortions in the images generated by customized SD models. Experimental results demonstrate that DUAW can distort the outputs of fine-tuned SD models, making them discernible to both human observers and a simple classifier, yielding more effective protection results than existing methods.

Chat is not available.