Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Measuring the Success of Diffusion Models at Imitating Human Artists (2307.04028v1)

Published 8 Jul 2023 in cs.CV, cs.AI, and cs.LG

Abstract: Modern diffusion models have set the state-of-the-art in AI image generation. Their success is due, in part, to training on Internet-scale data which often includes copyrighted work. This prompts questions about the extent to which these models learn from, imitate, or copy the work of human artists. This work suggests that tying copyright liability to the capabilities of the model may be useful given the evolving ecosystem of generative models. Specifically, much of the legal analysis of copyright and generative systems focuses on the use of protected data for training. As a result, the connections between data, training, and the system are often obscured. In our approach, we consider simple image classification techniques to measure a model's ability to imitate specific artists. Specifically, we use Contrastive Language-Image Pretrained (CLIP) encoders to classify images in a zero-shot fashion. Our process first prompts a model to imitate a specific artist. Then, we test whether CLIP can be used to reclassify the artist (or the artist's work) from the imitation. If these tests match the imitation back to the original artist, this suggests the model can imitate that artist's expression. Our approach is simple and quantitative. Furthermore, it uses standard techniques and does not require additional training. We demonstrate our approach with an audit of Stable Diffusion's capacity to imitate 70 professional digital artists with copyrighted work online. When Stable Diffusion is prompted to imitate an artist from this set, we find that the artist can be identified from the imitation with an average accuracy of 81.0%. Finally, we also show that a sample of the artist's work can be matched to these imitation images with a high degree of statistical reliability. Overall, these results suggest that Stable Diffusion is broadly successful at imitating individual human artists.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. Judging facts, judging norms: Training machine learning models to judge humans requires a modified approach to labeling data. Science Advances, 9(19):eabq0701, 2023.
  2. Judging similarity. Iowa L. Rev., 100:267, 2014.
  3. Copyright’s framing problem. UCLA L. Rev., 64:1102, 2017.
  4. Fair learning. Tex. L. Rev., 99:743, 2020.
  5. Midjourney. Midjourney, 2022. URL https://www.midjourney.com/.
  6. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
  7. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 2022.
  8. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  10684–10695, 2022.
  9. Sag, M. The new legal landscape for text mining and machine learning. J. Copyright Soc’y USA, 66:291, 2018.
  10. Laion-400m: Open dataset of clip-filtered 400 million image-text pairs. arXiv preprint arXiv:2111.02114, 2021.
  11. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pp. 2256–2265. PMLR, 2015.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Stephen Casper (40 papers)
  2. Zifan Guo (1 paper)
  3. Shreya Mogulothu (1 paper)
  4. Zachary Marinov (1 paper)
  5. Chinmay Deshpande (3 papers)
  6. Rui-Jie Yew (7 papers)
  7. Zheng Dai (7 papers)
  8. Dylan Hadfield-Menell (54 papers)
Citations (9)
Youtube Logo Streamline Icon: https://streamlinehq.com