Diffusion Attribution Score: Evaluating Training Data Influence in Diffusion Models (2410.18639v4)
Abstract: As diffusion models become increasingly popular, the misuse of copyrighted and private images has emerged as a major concern. One promising solution to mitigate this issue is identifying the contribution of specific training samples in generative models, a process known as data attribution. Existing data attribution methods for diffusion models typically quantify the contribution of a training sample by evaluating the change in diffusion loss when the sample is included or excluded from the training process. However, we argue that the direct usage of diffusion loss cannot represent such a contribution accurately due to the calculation of diffusion loss. Specifically, these approaches measure the divergence between predicted and ground truth distributions, which leads to an indirect comparison between the predicted distributions and cannot represent the variances between model behaviors. To address these issues, we aim to measure the direct comparison between predicted distributions with an attribution score to analyse the training sample importance, which is achieved by Diffusion Attribution Score (\textit{DAS}). Underpinned by rigorous theoretical analysis, we elucidate the effectiveness of DAS. Additionally, we explore strategies to accelerate DAS calculations, facilitating its application to large-scale diffusion models. Our extensive experiments across various datasets and diffusion models demonstrate that DAS significantly surpasses previous benchmarks in terms of the linear data-modelling score, establishing new state-of-the-art performance. Code is available at \hyperlink{here}{https://github.com/Jinxu-Lin/DAS}.
- Towards tracing knowledge in language models back to the training data. In Findings of the Association for Computational Linguistics: EMNLP 2022. Association for Computational Linguistics, 2022.
- Clustering with bregman divergences. Journal of machine learning research, 2005.
- Relatif: Identifying explanatory training samples via relative influence. In International Conference on Artificial Intelligence and Statistics, 2020.
- Influence functions in deep learning are fragile. In International Conference on Learning Representations, 2021.
- Machine unlearning. In 2021 IEEE Symposium on Security and Privacy (SP), 2021.
- Extracting training data from diffusion models. In 32nd USENIX Security Symposium (USENIX Security 23), 2023.
- Input similarity from the neural network perspective. Advances in Neural Information Processing Systems, 2019.
- Hydra: Hypergradient data relevance analysis for interpreting deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, 2021.
- R Dennis Cook. Detection of influential observation in linear regression. Technometrics, 1977.
- Training data attribution for diffusion models, 2023.
- What neural networks memorize and why: Discovering the long tail via influence estimation. Advances in Neural Information Processing Systems, 2020.
- The journey, not the destination: How data guides diffusion models. In Arxiv preprint arXiv:2312.06205, 2023.
- Data shapley: Equitable valuation of data for machine learning. In International conference on machine learning, 2019.
- Identifying a training-set attack’s target using renormalized influence estimation. In Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, 2022.
- Training data influence analysis and estimation: a survey. Machine Learning, 2024.
- Trevor Hastie. Ridge regularization: An essential concept in data science. Technometrics, 2020.
- Prompt-to-prompt image editing with cross-attention control. In The Eleventh International Conference on Learning Representations, 2023.
- Classifier-free diffusion guidance. In NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications, 2021.
- Denoising diffusion probabilistic models. Advances in neural information processing systems, 2020.
- Imagen video: High definition video generation with diffusion models. arXiv preprint arXiv:2210.02303, 2022.
- LoRA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
- Datamodels: Predicting predictions from training data. In ICML, 2022.
- Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 2018.
- Towards efficient data valuation based on the shapley value. In The 22nd International Conference on Artificial Intelligence and Statistics, 2019.
- Scalability vs. utility: Do we have to sacrifice one for the other in data importance quantification? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- Extensions of lipschitz maps into a hilbert space. Contemporary Mathematics, 1984.
- Interpreting black box predictions using fisher kernels. In The 22nd International Conference on Artificial Intelligence and Statistics, 2019.
- Understanding black-box predictions via influence functions. In International conference on machine learning, 2017.
- Interpretable diffusion via information decomposition. In The Twelfth International Conference on Learning Representations, 2024.
- Understanding instance-based interpretability of variational auto-encoders. Advances in Neural Information Processing Systems, 2021.
- A Krizhevsky. Learning multiple layers of features from tiny images. Master’s thesis, University of Tront, 2009.
- Diffusion-lm improves controllable text generation. Advances in Neural Information Processing Systems, 2022.
- The artbench dataset: Benchmarking generative models with artworks, 2022.
- Measuring the effect of training data on deep learning predictions via randomized experiments. In International Conference on Machine Learning, 2022.
- Influence selection for active learning. In Proceedings of the IEEE/CVF international conference on computer vision, 2021.
- Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, 2015.
- Decoupled weight decay regularization. In International Conference on Learning Representations, 2019.
- A unified approach to interpreting model predictions. Advances in neural information processing systems, 2017.
- Deepcache: Accelerating diffusion models for free. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15762–15772, 2024.
- A kernel-based view of language model fine-tuning. In International Conference on Machine Learning, 2023.
- Trak: Attributing model behavior at scale. In International Conference on Machine Learning, 2023.
- Daryl Pregibon. Logistic regression diagnostics. The annals of statistics, 1981.
- Estimating training data influence by tracing gradient descent. Advances in Neural Information Processing Systems, 2020.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, 2021.
- Model-agnostic interpretability of machine learning, 2016.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022.
- Palette: Image-to-image diffusion models. In ACM SIGGRAPH 2022 conference proceedings, 2022.
- Scaling up influence functions. In Proceedings of the AAAI Conference on Artificial Intelligence, 2022.
- Modeldiff: A framework for comparing learning algorithms. In International Conference on Machine Learning, 2023.
- Lloyd S Shapley et al. A value for n-person games. Machine Learning, 1953.
- Diffusion art or digital forgery? investigating data replication in diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
- Denoising diffusion implicit models. In International Conference on Learning Representations, 2021a.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b.
- Influence estimation for generative adversarial networks. In International Conference on Learning Representations, 2021.
- Neural discrete representation learning. Advances in neural information processing systems, 2017.
- Evaluating data attribution for text-to-image models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
- Representer point selection for explaining deep neural networks. Advances in neural information processing systems, 2018.
- Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
- Intriguing properties of data attribution on diffusion models. In The Twelfth International Conference on Learning Representations, 2024.