Large-scale Reinforcement Learning for Diffusion Models (2401.12244v1)
Abstract: Text-to-image diffusion models are a class of deep generative models that have demonstrated an impressive capacity for high-quality image generation. However, these models are susceptible to implicit biases that arise from web-scale text-image training pairs and may inaccurately model aspects of images we care about. This can result in suboptimal samples, model bias, and images that do not align with human ethics and preferences. In this paper, we present an effective scalable algorithm to improve diffusion models using Reinforcement Learning (RL) across a diverse set of reward functions, such as human preference, compositionality, and fairness over millions of images. We illustrate how our approach substantially outperforms existing methods for aligning diffusion models with human preferences. We further illustrate how this substantially improves pretrained Stable Diffusion (SD) models, generating samples that are preferred by humans 80.3% of the time over those from the base SD model while simultaneously improving both the composition and diversity of generated samples.
- A general language assistant as a laboratory for alignment, 2021.
- Training a helpful and harmless assistant with reinforcement learning from human feedback, 2022.
- Hrs-bench: Holistic, reliable and scalable benchmark for text-to-image models, 2023.
- Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. In 2023 ACM Conference on Fairness, Accountability, and Transparency. ACM, 2023.
- Training diffusion models with reinforcement learning, 2023.
- Conceptual 12m: Pushing web-scale image-text pre-training to recognize long-tail visual concepts, 2021.
- Attend-and-excite: Attention-based semantic guidance for text-to-image diffusion models, 2023.
- Training-free layout control with cross-attention guidance, 2023.
- Investigating gender and racial biases in dall-e mini images. manuscript.
- Dall-eval: Probing the reasoning skills and social biases of text-to-image generation models, 2023.
- Fair generative modeling via weak supervision, 2020.
- Debiasing vision-language models via biased prompts, 2023.
- Directly fine-tuning diffusion models on differentiable rewards, 2023.
- Emu: Enhancing image generation models using photogenic needles in a haystack, 2023.
- Raft: Reward ranked finetuning for generative foundation model alignment, 2023.
- Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. In International Conference on Machine Learning, pages 8489–8510. PMLR, 2023.
- Mitigating stereotypical biases in text to image generative systems, 2023.
- Dpok: Reinforcement learning for fine-tuning text-to-image diffusion models, 2023.
- Training-free structured diffusion guidance for compositional text-to-image synthesis, 2023a.
- Layoutgpt: Compositional visual planning and generation with large language models, 2023b.
- Fair diffusion: Instructing text-to-image generation models on fairness, 2023.
- Benchmarking spatial relationships in text-to-image generation, 2023.
- Classifier-free diffusion guidance, 2022.
- Denoising diffusion probabilistic models, 2020.
- T2i-compbench: A comprehensive benchmark for open-world compositional text-to-image generation, 2023.
- Aligning text-to-image models using human feedback, 2023.
- Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation, 2022.
- Gligen: Open-set grounded text-to-image generation, 2023.
- Llm-grounded diffusion: Enhancing prompt understanding of text-to-image diffusion models with large language models, 2023.
- Compositional visual generation with composable diffusion models. In European Conference on Computer Vision, pages 423–439. Springer, 2022.
- Decoupled weight decay regularization, 2019.
- Stable bias: Analyzing societal representations in diffusion models, 2023.
- Monte carlo gradient estimation in machine learning, 2020.
- Social biases through the text-to-image generation lens, 2023.
- Webgpt: Browser-assisted question-answering with human feedback, 2022.
- Training language models to follow instructions with human feedback, 2022.
- Learning transferable visual models from natural language supervision, 2021.
- Exploring the limits of transfer learning with a unified text-to-text transformer, 2023.
- High-resolution image synthesis with latent diffusion models, 2022.
- Laion-5b: An open large-scale dataset for training next generation image-text models, 2022.
- Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Balancing the picture: Debiasing vision-language datasets with synthetic contrast sets, 2023.
- Deep unsupervised learning using nonequilibrium thermodynamics, 2015.
- Denoising diffusion implicit models, 2022.
- Policy gradient methods for reinforcement learning with function approximation. In Advances in Neural Information Processing Systems. MIT Press, 1999.
- Christopher T. H Teo and Ngai-Man Cheung. Measuring fairness in generative models, 2021.
- DiffusionDB: A large-scale prompt gallery dataset for text-to-image generative models. arXiv:2210.14896 [cs], 2022.
- Ronald J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8:229–256, 2004.
- Harnessing the spatial-temporal attention of diffusion models for high-fidelity text-to-image synthesis, 2023a.
- Human preference score: Better aligning text-to-image models with human preference, 2023b.
- Imagereward: Learning and evaluating human preferences for text-to-image generation, 2023.
- Scaling autoregressive models for content-rich text-to-image generation, 2022.
- ITI-GEN: Inclusive text-to-image generation. In ICCV, 2023a.
- Auditing gender presentation differences in text-to-image models, 2023b.
- Simple multi-dataset detection, 2022.
- Yinan Zhang (31 papers)
- Eric Tzeng (17 papers)
- Yilun Du (113 papers)
- Dmitry Kislyuk (8 papers)