Latent Diffusion Energy-Based Model for Interpretable Text Modeling (2206.05895v4)
Abstract: Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in generative modeling. Fueled by its flexibility in the formulation and strong modeling power of the latent space, recent works built upon it have made interesting attempts aiming at the interpretability of text modeling. However, latent space EBMs also inherit some flaws from EBMs in data space; the degenerate MCMC sampling quality in practice can lead to poor generation quality and instability in training, especially on data with complex latent structures. Inspired by the recent efforts that leverage diffusion recovery likelihood learning as a cure for the sampling issue, we introduce a novel symbiosis between the diffusion models and latent space EBMs in a variational learning framework, coined as the latent diffusion energy-based model. We develop a geometric clustering-based regularization jointly with the information bottleneck to further improve the quality of the learned latent space. Experiments on several challenging tasks demonstrate the superior performance of our model on interpretable text modeling over strong counterparts.
- Modeling worlds in text. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Structured denoising diffusion models in discrete state-spaces. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Generalized denoising auto-encoders as generative models. In Advances in Neural Information Processing Systems (NeurIPS), 2013.
- Generating sentences from a continuous space. In Conference on Computational Natural Language Learning (CoNLL), 2016.
- The mathematics of statistical machine translation: Parameter estimation. Computational Linguistics, 1993.
- Importance weighted autoencoders. In International Conference on Learning Representations (ICLR), 2016.
- Neural models for documents with metadata. In Annual Meeting of the Association for Computational Linguistics (ACL), 2018.
- Learning phrase representations using rnn encoder-decoder for statistical machine translation. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014.
- Improved contrastive divergence training of energy based models. In International Conference on Machine Learning (ICML), 2021.
- Key-value retrieval networks for task-oriented dialogue. In Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDial), 2017.
- Implicit deep latent variable models for text generation. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019.
- Bootstrapping dialog systems with word embeddings. In Advances in Neural Information Processing Systems (NeurIPS), 2014.
- Cyclical annealing schedule: A simple approach to mitigating kl vanishing. In North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), 2019.
- Learning energy-based models by diffusion recovery likelihood. In International Conference on Learning Representations (ICLR), 2020.
- Your classifier is secretly an energy based model and you should treat it like one. In International Conference on Learning Representations (ICLR), 2019.
- A deep generative framework for paraphrase generation. In AAAI Conference on Artificial Intelligence (AAAI), 2018.
- Variational pretraining for semi-supervised text classification. In Annual Meeting of the Association for Computational Linguistics (ACL), 2019.
- Divergence triangle for joint training of generator model, energy-based model, and inferential model. In Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- Joint training of variational auto-encoder and latent energy-based model. In Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- Lagging inference networks and posterior collapse in variational autoencoders. In International Conference on Learning Representations (ICLR), 2018.
- beta-vae: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations (ICLR), 2016.
- Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Discrete latent variable representations for low-resource text classification. In Annual Meeting of the Association for Computational Linguistics (ACL), 2020.
- Learning and inferring movement with deep generative model. arXiv preprint arXiv:1805.07252, 2018.
- Task transfer by preference-based cost learning. In AAAI Conference on Artificial Intelligence (AAAI), 2019.
- Billion-scale similarity search with gpus. IEEE Transactions on Big Data, 2019.
- Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
- Semi-supervised learning with deep generative models. In Advances in Neural Information Processing Systems (NeurIPS), 2014.
- A surprisingly effective fix for deep latent variable modeling of text. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2019.
- Delete, retrieve, generate: a simple approach to sentiment and style transfer. In North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), 2018.
- Deep recurrent generative decoder for abstractive text summarization. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2017a.
- Dailydialog: A manually labelled multi-turn dialogue dataset. In Annual Meeting of the Association for Computational Linguistics (ACL), 2017b.
- Building a large annotated corpus of english: the penn treebank. Computational Linguistics, 1993.
- Regularizing and optimizing lstm language models. In International Conference on Learning Representations (ICLR), 2018.
- Neural variational inference for text processing. In International Conference on Machine Learning (ICML), 2016.
- Recurrent neural network based language model. In Interspeech, 2010.
- Vector-based models of semantic composition. In Annual Meeting of the Association for Computational Linguistics (ACL), 2008.
- Spectral normalization for generative adversarial networks. In International Conference on Learning Representations (ICLR), 2018.
- Controllable and compositional generation with latent-space energy-based models. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Learning non-convergent non-persistent short-run mcmc toward energy-based model. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
- On the anatomy of mcmc-based maximum likelihood learning of energy-based models. In AAAI Conference on Artificial Intelligence (AAAI), 2020.
- Latent space energy-based model of symbol-vector coupling for text generation and classification. In International Conference on Machine Learning (ICML), 2021.
- Learning latent space energy-based prior model. In Advances in Neural Information Processing Systems (NeurIPS), 2020a.
- Learning latent space energy-based prior model for molecule generation. arXiv preprint arXiv:2010.09351, 2020b.
- Trajectory prediction with latent belief energy-based model. In Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
- Bleu: a method for automatic evaluation of machine translation. In Annual Meeting of the Association for Computational Linguistics (ACL), 2002.
- Glove: Global vectors for word representation. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014.
- Stochastic backpropagation and approximate inference in deep generative models. In International Conference on Machine Learning (ICML), 2014.
- An optimal assessment of natural language student input using word-to-word similarity metrics. In International Conference on Intelligent Tutoring Systems, 2012.
- Building end-to-end dialogue systems using generative hierarchical neural network models. In AAAI Conference on Artificial Intelligence (AAAI), 2016.
- A hierarchical latent variable encoder-decoder model for generating dialogues. In AAAI Conference on Artificial Intelligence (AAAI), 2017.
- Dispersed exponential family mixture vaes for interpretable text generation. In International Conference on Machine Learning (ICML), 2020.
- D2c: Diffusion-denoising models for few-shot conditional generation. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning (ICML), 2015.
- Generative modeling by estimating gradients of the data distribution. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
- Improved techniques for training score-based generative models. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations (ICLR), 2020.
- Towards text generation with adversarially learned neural outlines. In Advances in Neural Information Processing Systems (NeurIPS), 2018.
- The information bottleneck method. arXiv preprint physics/0004057, 2000.
- NVAE: A deep hierarchical variational autoencoder. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Score-based generative modeling in latent space. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Vincent, P. A connection between score matching and denoising autoencoders. Neural Computation, 2011.
- Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. Journal of Machine Learning Research (JMLR), 2010.
- Topic-guided variational auto-encoder for text generation. In North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), 2019.
- Diffusion priors in variational autoencoders. arXiv preprint arXiv:2106.15671, 2021.
- Bayesian learning via stochastic gradient langevin dynamics. In International Conference on Machine Learning (ICML), 2011.
- Latent intention dialogue models. In International Conference on Machine Learning (ICML), 2017.
- A theory of generative convnet. In International Conference on Machine Learning (ICML), 2016.
- Pomdp-based statistical spoken dialog systems: A review. Proceedings of the IEEE, 2013.
- Unsupervised foreground extraction via deep region competition. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- Variational neural machine translation. In Annual Conference on Empirical Methods in Natural Language Processing (EMNLP), 2016.
- Character-level convolutional networks for text classification. In Advances in Neural Information Processing Systems (NeurIPS), 2015.
- Adversarially regularized autoencoders. In International Conference on Machine Learning (ICML), 2018a.
- Learning discourse-level diversity for neural dialog models using conditional variational autoencoders. In Annual Meeting of the Association for Computational Linguistics (ACL), 2017.
- Unsupervised discrete sentence representation learning for interpretable neural dialog generation. In Annual Meeting of the Association for Computational Linguistics (ACL), 2018b.