Fine-grained Text Style Transfer with Diffusion-Based Language Models (2305.19512v2)
Abstract: Diffusion probabilistic models have shown great success in generating high-quality images controllably, and researchers have tried to utilize this controllability into text generation domain. Previous works on diffusion-based LLMs have shown that they can be trained without external knowledge (such as pre-trained weights) and still achieve stable performance and controllability. In this paper, we trained a diffusion-based model on StylePTB dataset, the standard benchmark for fine-grained text style transfers. The tasks in StylePTB requires much more refined control over the output text compared to tasks evaluated in previous works, and our model was able to achieve state-of-the-art performance on StylePTB on both individual and compositional transfers. Moreover, our model, trained on limited data from StylePTB without external knowledge, outperforms previous works that utilized pretrained weights, embeddings, and external grammar parsers, and this may indicate that diffusion-based LLMs have great potential under low-resource settings.
- Structured denoising diffusion models in discrete state-spaces. Advances in Neural Information Processing Systems, 34:17981–17993.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
- Deep learning in spoken and text-based dialog systems. In Deep Learning in Natural Language Processing, pages 49–78. Springer.
- Diffuseq: Sequence to sequence text generation with diffusion models. arXiv preprint arXiv:2210.08933.
- Generating sentences by editing prototypes. Transactions of the Association for Computational Linguistics, 6:437–450.
- A retrieve-and-edit framework for predicting structured outputs. In Advances in Neural Information Processing Systems, pages 10052–10062.
- A probabilistic formulation of unsupervised text style transfer. arXiv preprint arXiv:2002.03912.
- Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851.
- Argmax flows and multinomial diffusion: Learning categorical distributions. Advances in Neural Information Processing Systems, 34:12454–12465.
- Eduard Hovy. 1987. Generating natural language under pragmatic constraints. Journal of Pragmatics, 11(6):689–719.
- Social robots as embedded reinforcers of social behavior in children with autism. Journal of autism and developmental disorders.
- Yoon Kim. 2021. Sequence-to-sequence learning with latent neural grammars. Advances in Neural Information Processing Systems, 34:26302–26317.
- Multiple-attribute text rewriting. In International Conference on Learning Representations.
- Delete, retrieve, generate: a simple approach to sentiment and style transfer. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1865–1874. Association for Computational Linguistics.
- Diffusion-lm improves controllable text generation. arXiv preprint arXiv:2205.14217.
- On emergent communication in competitive multi-agent teams. In AAMAS.
- Styleptb: A compositional benchmark for fine-grained controllable text style transfer. arXiv preprint arXiv:2104.05196.
- Politeness transfer: A tag and generate approach. arXiv preprint arXiv:2004.14257.
- Building a large annotated corpus of English: The Penn Treebank. Computational Linguistics, 19(2):313–330.
- Emotion recognition and adaptation in spoken dialogue systems. International Journal of Speech Technology.
- Style transfer through back-translation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 866–876. Association for Computational Linguistics.
- Language models are unsupervised multitask learners.
- Sudha Rao and Joel Tetreault. 2018. Dear sir or madam, may i introduce the gyafc dataset: Corpus, benchmarks and metrics for formality style transfer. arXiv preprint arXiv:1803.06535.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10684–10695.
- Tailor: Generating and perturbing text with semantic controls. arXiv preprint arXiv:2107.07150.
- Relevance of unsupervised metrics in task-oriented dialogue for evaluating natural language generation. CoRR, abs/1706.09799.
- Style transfer from non-parallel text by cross-alignment. In Advances in Neural Information Processing Systems, pages 6833–6844.
- Extracting latent steering vectors from pretrained language models. arXiv preprint arXiv:2205.05124.
- Transforming delete, retrieve, generate approach for controlled text style transfer. arXiv preprint arXiv:1908.09368.
- Sequence to sequence learning with neural networks. In Advances in neural information processing systems, pages 3104–3112.
- Paraphrasing for style. In Proceedings of COLING 2012, pages 2899–2914.
- Yiwei Lyu (30 papers)
- Tiange Luo (13 papers)
- Jiacheng Shi (11 papers)
- Todd C. Hollon (8 papers)
- Honglak Lee (174 papers)