PoetryDiffusion: Towards Joint Semantic and Metrical Manipulation in Poetry Generation (2306.08456v3)
Abstract: Controllable text generation is a challenging and meaningful field in natural language generation (NLG). Especially, poetry generation is a typical one with well-defined and strict conditions for text generation which is an ideal playground for the assessment of current methodologies. While prior works succeeded in controlling either semantic or metrical aspects of poetry generation, simultaneously addressing both remains a challenge. In this paper, we pioneer the use of the Diffusion model for generating sonnets and Chinese SongCi poetry to tackle such challenges. In terms of semantics, our PoetryDiffusion model, built upon the Diffusion model, generates entire sentences or poetry by comprehensively considering the entirety of sentence information. This approach enhances semantic expression, distinguishing it from autoregressive and LLMs. For metrical control, the separation feature of diffusion generation and its constraint control module enable us to flexibly incorporate a novel metrical controller to manipulate and evaluate metrics (format and rhythm). The denoising process in PoetryDiffusion allows for gradual enhancement of semantics and flexible integration of the metrical controller which can calculate and impose penalties on states that stray significantly from the target control distribution. Experimental results on two datasets demonstrate that our model outperforms existing models in automatic evaluation of semantic, metrical, and overall performance as well as human evaluation.
- Shall i compare thee to a machine-written sonnet? an approach to algorithmic sonnet generation. arXiv preprint arXiv:1811.05067.
- Controlled Text Generation with Adversarial Learning. In Proceedings of the 13th International Conference on Natural Language Generation, 29–34. Dublin, Ireland: Association for Computational Linguistics.
- Deep Generative Modelling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11): 7327–7347.
- Language models are few-shot learners. Advances in neural information processing systems, 33: 1877–1901.
- Maximum-likelihood augmented discrete generative adversarial networks (2017). arXiv preprint arXiv:1702.07983.
- Sentiment-Controllable Chinese Poetry Generation. In IJCAI, 4925–4931.
- Controllable Conversation Generation with Conversation Structures via Diffusion Models. In Findings of the Association for Computational Linguistics: ACL 2023, 7238–7251. Toronto, Canada: Association for Computational Linguistics.
- An iterative polishing framework based on quality aware masked language model for Chinese poetry generation. In Proceedings of the AAAI conference on artificial intelligence, volume 34, 7643–7650.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
- Generating topical poetry. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 1183–1191.
- Generative adversarial networks. Communications of the ACM, 63(11): 139–144.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
- Toward controlled generation of text. In International conference on machine learning, 1587–1596. PMLR.
- Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114.
- Deep-speare: A joint neural model of poetic language, meter and rhyme. arXiv preprint arXiv:1807.03491.
- A diversity-promoting objective function for neural conversation models. arXiv preprint arXiv:1510.03055.
- Rigid formats controlled text generation. In Proceedings of the 58th annual meeting of the association for computational linguistics, 742–751.
- Diffusion-LM Improves Controllable Text Generation. arXiv preprint arXiv:2205.14217.
- Lin, C.-Y. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, 74–81.
- Adversarial ranking for language generation. Advances in neural information processing systems, 30.
- OpenAI. 2021. GPT-3.5-turbo: Language Model. https://platform.openai.com/docs/guides/chatgpt.
- Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, 311–318.
- ChatGPT vs Human-authored Text: Insights into Controllable Text Summarization and Sentence Style Transfer. arXiv preprint arXiv:2306.07799.
- Songmass: Automatic song writing with pre-training and alignment constraint. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, 13798–13805.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, 2256–2265. PMLR.
- Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502.
- Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27.
- Zero-shot Sonnet Generation with Discourse-level Planning and Aesthetics Features. arXiv preprint arXiv:2205.01821.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
- Van de Cruys, T. 2020. Automatic poetry generation from prosaic text. In Proceedings of the 58th annual meeting of the association for computational linguistics, 2471–2480.
- Mixpoet: Diverse poetry generation via learning controllable mixed latent space. In Proceedings of the AAAI conference on artificial intelligence, volume 34, 9450–9457.
- Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI conference on artificial intelligence, volume 31.
- SummIt: Iterative Text Summarization via ChatGPT. arXiv preprint arXiv:2305.14835.
- DisCup: Discriminator cooperative unlikelihood prompt-tuning for controllable text generation. arXiv preprint arXiv:2210.09551.
- Generating Chinese CI with designated metrical structure. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, 7459–7467.
- Chinese poetry generation with recurrent neural networks. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 670–680.
- Controlled text generation with natural language instructions. arXiv preprint arXiv:2304.14293.
- Zhiyuan Hu (30 papers)
- Chumin Liu (2 papers)
- Yue Feng (55 papers)
- Anh Tuan Luu (69 papers)
- Bryan Hooi (159 papers)