Faithful Low-Resource Data-to-Text Generation through Cycle Training (2305.14793v2)
Abstract: Methods to generate text from structured data have advanced significantly in recent years, primarily due to fine-tuning of pre-trained LLMs on large datasets. However, such models can fail to produce output faithful to the input data, particularly on out-of-domain data. Sufficient annotated data is often not available for specific domains, leading us to seek an unsupervised approach to improve the faithfulness of output text. Since the problem is fundamentally one of consistency between the representations of the structured data and text, we evaluate the effectiveness of cycle training in this work. Cycle training uses two models which are inverses of each other: one that generates text from structured data, and one which generates the structured data from natural language text. We show that cycle training, when initialized with a small amount of supervised data (100 samples in our case), achieves nearly the same performance as fully supervised approaches for the data-to-text generation task on the WebNLG, E2E, WTQ, and WSQL datasets. We perform extensive empirical analysis with automated evaluation metrics and a newly designed human evaluation schema to reveal different cycle training strategies' effectiveness of reducing various types of generation errors. Our code is publicly available at https://github.com/Edillower/CycleNLG.
- Machine translation aided bilingual data-to-text generation and semantic parsing. In Proceedings of the 3rd International Workshop on Natural Language Generation from the Semantic Web (WebNLG+), pages 125–130, Dublin, Ireland (Virtual). Association for Computational Linguistics.
- Ron Artstein and Massimo Poesio. 2008. Survey article: Inter-coder agreement for computational linguistics. Computational Linguistics, 34(4):555–596.
- Satanjeev Banerjee and Alon Lavie. 2005. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65–72, Ann Arbor, Michigan. Association for Computational Linguistics.
- Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
- The 2020 bilingual, bi-directional WebNLG+ shared task: Overview and evaluation results (WebNLG+ 2020). In Proceedings of the 3rd International Workshop on Natural Language Generation from the Semantic Web (WebNLG+), pages 55–76, Dublin, Ireland (Virtual). Association for Computational Linguistics.
- WikiTableT: A Large-Scale Data-to-Text Dataset for Generating Wikipedia Article Sections. Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 193–209.
- TabFact: A Large-scale Dataset for Table-based Fact Verification. In International Conference on Learning Representations (ICLR), arXiv, Addis Ababa, Ethiopia.
- The WebNLG Challenge: Generating Text from DBPedia Data. Proceedings of the 9th International Natural Language Generation conference, pages 163–167.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
- Handling Divergent Reference Texts when Evaluating Table-to-Text Generation. arXiv. This is the PARENT evaluation metric paper.
- Fact Checking Machine Generated Text with Dependency Trees. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP).
- Creating Training Corpora for NLG Micro-Planners. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 179–188.
- The WebNLG Challenge: Generating Text from RDF Data. Proceedings of the 10th International Conference on Natural Language Generation, pages 124–133.
- CycleGT: Unsupervised graph-to-text and text-to-graph generation via cycle training. In Proceedings of the 3rd International Workshop on Natural Language Generation from the Semantic Web (WebNLG+), pages 77–88, Dublin, Ireland (Virtual). Association for Computational Linguistics.
- TaPas: Weakly Supervised Table Parsing via Pre-training. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4320–4333.
- Iterative back-translation for neural machine translation. In Proceedings of the 2nd workshop on neural machine translation and generation, pages 18–24.
- CycleNER: An Unsupervised Training Approach for Named Entity Recognition. Proceedings of the ACM Web Conference 2022, pages 2916–2924.
- CycleKQR: Unsupervised bidirectional keyword-question rewriting. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11875–11886, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Professor forcing: A new algorithm for training recurrent networks. Advances in neural information processing systems, 29.
- Unsupervised Machine Translation Using Monolingual Corpora Only. arXiv.
- BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, Online. Association for Computational Linguistics.
- Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
- TAPEX: Table Pre-training via Learning a Neural SQL Executor. arXiv.
- SemEval-2022 task 11: Multilingual complex named entity recognition (MultiCoNER). In Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022), pages 1412–1437, Seattle, United States. Association for Computational Linguistics.
- DART: Open-Domain Structured Data Record to Text Generation. arXiv.
- The E2E dataset: New challenges for end-to-end generation. In Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pages 201–206, Saarbrücken, Germany. Association for Computational Linguistics.
- Richard Yuanzhe Pang and Kevin Gimpel. 2019. Unsupervised Evaluation Metrics and Learning Criteria for Non-Parallel Textual Transfer. Proceedings of the 3rd Workshop on Neural Generation and Translation, pages 138–147.
- Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pages 311–318, Philadelphia, Pennsylvania, USA. Association for Computational Linguistics.
- ToTTo: A Controlled Table-To-Text Generation Dataset. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1173–1186.
- Panupong Pasupat and Percy Liang. 2015. Compositional semantic parsing on semi-structured tables. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1470–1480, Beijing, China. Association for Computational Linguistics.
- Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res., 21(1).
- Few-shot table-to-text generation with prototype memory. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 910–917, Punta Cana, Dominican Republic. Association for Computational Linguistics.
- What matters for shoppers: Investigating key attributes for online product comparison. In European Conference on Information Retrieval, pages 231–239. Springer.
- Generating explainable product comparisons for online shopping. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, pages 949–957.
- Ronald J Williams and David Zipser. 1989. A learning algorithm for continually running fully recurrent neural networks. Neural computation, 1(2):270–280.
- ASDOT: Any-shot data-to-text generation with pretrained language models. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 1886–1899, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- TableFormer: Robust Transformer Modeling for Table-Text Encoding. arXiv. Very interesting approach to use scalar attention biases between different types of content, e.g. table columns and the input query.
- Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations.
- Seq2sql: Generating structured queries from natural language using reinforcement learning. arXiv preprint arXiv:1709.00103.
- Learning Dense Correspondence via 3D-Guided Cycle Consistency. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 117–126.
- Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. 2017 IEEE International Conference on Computer Vision (ICCV), pages 2242–2251.
- Zhuoer Wang (9 papers)
- Marcus Collins (5 papers)
- Nikhita Vedula (11 papers)
- Simone Filice (9 papers)
- Shervin Malmasi (40 papers)
- Oleg Rokhlenko (22 papers)