Challenges in Domain-Specific Abstractive Summarization and How to Overcome them (2307.00963v1)
Abstract: LLMs work quite well with general-purpose data and many tasks in Natural Language Processing. However, they show several limitations when used for a task such as domain-specific abstractive text summarization. This paper identifies three of those limitations as research problems in the context of abstractive text summarization: 1) Quadratic complexity of transformer-based models with respect to the input text length; 2) Model Hallucination, which is a model's ability to generate factually incorrect text; and 3) Domain Shift, which happens when the distribution of the model's training and test corpus is not the same. Along with a discussion of the open research questions, this paper also provides an assessment of existing state-of-the-art techniques relevant to domain-specific text summarization to address the research gaps.
- Text summarization techniques: A brief survey.
- METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65–72, Ann Arbor, Michigan. Association for Computational Linguistics.
- Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.
- CLIFF: Contrastive learning for improving faithfulness and factuality in abstractive summarization. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Improving faithfulness in abstractive summarization with contrast candidate generation and selection. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online. Association for Computational Linguistics.
- Rethinking attention with performers. arXiv preprint arXiv:2009.14794.
- Neural path hunter: Reducing hallucination in dialogue systems via path grounding. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 2197–2214, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Ranking generated summaries by correctness: An interesting but challenging application for natural language inference. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy. Association for Computational Linguistics.
- GO FIGURE: A meta evaluation of factuality in summarization. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 478–487, Online. Association for Computational Linguistics.
- The reversible residual network: backpropagation without storing activations. Proceedings of the 31st International Conference on Neural Information Processing Systems, pages 2211–2221.
- Assessing the factual accuracy of generated text. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &; Data Mining, KDD ’19, page 166–175, New York, NY, USA. Association for Computing Machinery.
- Mind the facts: Knowledge-boosted coherent abstractive text summarization. NeurIPS, Knowledge Representation & Reasoning Meets Machine Learning (KR2ML workshop), abs/2006.15435.
- Abstractive summarization: An overview of the state of the art. Expert Systems with Applications, 121:49–65.
- Long short-term memory. Neural Computation, 9(8):1735–1780.
- The factual inconsistency problem in abstractive text summarization: A survey. CoRR, abs/2104.14839.
- UDALM: Unsupervised domain adaptation through language modeling. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2579–2590, Online. Association for Computational Linguistics.
- Reformer: The efficient transformer. CoRR, abs/2001.04451.
- Automatic text summarization: A state-of-the-art review. ICEIS (1), pages 648–655.
- Explainable automated fact-checking for public health claims. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online. Association for Computational Linguistics.
- Evaluating the factual consistency of abstractive text summarization. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 9332–9346, Online. Association for Computational Linguistics.
- Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. CoRR, abs/1808.06226.
- BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. CoRR, abs/1910.13461.
- Lin, C.-Y. (2004). ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
- Roberta: A robustly optimized bert pretraining approach.
- On faithfulness and factuality in abstractive summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1906–1919, Online. Association for Computational Linguistics.
- Summarization of biomedical articles using domain-specific word embeddings and graph ranking. Journal of Biomedical Informatics, 107:103452.
- Bleu: A method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, ACL ’02, page 311–318, USA. Association for Computational Linguistics.
- Efficient domain adaptation of language models via adaptive tokenization. Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, pages 155–165.
- A survey of evaluation metrics used for nlg systems. ACM Computing Surveys (CSUR), 55(2):1–39.
- Joint parsing and generation for abstractive summarization. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05):8894–8901.
- Linear-time wordpiece tokenization. CoRR, abs/2012.15524.
- Efficient transformers: A survey. CoRR, abs/2009.06732.
- Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems, pages 6000–6010.
- Big bird: Transformers for longer sequences. CoRR, abs/2007.14062.
- Automated fact-checking: A survey. Language and Linguistics Compass, 15(10):e12438.
- PEGASUS: pre-training with extracted gap-sentences for abstractive summarization. CoRR, abs/1912.08777.
- Bertscore: Evaluating text generation with BERT. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net.
- Mtl-das: Automatic text summarization for domain adaptation. Intell. Neuroscience, 2022.
- Detecting hallucinated content in conditional neural sequence generation. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021.