Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

JADS: A Framework for Self-supervised Joint Aspect Discovery and Summarization (2405.18642v1)

Published 28 May 2024 in cs.AI and cs.CL

Abstract: To generate summaries that include multiple aspects or topics for text documents, most approaches use clustering or topic modeling to group relevant sentences and then generate a summary for each group. These approaches struggle to optimize the summarization and clustering algorithms jointly. On the other hand, aspect-based summarization requires known aspects. Our solution integrates topic discovery and summarization into a single step. Given text data, our Joint Aspect Discovery and Summarization algorithm (JADS) discovers aspects from the input and generates a summary of the topics, in one step. We propose a self-supervised framework that creates a labeled dataset by first mixing sentences from multiple documents (e.g., CNN/DailyMail articles) as the input and then uses the article summaries from the mixture as the labels. The JADS model outperforms the two-step baselines. With pretraining, the model achieves better performance and stability. Furthermore, embeddings derived from JADS exhibit superior clustering capabilities. Our proposed method achieves higher semantic alignment with ground truth and is factual.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Unsupervised neural networks for automatic arabic text summarization using document clustering and topic modeling. Expert Systems with Applications, 172:114652.
  2. Longformer: The long-document transformer. arXiv:2004.05150.
  3. Latent dirichlet allocation. Journal of machine Learning research, 3(Jan):993–1022.
  4. Unsupervised aspect-based multi-document abstractive summarization. In Proceedings of the 2nd Workshop on New Frontiers in Summarization, pages 42–47, Hong Kong, China. Association for Computational Linguistics.
  5. A discourse-aware attention model for abstractive summarization of long documents. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 615–621, New Orleans, Louisiana. Association for Computational Linguistics.
  6. Lea Frermann and Alexandre Klementiev. 2019. Inducing document structure for aspect-based summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6263–6273.
  7. Supert: Towards new frontiers in unsupervised evaluation metrics for multi-document summarization. arXiv preprint arXiv:2005.03724.
  8. Maarten Grootendorst. 2022. Bertopic: Neural topic modeling with a class-based tf-idf procedure. arXiv preprint arXiv:2203.05794.
  9. Longt5: Efficient text-to-text transformer for long sequences. arXiv preprint arXiv:2112.07916.
  10. Teaching machines to read and comprehend. In Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 1, NIPS’15, page 1693–1701, Cambridge, MA, USA. MIT Press.
  11. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
  12. Review summary generation in online systems: Frameworks for supervised and unsupervised scenarios. ACM Trans. Web, 15(3).
  13. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461.
  14. Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
  15. Generating wikipedia by summarizing long sequences. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net.
  16. hdbscan: Hierarchical density based clustering. The Journal of Open Source Software, 2(11):205.
  17. Umap: Uniform manifold approximation and projection. The Journal of Open Source Software, 3(29):861.
  18. Naresh Kumar Nagwani. 2015. Summarizing large text collection using topic modeling and clustering based on mapreduce framework. Journal of Big Data, 2(1):1–18.
  19. Don’t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization. arXiv preprint arXiv:1808.08745.
  20. Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence embeddings using Siamese BERT-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3982–3992, Hong Kong, China. Association for Computational Linguistics.
  21. A neural attention model for abstractive sentence summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 379–389, Lisbon, Portugal. Association for Computational Linguistics.
  22. A simple and effective self-supervised contrastive learning framework for aspect detection. In Proceedings of the AAAI conference on artificial intelligence, volume 15, pages 13815–13824.
  23. Summarizing text on any aspects: A knowledge-informed weakly-supervised approach. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6301–6309, Online. Association for Computational Linguistics.
  24. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers.
  25. S2vntm: Semi-supervised vmf neural topic modeling. arXiv preprint arXiv:2307.04804.
  26. DeTiME: Diffusion-enhanced topic modeling using encoder-decoder based LLM. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 9040–9057, Singapore. Association for Computational Linguistics.
  27. HR-MultiWOZ: A task oriented dialogue (TOD) dataset for HR LLM agent. In Proceedings of the First Workshop on Natural Language Processing for Human Resources (NLP4HR 2024), pages 59–72, St. Julian’s, Malta. Association for Computational Linguistics.
  28. Kdstm: Neural semi-supervised topic modeling with knowledge distillation. arXiv preprint arXiv:2307.01878.
  29. vONTSS: vMF based semi-supervised neural topic modeling with optimal transport. In Findings of the Association for Computational Linguistics: ACL 2023, pages 4433–4457, Toronto, Canada. Association for Computational Linguistics.
  30. Targeted feedback generation for constructed-response questions. aaai.
  31. End-to-end question generation to assist formative assessment design for conceptual knowledge learning. amazon.science.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Xiaobo Guo (32 papers)
  2. Jay Desai (11 papers)
  3. Srinivasan H. Sengamedu (10 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets