Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Personalized Abstractive Summarization by Tri-agent Generation Pipeline (2305.02483v2)

Published 4 May 2023 in cs.CL

Abstract: Tailoring outputs from LLMs, like ChatGPT, to implicit user preferences remains a challenge despite their impressive generative capabilities. In this paper, we propose a tri-agent generation pipeline comprising a generator, an instructor, and an editor to enhance output personalization. The generator produces an initial output, the instructor automatically generates editing instructions based on user preferences, and the editor refines the output to align with those preferences. The inference-only LLM (ChatGPT) serves as both the generator and editor, with a smaller model acting as the instructor to guide output generation. We train the instructor using editor-steered reinforcement learning, leveraging feedback from a large-scale editor model to optimize instruction generation. Experimental results on two abstractive summarization datasets demonstrate the effectiveness of our approach in generating outputs that better meet user expectations. Code is available at \url{https://github.com/Wendy-Xiao/chatgpt_editing_summ}

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Language models are few-shot learners. CoRR, abs/2005.14165.
  2. Sparks of artificial general intelligence: Early experiments with gpt-4.
  3. Palm: Scaling language modeling with pathways.
  4. Scaling instruction-finetuned language models.
  5. QAFactEval: Improved QA-based factual consistency evaluation for summarization. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2587–2601, Seattle, United States. Association for Computational Linguistics.
  6. Text editing by command. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5259–5274, Online. Association for Computational Linguistics.
  7. Tanya Goyal and Greg Durrett. 2021. Annotating and modeling fine-grained factuality in summarization. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
  8. News summarization and evaluation in the era of gpt-3.
  9. Levenshtein transformer. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  10. Z-code++: A pre-trained language model optimized for abstractive summarization.
  11. Second thoughts are best: Learning to re-align with human values from text edits. In Advances in Neural Information Processing Systems, volume 35, pages 181–196. Curran Associates, Inc.
  12. On improving summarization factual consistency from natural language feedback.
  13. Self-refine: Iterative refinement with self-feedback.
  14. Encode, tag, realize: High-precision text editing. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5054–5065, Hong Kong, China. Association for Computational Linguistics.
  15. Abstractive text summarization using sequence-to-sequence RNNs and beyond. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, pages 280–290, Berlin, Germany. Association for Computational Linguistics.
  16. Training language models to follow instructions with human feedback.
  17. Is reinforcement learning (not) for natural language processing: Benchmarks, baselines, and building blocks for natural language policy optimization.
  18. Machel Reid and Graham Neubig. 2022. Learning to model editing processes. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 3822–3832, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  19. Machel Reid and Victor Zhong. 2021. LEWIS: Levenshtein editing for unsupervised text style transfer. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 3932–3944, Online. Association for Computational Linguistics.
  20. Multitask prompted training enables zero-shot task generalization. CoRR, abs/2110.08207.
  21. Newsedits: A news article revision dataset and a document-level reasoning challenge.
  22. Lamda: Language models for dialog applications. CoRR, abs/2201.08239.
  23. Llama: Open and efficient foundation language models.
  24. Self-instruct: Aligning language model with self generated instructions.
  25. Finetuned language models are zero-shot learners. CoRR, abs/2109.01652.
  26. Generating sequences by learning to self-correct.
  27. PRIMER: pyramid-based masked sentence pre-training for multi-document summarization. CoRR, abs/2110.08499.
  28. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org.
  29. Opt: Open pre-trained transformer language models.
  30. Benchmarking large language models for news summarization.
Citations (3)

Summary

We haven't generated a summary for this paper yet.