Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low-Resource Court Judgment Summarization for Common Law Systems (2403.04454v1)

Published 7 Mar 2024 in cs.CL and cs.AI

Abstract: Common law courts need to refer to similar precedents' judgments to inform their current decisions. Generating high-quality summaries of court judgment documents can facilitate legal practitioners to efficiently review previous cases and assist the general public in accessing how the courts operate and how the law is applied. Previous court judgment summarization research focuses on civil law or a particular jurisdiction's judgments. However, judges can refer to the judgments from all common law jurisdictions. Current summarization datasets are insufficient to satisfy the demands of summarizing precedents across multiple jurisdictions, especially when labeled data are scarce for many jurisdictions. To address the lack of datasets, we present CLSum, the first dataset for summarizing multi-jurisdictional common law court judgment documents. Besides, this is the first court judgment summarization work adopting LLMs in data augmentation, summary generation, and evaluation. Specifically, we design an LLM-based data augmentation method incorporating legal knowledge. We also propose a legal knowledge enhanced evaluation metric based on LLM to assess the quality of generated judgment summaries. Our experimental results verify that the LLM-based summarization methods can perform well in the few-shot and zero-shot settings. Our LLM-based data augmentation method can mitigate the impact of low data resources. Furthermore, we carry out comprehensive comparative experiments to find essential model components and settings that are capable of enhancing summarization performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (55)
  1. HKSAR DOJ. The common law and rules of equity. URL www.doj.gov.hk/en/our_legal_system/the_common_law.html.
  2. West’s encyclopedia of American law, volume 6. 2005.
  3. A comparative study of summarization algorithms applied to legal case judgments. In Proc. ECIR, Part I 41, pages 413–428. Springer, 2019.
  4. Text summarization from legal documents: a survey. Artificial Intelligence Review, 51:371–402, 2019.
  5. Diego de Vargas Feijó and Viviane Pereira Moreira. Rulingbr: A summarization dataset for legal texts. In Proc. PROPOR, pages 255–264. Springer, 2018.
  6. Summarization of german court rulings. In Proc. NLLP, pages 180–189, 2021.
  7. Long document summarization in a low resource setting using pretrained language models. In Proc. ACL-SRW, pages 71–80, 2021.
  8. Legal case document summarization: Extractive and abstractive methods and their evaluation. In Proc. AACL-IJCNLP, pages 1048–1064, 2022.
  9. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150, 2020.
  10. Flashattention: Fast and memory-efficient exact attention with io-awareness. Proc. NeurIPS, 35:16344–16359, 2022.
  11. Llm. int8 (): 8-bit matrix multiplication for transformers at scale. arXiv preprint arXiv:2208.07339, 2022a.
  12. Understanding int4 quantization for transformer models: Latency speedup, composability, and failure cases. arXiv preprint arXiv:2301.12017, 2023.
  13. 8-bit optimizers via block-wise quantization. Proc. ICLR, 2022b.
  14. Zero: Memory optimizations toward training trillion parameter models. In Proc. SC, pages 1–16. IEEE, 2020.
  15. Qlora: Efficient finetuning of quantized llms. arXiv preprint arXiv:2305.14314, 2023.
  16. Lora: Low-rank adaptation of large language models. In Proc. ICLR, 2021.
  17. Long text and multi-table summarization: Dataset and method. In Findings of EMNLP 2022, pages 1995–2010, 2022a.
  18. Neural abstractive summarization for long text and multiple tables. IEEE Transactions on Knowledge and Data Engineering, 2023.
  19. Newsroom: A dataset of 1.3 million summaries with diverse extractive strategies. In Proc. NAACL-HLT, pages 708–719, 2018.
  20. Movie review mining and summarization. In Proc. CIKM, pages 43–50, 2006.
  21. Multi-news: A large-scale multi-document summarization dataset and abstractive hierarchical model. In Proc. ACL, pages 1074–1084, 2019.
  22. Highlight-transformer: Leveraging key phrase aware attention to improve abstractive multi-document summarization. In Findings of ACL, pages 5021–5027, 2021.
  23. Generating a structured summary of numerous academic papers: Dataset and method. In Proc. IJCAI, pages 4259–4265, 2022b.
  24. Efficient attentions for long document summarization. In Proc. NAACL-HLT, pages 1419–1436, 2021.
  25. Legal systems in india: Overview. URL uk.practicallaw.thomsonreuters.com/w-017-5278.
  26. Letsum, an automatic text summarization system in law field. JURIX, 2004.
  27. Casesummarizer: a system for automated summarization of legal texts. In Proc. COLING: System Demonstrations, pages 258–262, 2016.
  28. Improving legal document summarization using graphical models. FAIA, 152:51, 2006.
  29. Diego de Vargas Feijo and Viviane P Moreira. Improving abstractive summarization of legal rulings through textual entailment. Artificial intelligence and law, 31(1):91–113, 2023.
  30. Lexrank: Graph-based lexical centrality as salience in text summarization. JAIR, 22:457–479, 2004.
  31. Textrank: Bringing order into text. In Proc. EMNLP, pages 404–411, 2004.
  32. Mrc-sum: An mrc framework for extractive summarization of academic articles in natural sciences and medicine. Information Processing & Management, 60(5):103467, 2023.
  33. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proc. ICML, pages 11328–11339. PMLR, 2020.
  34. Key phrase aware transformer for abstractive summarization. Information Processing & Management, 59(3):102913, 2022c.
  35. Term-weighting approaches in automatic text retrieval. Information processing & management, 24(5):513–523, 1988.
  36. R Rehurek. Software framework for topic modelling with large corpora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks, page 45, 2010.
  37. Exploring text links for coherent multi-document summarization. In Proc. COLING: Technical Papers, pages 213–223, 2016.
  38. Recent advances in document summarization. Knowledge and Information Systems, 53(2):297–336, 2017.
  39. Attention is all you need. In Proc. NeurIPS, pages 5998–6008, 2017.
  40. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proc. NAACL-HLT, Volume 1 (Long and Short Papers), pages 4171–4186, 2019.
  41. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proc. ACL, pages 7871–7880, 2020.
  42. Improving neural machine translation models with monolingual data. In Proc. ACL, pages 86–96, 2016.
  43. arXiv preprint arXiv:2302.13007, 2023.
  44. LongT5: Efficient text-to-text transformer for long sequences. In Findings of NAACL, pages 724–736, 2022.
  45. Exploring the limits of transfer learning with a unified text-to-text transformer. JMLR, 21(140):1–67, 2020.
  46. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100, 2022.
  47. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  48. Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. lmsys.org/blog/2023-03-30-vicuna/, 2023.
  49. OpenAI. Introducing chatgpt. openai.com/blog/chatgpt, 2022.
  50. Adam: A method for stochastic optimization. CoRR, abs/1412.6980, 2015.
  51. Adafactor: Adaptive learning rates with sublinear memory cost. In Proc. ICML, pages 4596–4604. PMLR, 2018.
  52. Transformers: State-of-the-art natural language processing. In EMNLP, pages 38–45, 2020.
  53. Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
  54. Bartscore: Evaluating generated text as text generation. Proc. NeurIPS, 34:27263–27277, 2021.
  55. Joseph L Fleiss. Measuring nominal scale agreement among many raters. Psychological bulletin, 76(5):378, 1971.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Shuaiqi Liu (12 papers)
  2. Jiannong Cao (73 papers)
  3. Yicong Li (34 papers)
  4. Ruosong Yang (8 papers)
  5. Zhiyuan Wen (11 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.