Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
96 tokens/sec
Gemini 2.5 Pro Premium
42 tokens/sec
GPT-5 Medium
20 tokens/sec
GPT-5 High Premium
27 tokens/sec
GPT-4o
100 tokens/sec
DeepSeek R1 via Azure Premium
86 tokens/sec
GPT OSS 120B via Groq Premium
464 tokens/sec
Kimi K2 via Groq Premium
181 tokens/sec
2000 character limit reached

Bonafide at LegalLens 2024 Shared Task: Using Lightweight DeBERTa Based Encoder For Legal Violation Detection and Resolution (2410.22977v1)

Published 30 Oct 2024 in cs.CL

Abstract: In this work, we present two systems -- Named Entity Resolution (NER) and Natural Language Inference (NLI) -- for detecting legal violations within unstructured textual data and for associating these violations with potentially affected individuals, respectively. Both these systems are lightweight DeBERTa based encoders that outperform the LLM baselines. The proposed NER system achieved an F1 score of 60.01\% on Subtask A of the LegalLens challenge, which focuses on identifying violations. The proposed NLI system achieved an F1 score of 84.73\% on Subtask B of the LegalLens challenge, which focuses on resolving these violations by matching them with pre-existing legal complaints of class action cases. Our NER system ranked sixth and NLI system ranked fifth on the LegalLens leaderboard. We release the trained models and inference scripts.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Falcon-40b: an open large language model with state-of-the-art performance.
  2. Legallens: Leveraging llms for legal violation identification in unstructured text. arXiv preprint arXiv:2402.04335.
  3. An algorithm that learns what’s in a name. Machine learning, 34:211–231.
  4. Nyu: Description of the mene named entity system as used in muc-7. In Seventh Message Understanding Conference (MUC-7): Proceedings of a Conference Held in Fairfax, Virginia, April 29-May 1, 1998.
  5. A large annotated corpus for learning natural language inference. arXiv preprint arXiv:1508.05326.
  6. Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
  7. Legal-bert: The muppets straight out of law school. arXiv preprint arXiv:2010.02559.
  8. Xnli: Evaluating cross-lingual sentence representations. arXiv preprint arXiv:1809.05053.
  9. Xiang Dai. 2018. Recognizing complex entity mentions: A review and future directions. In Proceedings of ACL 2018, Student Research Workshop, pages 37–44.
  10. Jacob Devlin. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  11. Wee Chung Gan and Hwee Tou Ng. 2019. Improving the robustness of question answering systems to question paraphrasing. In Proceedings of the 57th annual meeting of the association for computational linguistics, pages 6065–6075.
  12. The third pascal recognizing textual entailment challenge. In Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing, pages 1–9.
  13. Legallens shared task 2024: Legal violation identification in unstructured text. Preprint, arXiv:2410.12064.
  14. Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing. Preprint, arXiv:2111.09543.
  15. Generate, annotate, and learn: Generative models advance self-training and knowledge distillation. arXiv:2106.06168.
  16. A synthetic data approach for domain generalization of nli models. arXiv preprint arXiv:2402.12368.
  17. Bidirectional lstm-crf models for sequence tagging. arXiv preprint arXiv:1508.01991.
  18. AntContentTech at SemEval-2023 task 6: Domain-adaptive pretraining and auxiliary-task learning for understanding Indian legal texts. In Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), pages 402–408, Toronto, Canada. Association for Computational Linguistics.
  19. Mixtral of experts. arXiv preprint arXiv:2401.04088.
  20. Named entity recognition in indian court judgments. arXiv preprint arXiv:2211.03442.
  21. Neural architectures for named entity recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 260–270, San Diego, California. Association for Computational Linguistics.
  22. Yinhan Liu. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
  23. Xuezhe Ma and Eduard Hovy. 2016. End-to-end sequence labeling via bi-directional LSTM-CNNs-CRF. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1064–1074, Berlin, Germany. Association for Computational Linguistics.
  24. Andrew McCallum and Wei Li. 2003. Early results for named entity recognition with conditional random fields, feature induction and web-enhanced lexicons. In Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003, pages 188–191.
  25. Semeval 2023 task 6: Legaleval-understanding legal texts. arXiv preprint arXiv:2304.09548.
  26. Researchteam_hcn at semeval-2023 task 6: A knowledge enhanced transformers based legal nlp system. In Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), pages 1245–1253.
  27. Damien Sileo. 2024. tasksource: A large collection of NLP tasks with a structured dataset preprocessing framework. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 15655–15684, Torino, Italia. ELRA and ICCL.
  28. Attention is all you need.(nips), 2017. Advances in Neural Information Processing Systems, 10:S0140525X16001837.
  29. A broad-coverage challenge corpus for sentence understanding through inference. arXiv preprint arXiv:1704.05426.
  30. Zhilin Yang. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. arXiv preprint arXiv:1906.08237.
  31. Gliner: Generalist model for named entity recognition using bidirectional transformer. Preprint, arXiv:2311.08526.
  32. Universalner: Targeted distillation from large language models for open named entity recognition. arXiv preprint arXiv:2308.03279.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (1)

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets