Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Legal Documents Drafting with Fine-Tuned Pre-Trained Large Language Model (2406.04202v1)

Published 6 Jun 2024 in cs.CL and cs.AI

Abstract: With the development of large-scale LLMs (LLM), fine-tuning pre-trained LLM has become a mainstream paradigm for solving downstream tasks of natural language processing. However, training a LLM in the legal field requires a large number of legal documents so that the LLM can learn legal terminology and the particularity of the format of legal documents. The typical NLP approaches usually rely on many manually annotated data sets for training. However, in the legal field application, it is difficult to obtain a large number of manually annotated data sets, which restricts the typical method applied to the task of drafting legal documents. The experimental results of this paper show that not only can we leverage a large number of annotation-free legal documents without Chinese word segmentation to fine-tune a large-scale LLM, but more importantly, it can fine-tune a pre-trained LLM on the local computer to achieve the generating legal document drafts task, and at the same time achieve the protection of information privacy and to improve information security issues.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Chun-Hsien Lin (4 papers)
  2. Pu-Jen Cheng (12 papers)
Citations (3)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets