Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LawGPT: A Chinese Legal Knowledge-Enhanced Large Language Model (2406.04614v1)

Published 7 Jun 2024 in cs.CL and cs.AI

Abstract: LLMs, including both proprietary and open-source models, have showcased remarkable capabilities in addressing a wide range of downstream tasks. Nonetheless, when it comes to practical Chinese legal tasks, these models fail to meet the actual requirements. Proprietary models do not ensure data privacy for sensitive legal cases, while open-source models demonstrate unsatisfactory performance due to their lack of legal knowledge. To address this problem, we introduce LawGPT, the first open-source model specifically designed for Chinese legal applications. LawGPT comprises two key components: legal-oriented pre-training and legal supervised fine-tuning. Specifically, we employ large-scale Chinese legal documents for legal-oriented pre-training to incorporate legal domain knowledge. To further improve the model's performance on downstream legal tasks, we create a knowledge-driven instruction dataset for legal supervised fine-tuning. Our experimental results demonstrate that LawGPT outperforms the open-source LLaMA 7B model. Our code and resources are publicly available at https://github.com/pengxiao-song/LaWGPT and have received 5.7K stars on GitHub.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhi Zhou (135 papers)
  2. Jiang-Xin Shi (13 papers)
  3. Peng-Xiao Song (3 papers)
  4. Xiao-Wen Yang (7 papers)
  5. Yi-Xuan Jin (3 papers)
  6. Lan-Zhe Guo (33 papers)
  7. Yu-Feng Li (96 papers)
Citations (8)
X Twitter Logo Streamline Icon: https://streamlinehq.com