Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LexGPT 0.1: pre-trained GPT-J models with Pile of Law (2306.05431v1)

Published 5 Jun 2023 in cs.CL

Abstract: This research aims to build generative LLMs specialized for the legal domain. The manuscript presents the development of LexGPT models based on GPT-J models and pre-trained with Pile of Law. The foundation model built in this manuscript is the initial step for the development of future applications in the legal domain, such as further training with reinforcement learning from human feedback. Another objective of this manuscript is to assist legal professionals in utilizing LLMs through the ``No Code'' approach. By fine-tuning models with specialized data and without modifying any source code, legal professionals can create custom LLMs for downstream tasks with minimum effort and technical knowledge. The downstream task in this manuscript is to turn a LexGPT model into a classifier, although the performance is notably lower than the state-of-the-art result. How to enhance downstream task performance without modifying the model or its source code is a research topic for future exploration.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Jieh-Sheng Lee (10 papers)
Citations (8)