LexGPT 0.1: pre-trained GPT-J models with Pile of Law

06/05/2023
by   Jieh-Sheng Lee, et al.
0

This research aims to build generative language models specialized for the legal domain. The manuscript presents the development of LexGPT models based on GPT-J models and pre-trained with Pile of Law. The foundation model built in this manuscript is the initial step for the development of future applications in the legal domain, such as further training with reinforcement learning from human feedback. Another objective of this manuscript is to assist legal professionals in utilizing language models through the “No Code” approach. By fine-tuning models with specialized data and without modifying any source code, legal professionals can create custom language models for downstream tasks with minimum effort and technical knowledge. The downstream task in this manuscript is to turn a LexGPT model into a classifier, although the performance is notably lower than the state-of-the-art result. How to enhance downstream task performance without modifying the model or its source code is a research topic for future exploration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2023

LeXFiles and LegalLAMA: Facilitating English Multinational Legal Language Model Development

In this work, we conduct a detailed analysis on the performance of legal...
research
06/23/2022

Evaluating Generative Patent Language Models

This research aims to build generative language models in the patent dom...
research
03/28/2023

Foundation Models and Fair Use

Existing foundation models are trained on copyrighted material. Deployin...
research
05/25/2023

Tuning Models of Code with Compiler-Generated Reinforcement Learning Feedback

Large Language Models (LLMs) pre-trained on code have recently emerged a...
research
06/24/2023

Can GPT-4 Support Analysis of Textual Data in Tasks Requiring Highly Specialized Domain Expertise?

We evaluated the capability of generative pre-trained transformers (GPT-...
research
01/04/2023

Extending Source Code Pre-Trained Language Models to Summarise Decompiled Binaries

Reverse engineering binaries is required to understand and analyse progr...
research
12/02/2021

Unsupervised Law Article Mining based on Deep Pre-Trained Language Representation Models with Application to the Italian Civil Code

Modeling law search and retrieval as prediction problems has recently em...

Please sign up or login with your details

Forgot password? Click here to reset