DeepAI AI Chat
Log In Sign Up

On the Fairness of 'Fake' Data in Legal AI

by   Lauren Boswell, et al.
The University of Sydney

The economics of smaller budgets and larger case numbers necessitates the use of AI in legal proceedings. We examine the concept of disparate impact and how biases in the training data lead to the search for fairer AI. This paper seeks to begin the discourse on what such an implementation would actually look like with a criticism of pre-processing methods in a legal context . We outline how pre-processing is used to correct biased data and then examine the legal implications of effectively changing cases in order to achieve a fairer outcome including the black box problem and the slow encroachment on legal precedent. Finally we present recommendations on how to avoid the pitfalls of pre-processed data with methods that either modify the classifier or correct the output in the final step.


page 1

page 2

page 3

page 4


Human Indignity: From Legal AI Personhood to Selfish Memes

It is possible to rely on current corporate law to grant legal personhoo...

Legal Judgment Prediction (LJP) Amid the Advent of Autonomous AI Legal Reasoning

Legal Judgment Prediction (LJP) is a longstanding and open topic in the ...

Are Models Trained on Indian Legal Data Fair?

Recent advances and applications of language technology and artificial i...

An Uncommon Task: Participatory Design in Legal AI

Despite growing calls for participation in AI design, there are to date ...

THUIR@COLIEE 2023: Incorporating Structural Knowledge into Pre-trained Language Models for Legal Case Retrieval

Legal case retrieval techniques play an essential role in modern intelli...

Exploring the psychology of GPT-4's Moral and Legal Reasoning

Large language models have been used as the foundation of highly sophist...

PRESISTANT: Learning based assistant for data pre-processing

Data pre-processing is one of the most time consuming and relevant steps...