Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators

07/08/2023
by   Andreas Liesenfeld, et al.
0

Large language models that exhibit instruction-following behaviour represent one of the biggest recent upheavals in conversational interfaces, a trend in large part fuelled by the release of OpenAI's ChatGPT, a proprietary large language model for text generation fine-tuned through reinforcement learning from human feedback (LLM+RLHF). We review the risks of relying on proprietary software and survey the first crop of open-source projects of comparable architecture and functionality. The main contribution of this paper is to show that openness is differentiated, and to offer scientific documentation of degrees of openness in this fast-moving field. We evaluate projects in terms of openness of code, training data, model weights, RLHF data, licensing, scientific documentation, and access methods. We find that while there is a fast-growing list of projects billing themselves as 'open source', many inherit undocumented data of dubious legality, few share the all-important instruction-tuning (a key site where human annotation labour is involved), and careful scientific documentation is exceedingly rare. Degrees of openness are relevant to fairness and accountability at all points, from data collection and curation to model architecture, and from training and fine-tuning to release and deployment.

READ FULL TEXT
research
07/29/2023

Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback

A key technology for the development of large language models (LLMs) inv...
research
08/01/2023

Instructed to Bias: Instruction-Tuned Language Models Exhibit Emergent Cognitive Bias

Recent studies show that instruction tuning and learning from human feed...
research
08/25/2023

DARWIN Series: Domain Specific Large Language Models for Natural Science

Emerging tools bring forth fresh approaches to work, and the field of na...
research
07/05/2023

Flacuna: Unleashing the Problem Solving Power of Vicuna using FLAN Fine-Tuning

Recently, the release of INSTRUCTEVAL has provided valuable insights int...
research
08/25/2023

SoTaNa: The Open-Source Software Development Assistant

Software development plays a crucial role in driving innovation and effi...
research
04/14/2023

OpenAssistant Conversations – Democratizing Large Language Model Alignment

Aligning large language models (LLMs) with human preferences has proven ...
research
12/20/2022

Is GPT-3 a Psychopath? Evaluating Large Language Models from a Psychological Perspective

Are large language models (LLMs) like GPT-3 psychologically safe? In thi...

Please sign up or login with your details

Forgot password? Click here to reset