FootGPT : A Large Language Model Development Experiment on a Minimal Setting

08/16/2023
by   Eren Unlu, et al.
0

With recent empirical observations, it has been argued that the most significant aspect of developing accurate language models may be the proper dataset content and training strategy compared to the number of neural parameters, training duration or dataset size. Following this argument, we opted to fine tune a one billion parameter size trained general purpose causal language model with a dataset curated on team statistics of the Italian football league first ten game weeks, using low rank adaptation. The limited training dataset was compiled based on a framework where a powerful commercial large language model provides distilled paragraphs and question answer pairs as intended. The training duration was kept relatively short to provide a basis for our minimal setting exploration. We share our key observations on the process related to developing a specific purpose language model which is intended to interpret soccer data with constrained resources in this article.

READ FULL TEXT

page 3

page 4

page 5

page 6

page 7

page 8

page 9

research
11/09/2021

FPM: A Collection of Large-scale Foundation Pre-trained Language Models

Recent work in language modeling has shown that training large-scale Tra...
research
03/30/2023

BloombergGPT: A Large Language Model for Finance

The use of NLP in the realm of financial technology is broad and complex...
research
04/25/2022

Super-Prompting: Utilizing Model-Independent Contextual Data to Reduce Data Annotation Required in Visual Commonsense Tasks

Pre-trained language models have shown excellent results in few-shot lea...
research
09/17/2023

Performance of the Pre-Trained Large Language Model GPT-4 on Automated Short Answer Grading

Automated Short Answer Grading (ASAG) has been an active area of machine...
research
04/19/2023

A Theory on Adam Instability in Large-Scale Machine Learning

We present a theory for the previously unexplained divergent behavior no...
research
05/01/2020

Selecting Informative Contexts Improves Language Model Finetuning

We present a general finetuning meta-method that we call information gai...
research
05/22/2023

GPT-SW3: An Autoregressive Language Model for the Nordic Languages

This paper details the process of developing the first native large gene...

Please sign up or login with your details

Forgot password? Click here to reset