OpenFlamingo: An Open-Source Framework for Training Large Autoregressive Vision-Language Models

08/02/2023
by   Anas Awadalla, et al.
0

We introduce OpenFlamingo, a family of autoregressive vision-language models ranging from 3B to 9B parameters. OpenFlamingo is an ongoing effort to produce an open-source replication of DeepMind's Flamingo models. On seven vision-language datasets, OpenFlamingo models average between 80 - 89 corresponding Flamingo performance. This technical report describes our models, training data, hyperparameters, and evaluation suite. We share our models and code at https://github.com/mlfoundations/open_flamingo.

READ FULL TEXT

page 2

page 4

page 16

page 17

research
04/14/2022

GPT-NeoX-20B: An Open-Source Autoregressive Language Model

We introduce GPT-NeoX-20B, a 20 billion parameter autoregressive languag...
research
03/31/2022

Scaling Up Models and Data with and

Recent neural network-based language models have benefited greatly from ...
research
05/15/2023

Memorization for Good: Encryption with Autoregressive Language Models

Over-parameterized neural language models (LMs) can memorize and recite ...
research
04/03/2023

Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling

How do large language models (LLMs) develop and evolve over the course o...
research
06/28/2023

Towards Language Models That Can See: Computer Vision Through the LENS of Natural Language

We propose LENS, a modular approach for tackling computer vision problem...
research
07/17/2023

Mini-Giants: "Small" Language Models and Open Source Win-Win

ChatGPT is phenomenal. However, it is prohibitively expensive to train a...
research
07/07/2023

QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models

We present ongoing work on a new automatic code generation approach for ...

Please sign up or login with your details

Forgot password? Click here to reset