Understanding BLOOM: An empirical study on diverse NLP tasks

11/27/2022
by   Parag Pravin Dakle, et al.
0

In this work, we present an evaluation of smaller BLOOM model variants (350m/560m and 1b3/1b7) on various natural language processing tasks. This includes GLUE - language understanding, prompt-based zero-shot and few-shot text classification and extraction, question answering, prompt-based text generation, and multi-lingual text classification to understand model strengths/weaknesses and behavior. Empirical results show that BLOOM variants under-perform on all GLUE tasks (except WNLI), question-answering, and text generation. The variants bloom for WNLI, with an accuracy of 56.3 prompt-based few-shot text extraction on MIT Movies and ATIS datasets. The BLOOM variants on average have 7 models on Director and Airline Name extraction from MIT Movies and ATIS datasets, respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/16/2022

ZeroGen: Efficient Zero-shot Learning via Dataset Generation

There is a growing interest in dataset generation recently due to the su...
research
04/19/2019

Unifying Question Answering and Text Classification via Span Extraction

Even as pre-trained language encoders such as BERT are shared across man...
research
05/18/2022

Evaluation of Transfer Learning for Polish with a Text-to-Text Model

We introduce a new benchmark for assessing the quality of text-to-text m...
research
04/16/2021

An Empirical Study of Extrapolation in Text Generation with Scalar Control

We conduct an empirical evaluation of extrapolation performance when con...
research
07/06/2023

Text Alignment Is An Efficient Unified Model for Massive NLP Tasks

Large language models (LLMs), typically designed as a function of next-w...
research
05/13/2023

On the Hidden Mystery of OCR in Large Multimodal Models

Large models have recently played a dominant role in natural language pr...
research
07/17/2016

An Empirical Evaluation of various Deep Learning Architectures for Bi-Sequence Classification Tasks

Several tasks in argumentation mining and debating, question-answering, ...

Please sign up or login with your details

Forgot password? Click here to reset