Investigating Numeracy Learning Ability of a Text-to-Text Transfer Model

09/10/2021
by   Kuntal Kumar Pal, et al.
17

The transformer-based pre-trained language models have been tremendously successful in most of the conventional NLP tasks. But they often struggle in those tasks where numerical understanding is required. Some possible reasons can be the tokenizers and pre-training objectives which are not specifically designed to learn and preserve numeracy. Here we investigate the ability of text-to-text transfer learning model (T5), which has outperformed its predecessors in the conventional NLP tasks, to learn numeracy. We consider four numeracy tasks: numeration, magnitude order prediction, finding minimum and maximum in a series, and sorting. We find that, although T5 models perform reasonably well in the interpolation setting, they struggle considerably in the extrapolation setting across all four tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2023

Pre-Training to Learn in Context

In-context learning, where pre-trained language models learn to perform ...
research
10/03/2022

Characterization of effects of transfer learning across domains and languages

With ever-expanding datasets of domains, tasks and languages, transfer l...
research
10/23/2019

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

Transfer learning, where a model is first pre-trained on a data-rich tas...
research
06/20/2023

Exploring New Frontiers in Agricultural NLP: Investigating the Potential of Large Language Models for Food Applications

This paper explores new frontiers in agricultural natural language proce...
research
01/10/2021

Transfer Learning and Augmentation for Word Sense Disambiguation

Many downstream NLP tasks have shown significant improvement through con...
research
02/03/2021

Neural Transfer Learning with Transformers for Social Science Text Analysis

During the last years, there have been substantial increases in the pred...
research
12/08/2022

Investigating Glyph Phonetic Information for Chinese Spell Checking: What Works and What's Next

While pre-trained Chinese language models have demonstrated impressive p...

Please sign up or login with your details

Forgot password? Click here to reset