ToddlerBERTa: Exploiting BabyBERTa for Grammar Learning and Language Understanding

08/30/2023
by   Omer Veysel Cagatan, et al.
0

We present ToddlerBERTa, a BabyBERTa-like language model, exploring its capabilities through five different models with varied hyperparameters. Evaluating on BLiMP, SuperGLUE, MSGS, and a Supplement benchmark from the BabyLM challenge, we find that smaller models can excel in specific tasks, while larger models perform well with substantial data. Despite training on a smaller dataset, ToddlerBERTa demonstrates commendable performance, rivalling the state-of-the-art RoBERTa-base. The model showcases robust language understanding, even with single-sentence pretraining, and competes with baselines that leverage broader contextual information. Our work provides insights into hyperparameter choices, and data utilization, contributing to the advancement of language models.

READ FULL TEXT
research
12/03/2021

The Catalan Language CLUB

The Catalan Language Understanding Benchmark (CLUB) encompasses various ...
research
01/12/2022

How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets

A central question in natural language understanding (NLU) research is w...
research
07/19/2022

Analyzing Bagging Methods for Language Models

Modern language models leverage increasingly large numbers of parameters...
research
11/02/2018

Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks

Pretraining with language modeling and related unsupervised tasks has re...
research
05/22/2023

Can ChatGPT Detect Intent? Evaluating Large Language Models for Spoken Language Understanding

Recently, large pretrained language models have demonstrated strong lang...
research
01/24/2021

WangchanBERTa: Pretraining transformer-based Thai Language Models

Transformer-based language models, more specifically BERT-based architec...
research
07/26/2019

RoBERTa: A Robustly Optimized BERT Pretraining Approach

Language model pretraining has led to significant performance gains but ...

Please sign up or login with your details

Forgot password? Click here to reset