CMV-BERT: Contrastive multi-vocab pretraining of BERT

12/29/2020
by   Wei Zhu, et al.
0

In this work, we represent CMV-BERT, which improves the pretraining of a language model via two ingredients: (a) contrastive learning, which is well studied in the area of computer vision; (b) multiple vocabularies, one of which is fine-grained and the other is coarse-grained. The two methods both provide different views of an original sentence, and both are shown to be beneficial. Downstream tasks demonstrate our proposed CMV-BERT are effective in improving the pretrained language models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/16/2020

CERT: Contrastive Self-supervised Learning for Language Understanding

Pretrained language models such as BERT, GPT have shown great effectiven...
research
04/15/2021

How to Train BERT with an Academic Budget

While large language models à la BERT are used ubiquitously in NLP, pret...
research
05/13/2021

Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level

Larger language models have higher accuracy on average, but are they bet...
research
06/30/2021

The MultiBERTs: BERT Reproductions for Robustness Analysis

Experiments with pretrained models such as BERT are often based on a sin...
research
06/30/2023

Ticket-BERT: Labeling Incident Management Tickets with Language Models

An essential aspect of prioritizing incident tickets for resolution is e...
research
04/11/2022

Tokenwise Contrastive Pretraining for Finer Speech-to-BERT Alignment in End-to-End Speech-to-Intent Systems

Recent advances in End-to-End (E2E) Spoken Language Understanding (SLU) ...
research
12/10/2021

Tradeoffs Between Contrastive and Supervised Learning: An Empirical Study

Contrastive learning has made considerable progress in computer vision, ...

Please sign up or login with your details

Forgot password? Click here to reset