Learned in Translation: Contextualized Word Vectors

08/01/2017
by   Bryan McCann, et al.
0

Computer vision has benefited from initializing multiple deep layers with weights pretrained on large supervised training sets like ImageNet. Natural language processing (NLP) typically sees initialization of only the lowest layer of deep models with pretrained word vectors. In this paper, we use a deep LSTM encoder from an attentional sequence-to-sequence model trained for machine translation (MT) to contextualize word vectors. We show that adding these context vectors (CoVe) improves performance over using only unsupervised word and character vectors on a wide variety of common NLP tasks: sentiment analysis (SST, IMDb), question classification (TREC), entailment (SNLI), and question answering (SQuAD). For fine-grained sentiment analysis and entailment, CoVe improves performance of our baseline models to the state of the art.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/19/2020

Towards Computational Linguistics in Minangkabau Language: Studies on Sentiment Analysis and Machine Translation

Although some linguists (Rusmali et al., 1985; Crouch, 2009) have fairly...
research
02/15/2018

Deep contextualized word representations

We introduce a new type of deep contextualized word representation that ...
research
12/02/2019

EduBERT: Pretrained Deep Language Models for Learning Analytics

The use of large pretrained neural networks to create contextualized wor...
research
04/16/2021

Word2rate: training and evaluating multiple word embeddings as statistical transitions

Using pretrained word embeddings has been shown to be a very effective w...
research
11/10/2017

Bayesian Paragraph Vectors

Word2vec (Mikolov et al., 2013) has proven to be successful in natural l...
research
06/10/2016

WordNet2Vec: Corpora Agnostic Word Vectorization Method

A complex nature of big data resources demands new methods for structuri...
research
02/15/2019

Contextual Word Representations: A Contextual Introduction

This introduction aims to tell the story of how we put words into comput...

Please sign up or login with your details

Forgot password? Click here to reset