Understanding Learning Dynamics Of Language Models with SVCCA

11/01/2018
by   Naomi Saphra, et al.
0

Recent work has demonstrated that neural language models encode linguistic structure implicitly in a number of ways. However, existing research has not shed light on the process by which this structure is acquired during training. We use SVCCA as a tool for understanding how a language model is implicitly predicting a variety of word cluster tags. We present experiments suggesting that a single recurrent layer of a language model learns linguistic structure in phases. We find, for example, that a language model naturally stabilizes its representation of part of speech earlier than it learns semantic and topic information.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2021

Probing Linguistic Information For Logical Inference In Pre-trained Language Models

Progress in pre-trained language models has led to a surge of impressive...
research
09/16/2021

Do Language Models Know the Way to Rome?

The global geometry of language models is important for a range of appli...
research
06/01/2021

PIGLeT: Language Grounding Through Neuro-Symbolic Interaction in a 3D World

We propose PIGLeT: a model that learns physical commonsense knowledge th...
research
04/20/2018

What's Going On in Neural Constituency Parsers? An Analysis

A number of differences have emerged between modern and classic approach...
research
07/28/2023

The Hydra Effect: Emergent Self-repair in Language Model Computations

We investigate the internal structure of language model computations usi...
research
09/05/2020

Visually Analyzing Contextualized Embeddings

In this paper we introduce a method for visually analyzing contextualize...
research
04/20/2021

Frustratingly Easy Edit-based Linguistic Steganography with a Masked Language Model

With advances in neural language models, the focus of linguistic stegano...

Please sign up or login with your details

Forgot password? Click here to reset