Low-Dimensional Structure in the Space of Language Representations is Reflected in Brain Responses

06/09/2021
by   Richard Antonello, et al.
0

How related are the representations learned by neural language models, translation models, and language tagging tasks? We answer this question by adapting an encoder-decoder transfer learning method from computer vision to investigate the structure among 100 different feature spaces extracted from hidden representations of various networks trained on language tasks. This method reveals a low-dimensional structure where language models and translation models smoothly interpolate between word embeddings, syntactic and semantic tasks, and future word embeddings. We call this low-dimensional structure a language representation embedding because it encodes the relationships between representations needed to process language for a variety of NLP tasks. We find that this representation embedding can predict how well each individual feature space maps to human brain responses to natural language stimuli recorded using fMRI. Additionally, we find that the principal dimension of this structure can be used to create a metric which highlights the brain's natural language processing hierarchy. This suggests that the embedding captures some part of the brain's natural language representation structure.

READ FULL TEXT

page 5

page 7

page 8

page 16

page 17

page 18

page 20

page 21

research
08/31/2021

Sense representations for Portuguese: experiments with sense embeddings and deep neural language models

Sense representations have gone beyond word representations like Word2Ve...
research
03/23/2020

Data-driven models and computational tools for neurolinguistics: a language technology perspective

In this paper, our focus is the connection and influence of language tec...
research
10/13/2016

Mapping Between fMRI Responses to Movies and their Natural Language Annotations

Several research groups have shown how to correlate fMRI responses to th...
research
10/14/2020

From Language to Language-ish: How Brain-Like is an LSTM's Representation of Nonsensical Language Stimuli?

The representations generated by many models of language (word embedding...
research
06/20/2019

Low-dimensional Embodied Semantics for Music and Language

Embodied cognition states that semantics is encoded in the brain as firi...
research
12/04/2019

Natural Alpha Embeddings

Learning an embedding for a large collection of items is a popular appro...
research
04/18/2021

"Average" Approximates "First Principal Component"? An Empirical Analysis on Representations from Neural Language Models

Contextualized representations based on neural language models have furt...

Please sign up or login with your details

Forgot password? Click here to reset