Scaling laws for language encoding models in fMRI

05/19/2023
by   Richard Antonello, et al.
0

Representations from transformer-based unidirectional language models are known to be effective at predicting brain responses to natural language. However, most studies comparing language models to brains have used GPT-2 or similarly sized language models. Here we tested whether larger open-source models such as those from the OPT and LLaMA families are better at predicting brain responses recorded using fMRI. Mirroring scaling results from other contexts, we found that brain prediction performance scales log-linearly with model size from 125M to 30B parameter models, with  15 performance as measured by correlation with a held-out test set across 3 subjects. Similar log-linear behavior was observed when scaling the size of the fMRI training set. We also characterized scaling for acoustic encoding models that use HuBERT, WavLM, and Whisper, and we found comparable improvements with model size. A noise ceiling analysis of these large, high-performance encoding models showed that performance is nearing the theoretical maximum for brain areas such as the precuneus and higher auditory cortex. These results suggest that increasing scale in both models and data will yield incredibly effective models of language processing in the brain, enabling better scientific understanding as well as applications such as decoding.

READ FULL TEXT

page 8

page 9

page 15

page 18

page 19

page 21

page 23

page 24

research
08/01/2023

Applicability of scaling laws to vision encoding models

In this paper, we investigated how to build a high-performance vision en...
research
10/02/2019

Linking artificial and human neural representations of language

What information from an act of sentence understanding is robustly repre...
research
04/04/2019

Robust Evaluation of Language-Brain Encoding Experiments

Language-brain encoding experiments evaluate the ability of language mod...
research
05/03/2022

Neural Language Taskonomy: Which NLP Tasks are the most Predictive of fMRI Brain Activity?

Several popular Transformer based language models have been found to be ...
research
10/13/2016

Mapping Between fMRI Responses to Movies and their Natural Language Annotations

Several research groups have shown how to correlate fMRI responses to th...
research
05/20/2023

Brain encoding models based on multimodal transformers can transfer across language and vision

Encoding models have been used to assess how the human brain represents ...
research
05/12/2022

Predicting Human Psychometric Properties Using Computational Language Models

Transformer-based language models (LMs) continue to achieve state-of-the...

Please sign up or login with your details

Forgot password? Click here to reset