minicons: Enabling Flexible Behavioral and Representational Analyses of Transformer Language Models

03/24/2022
by   Kanishka Misra, et al.
4

We present minicons, an open source library that provides a standard API for researchers interested in conducting behavioral and representational analyses of transformer-based language models (LMs). Specifically, minicons enables researchers to apply analysis methods at two levels: (1) at the prediction level – by providing functions to efficiently extract word/sentence level probabilities; and (2) at the representational level – by also facilitating efficient extraction of word/phrase level vectors from one or more layers. In this paper, we describe the library and apply it to two motivating case studies: One focusing on the learning dynamics of the BERT architecture on relative grammatical judgments, and the other on benchmarking 23 different LMs on zero-shot abductive reasoning. minicons is available at https://github.com/kanishkamisra/minicons

READ FULL TEXT
research
11/13/2020

diagNNose: A Library for Neural Activation Analysis

In this paper we introduce diagNNose, an open source library for analysi...
research
07/31/2023

Camoscio: an Italian Instruction-tuned LLaMA

In recent years Large Language Models (LLMs) have increased the state of...
research
09/15/2022

LAVIS: A Library for Language-Vision Intelligence

We introduce LAVIS, an open-source deep learning library for LAnguage-VI...
research
04/27/2020

DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference

Large-scale pre-trained language models such as BERT have brought signif...
research
07/28/2022

LAD: Language Models as Data for Zero-Shot Dialog

To facilitate zero-shot generalization in taskoriented dialog, this pape...
research
07/17/2023

A mixed policy to improve performance of language models on math problems

When to solve math problems, most language models take a sampling strate...
research
02/03/2020

Torch-Struct: Deep Structured Prediction Library

The literature on structured prediction for NLP describes a rich collect...

Please sign up or login with your details

Forgot password? Click here to reset