How Reliable are Model Diagnostics?

05/12/2021
by   Vamsi Aribandi, et al.
6

In the pursuit of a deeper understanding of a model's behaviour, there is recent impetus for developing suites of probes aimed at diagnosing models beyond simple metrics like accuracy or BLEU. This paper takes a step back and asks an important and timely question: how reliable are these diagnostics in providing insight into models and training setups? We critically examine three recent diagnostic tests for pre-trained language models, and find that likelihood-based and representation-based model diagnostics are not yet as reliable as previously assumed. Based on our empirical findings, we also formulate recommendations for practitioners and researchers.

READ FULL TEXT

page 3

page 4

research
01/28/2023

On Pre-trained Language Models for Antibody

Antibodies are vital proteins offering robust protection for the human b...
research
09/11/2021

Multilingual Translation via Grafting Pre-trained Language Models

Can pre-trained BERT for one language and GPT for another be glued toget...
research
09/11/2022

Testing Pre-trained Language Models' Understanding of Distributivity via Causal Mediation Analysis

To what extent do pre-trained language models grasp semantic knowledge r...
research
01/30/2020

Are Pre-trained Language Models Aware of Phrases? Simple but Strong Baselines for Grammar Induction

With the recent success and popularity of pre-trained language models (L...
research
09/27/2021

Pragmatic competence of pre-trained language models through the lens of discourse connectives

As pre-trained language models (LMs) continue to dominate NLP, it is inc...
research
06/17/2021

Knowledgeable or Educated Guess? Revisiting Language Models as Knowledge Bases

Previous literatures show that pre-trained masked language models (MLMs)...

Please sign up or login with your details

Forgot password? Click here to reset