BERT Family Eat Word Salad: Experiments with Text Understanding

01/10/2021
by   Ashim Gupta, et al.
6

In this paper, we study the response of large models from the BERT family to incoherent inputs that should confuse any model that claims to understand natural language. We define simple heuristics to construct such examples. Our experiments show that state-of-the-art models consistently fail to recognize them as ill-formed, and instead produce high confidence predictions on them. Finally, we show that if models are explicitly trained to recognize invalid inputs, they can be robust to such attacks without a drop in performance.

READ FULL TEXT

page 12

page 14

research
05/04/2020

Robust Encodings: A Framework for Combating Adversarial Typos

Despite excellent performance on many tasks, NLP systems are easily fool...
research
02/14/2020

Understanding patient complaint characteristics using contextual clinical BERT embeddings

In clinical conversational applications, extracted entities tend to capt...
research
01/20/2023

Phoneme-Level BERT for Enhanced Prosody of Text-to-Speech with Grapheme Predictions

Large-scale pre-trained language models have been shown to be helpful in...
research
07/27/2019

Is BERT Really Robust? Natural Language Attack on Text Classification and Entailment

Machine learning algorithms are often vulnerable to adversarial examples...
research
09/15/2021

BERT is Robust! A Case Against Synonym-Based Adversarial Examples in Text Classification

Deep Neural Networks have taken Natural Language Processing by storm. Wh...
research
02/04/2019

Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference

Machine learning systems can often achieve high performance on a test se...
research
09/18/2020

Will it Unblend?

Natural language processing systems often struggle with out-of-vocabular...

Please sign up or login with your details

Forgot password? Click here to reset