Evaluating Neural Machine Comprehension Model Robustness to Noisy Inputs and Adversarial Attacks

05/01/2020
by   Winston Wu, et al.
0

We evaluate machine comprehension models' robustness to noise and adversarial attacks by performing novel perturbations at the character, word, and sentence level. We experiment with different amounts of perturbations to examine model confidence and misclassification rate, and contrast model performance in adversarial training with different embedding types on two benchmark datasets. We demonstrate improving model performance with ensembling. Finally, we analyze factors that effect model behavior under adversarial training and develop a model to predict model errors during adversarial attacks.

READ FULL TEXT
research
06/08/2020

Adversarial Feature Desensitization

Deep neural networks can now perform many tasks that were once thought t...
research
09/06/2019

Learning to Discriminate Perturbations for Blocking Adversarial Attacks in Text Classification

Adversarial attacks against machine learning models have threatened vari...
research
04/23/2021

Evaluating Deception Detection Model Robustness To Linguistic Variation

With the increasing use of machine-learning driven algorithmic judgement...
research
04/29/2022

Logically Consistent Adversarial Attacks for Soft Theorem Provers

Recent efforts within the AI community have yielded impressive results t...
research
03/23/2023

Decentralized Adversarial Training over Graphs

The vulnerability of machine learning models to adversarial attacks has ...
research
05/18/2023

Quantifying the robustness of deep multispectral segmentation models against natural perturbations and data poisoning

In overhead image segmentation tasks, including additional spectral band...
research
06/17/2021

Evaluating the Robustness of Bayesian Neural Networks Against Different Types of Attacks

To evaluate the robustness gain of Bayesian neural networks on image cla...

Please sign up or login with your details

Forgot password? Click here to reset