Targeted Adversarial Training for Natural Language Understanding

04/12/2021
by   Lis Pereira, et al.
0

We present a simple yet effective Targeted Adversarial Training (TAT) algorithm to improve adversarial training for natural language understanding. The key idea is to introspect current mistakes and prioritize adversarial training steps to where the model errs the most. Experiments show that TAT can significantly improve accuracy over standard adversarial training on GLUE and attain new state-of-the-art zero-shot results on XNLI. Our code will be released at: https://github.com/namisan/mt-dnn.

READ FULL TEXT

page 1

page 4

research
04/11/2021

Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach

Adversarial training has been shown to improve the generalization perfor...
research
09/25/2019

FreeLB: Enhanced Adversarial Training for Language Understanding

Adversarial training, which minimizes the maximal risk for label-preserv...
research
09/29/2020

A Simple but Tough-to-Beat Data Augmentation Approach for Natural Language Understanding and Generation

Adversarial training has been shown effective at endowing the learned re...
research
03/12/2022

Towards Equal Opportunity Fairness through Adversarial Learning

Adversarial training is a common approach for bias mitigation in natural...
research
04/30/2020

TextAT: Adversarial Training for Natural Language Understanding with Token-Level Perturbation

Adversarial training is effective in improving the robustness of neural ...
research
06/17/2019

MixUp as Directional Adversarial Training

In this work, we explain the working mechanism of MixUp in terms of adve...
research
07/14/2023

Adversarial Training Over Long-Tailed Distribution

In this paper, we study adversarial training on datasets that obey the l...

Please sign up or login with your details

Forgot password? Click here to reset