Adversarial Over-Sensitivity and Over-Stability Strategies for Dialogue Models

09/06/2018
by   Tong Niu, et al.
0

We present two categories of model-agnostic adversarial strategies that reveal the weaknesses of several generative, task-oriented dialogue models: Should-Not-Change strategies that evaluate over-sensitivity to small and semantics-preserving edits, as well as Should-Change strategies that test if a model is over-stable against subtle yet semantics-changing modifications. We next perform adversarial training with each strategy, employing a max-margin approach for negative generative examples. This not only makes the target dialogue model more robust to the adversarial inputs, but also helps it perform significantly better on the original inputs. Moreover, training on all strategies combined achieves further improvements, achieving a new state-of-the-art performance on the original task (also verified via human evaluation). In addition to adversarial training, we also address the robustness task at the model-level, by feeding it subword units as both inputs and outputs, and show that the resulting model is equally competitive, requires only 1/4 of the original vocabulary size, and is robust to one of the adversarial strategies (to which the original model is vulnerable) even without adversarial training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/23/2020

Semantics-Preserving Adversarial Training

Adversarial training is a defense technique that improves adversarial ro...
research
01/23/2017

Adversarial Learning for Neural Dialogue Generation

In this paper, drawing intuition from the Turing test, we propose using ...
research
08/20/2021

Towards Understanding the Generative Capability of Adversarially Robust Classifiers

Recently, some works found an interesting phenomenon that adversarially ...
research
04/17/2018

Robust Machine Comprehension Models via Adversarial Training

It is shown that many published models for the Stanford Question Answeri...
research
10/03/2020

Efficient Robust Training via Backward Smoothing

Adversarial training is so far the most effective strategy in defending ...
research
11/12/2018

Improved Dynamic Memory Network for Dialogue Act Classification with Adversarial Training

Dialogue Act (DA) classification is a challenging problem in dialogue in...
research
08/23/2022

Predicting Query-Item Relationship using Adversarial Training and Robust Modeling Techniques

We present an effective way to predict search query-item relationship. W...

Please sign up or login with your details

Forgot password? Click here to reset