Don't Search for a Search Method – Simple Heuristics Suffice for Adversarial Text Attacks

09/16/2021
by   Nathaniel Berger, et al.
0

Recently more attention has been given to adversarial attacks on neural networks for natural language processing (NLP). A central research topic has been the investigation of search algorithms and search constraints, accompanied by benchmark algorithms and tasks. We implement an algorithm inspired by zeroth order optimization-based attacks and compare with the benchmark results in the TextAttack framework. Surprisingly, we find that optimization-based methods do not yield any improvement in a constrained setup and slightly benefit from approximate gradient information only in unconstrained setups where search spaces are larger. In contrast, simple heuristics exploiting nearest neighbors without querying the target function yield substantial success rates in constrained setups, and nearly full success rate in unconstrained setups, at an order of magnitude fewer queries. We conclude from these results that current TextAttack benchmark tasks are too easy and constraints are too strict, preventing meaningful research on black-box adversarial text attacks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/10/2021

A Strong Baseline for Query Efficient Attacks in a Black Box Setting

Existing black box search methods have achieved high success rate in gen...
research
12/22/2021

An Attention Score Based Attacker for Black-box NLP Classifier

Deep neural networks have a wide range of applications in solving variou...
research
04/29/2020

TextAttack: A Framework for Adversarial Attacks in Natural Language Processing

TextAttack is a library for running adversarial attacks against natural ...
research
10/05/2019

Yet another but more efficient black-box adversarial attack: tiling and evolution strategies

We introduce a new black-box attack achieving state of the art performan...
research
03/16/2022

Attacking deep networks with surrogate-based adversarial black-box methods is easy

A recent line of work on black-box adversarial attacks has revived the u...
research
10/08/2020

Gaussian MRF Covariance Modeling for Efficient Black-Box Adversarial Attacks

We study the problem of generating adversarial examples in a black-box s...
research
01/22/2019

Universal Rules for Fooling Deep Neural Networks based Text Classification

Recently, deep learning based natural language processing techniques are...

Please sign up or login with your details

Forgot password? Click here to reset