BERT Rankers are Brittle: a Study using Adversarial Document Perturbations

06/23/2022
by   Yumeng Wang, et al.
0

Contextual ranking models based on BERT are now well established for a wide range of passage and document ranking tasks. However, the robustness of BERT-based ranking models under adversarial inputs is under-explored. In this paper, we argue that BERT-rankers are not immune to adversarial attacks targeting retrieved documents given a query. Firstly, we propose algorithms for adversarial perturbation of both highly relevant and non-relevant documents using gradient-based optimization methods. The aim of our algorithms is to add/replace a small number of tokens to a highly relevant or non-relevant document to cause a large rank demotion or promotion. Our experiments show that a small number of tokens can already result in a large change in the rank of a document. Moreover, we find that BERT-rankers heavily rely on the document start/head for relevance prediction, making the initial part of the document more susceptible to adversarial attacks. More interestingly, we find a small set of recurring adversarial words that when added to documents result in successful rank demotion/promotion of any relevant/non-relevant document respectively. Finally, our adversarial tokens also show particular topic preferences within and across datasets, exposing potential biases from BERT pre-training or downstream datasets.

READ FULL TEXT
research
04/17/2020

Learning-to-Rank with BERT in TF-Ranking

This paper describes a machine learning algorithm for document (re)ranki...
research
08/05/2020

One word at a time: adversarial attacks on retrieval models

Adversarial examples, generated by applying small perturbations to input...
research
04/16/2019

Understanding the Behaviors of BERT in Ranking

This paper studies the performances and behaviors of BERT in ranking tas...
research
05/23/2022

Learning to Ignore Adversarial Attacks

Despite the strong performance of current NLP models, they can be brittl...
research
07/30/2020

What does BERT know about books, movies and music? Probing BERT for Conversational Recommendation

Heavily pre-trained transformer models such as BERT have recently shown ...
research
05/03/2023

Towards Imperceptible Document Manipulations against Neural Ranking Models

Adversarial attacks have gained traction in order to identify potential ...
research
03/01/2021

BERT based patent novelty search by training claims to their own description

In this paper we present a method to concatenate patent claims to their ...

Please sign up or login with your details

Forgot password? Click here to reset