Defending Against Poisoning Attacks in Open-Domain Question Answering

12/20/2022
by   Orion Weller, et al.
3

Recent work in open-domain question answering (ODQA) has shown that adversarial poisoning of the input contexts can cause large drops in accuracy for production systems. However, little to no work has proposed methods to defend against these attacks. To do so, we introduce a new method that uses query augmentation to search for a diverse set of retrieved passages that could answer the original question. We integrate these new passages into the model through the design of a novel confidence method, comparing the predicted answer to its appearance in the retrieved contexts (what we call Confidence from Answer Redundancy, e.g. CAR). Together these methods allow for a simple but effective way to defend against poisoning attacks and provide gains of 5-20 exact match across varying levels of data poisoning.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset