On Training Robust PDF Malware Classifiers

04/06/2019
by   Yizheng Chen, et al.
0

Although state-of-the-art PDF malware classifiers can be trained with almost perfect test accuracy (99 it has been shown that even a simple adversary can evade them. A practically useful malware classifier must be robust against evasion attacks. However, achieving such robustness is an extremely challenging task. In this paper, we take the first steps towards training robust PDF malware classifiers with verifiable robustness properties. For instance, a robustness property can enforce that no matter how many pages from benign documents are inserted into a PDF malware, the classifier must still classify it as malicious. We demonstrate how the worst-case behavior of a malware classifier with respect to specific robustness properties can be formally verified. Furthermore, we find that training classifiers that satisfy formally verified robustness properties can increase the computation cost of unbounded (i.e., not bounded by the robustness properties) attackers by eliminating simple evasion attacks. Specifically, we propose a new distance metric that operates on the PDF tree structure and specify two classes of robustness properties including subtree insertions and deletions. We utilize state-of-the-art verifiably robust training method to build robust PDF malware classifiers. Our results show that, we can achieve 99 and 0.41 state-of-the-art unbounded attacker found no successful evasion on the robust classifier in 6 hours. Even for a new unbounded adaptive attacker we have designed, the number of successful evasions within a fixed time budget is cut down by 4x.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2021

Learning Security Classifiers with Verified Global Robustness Properties

Recent works have proposed methods to train classifiers with local robus...
research
10/30/2020

Being Single Has Benefits. Instance Poisoning to Deceive Malware Classifiers

The performance of a machine learning-based malware classifier depends o...
research
03/04/2022

Adversarial Patterns: Building Robust Android Malware Classifiers

Deep learning-based classifiers have substantially improved recognition ...
research
05/07/2020

Defending Hardware-based Malware Detectors against Adversarial Attacks

In the era of Internet of Things (IoT), Malware has been proliferating e...
research
10/22/2021

Improving Robustness of Malware Classifiers using Adversarial Strings Generated from Perturbed Latent Representations

In malware behavioral analysis, the list of accessed and created files v...
research
05/04/2020

Do Gradient-based Explanations Tell Anything About Adversarial Robustness to Android Malware?

Machine-learning algorithms trained on features extracted from static co...
research
06/12/2019

A Stratified Approach to Robustness for Randomly Smoothed Classifiers

Strong theoretical guarantees of robustness can be given for ensembles o...

Please sign up or login with your details

Forgot password? Click here to reset