Best-Effort Adversarial Approximation of Black-Box Malware Classifiers

06/28/2020
by   Abdullah Ali, et al.
0

An adversary who aims to steal a black-box model repeatedly queries the model via a prediction API to learn a function that approximates its decision boundary. Adversarial approximation is non-trivial because of the enormous combinations of model architectures, parameters, and features to explore. In this context, the adversary resorts to a best-effort strategy that yields the closest approximation. This paper explores best-effort adversarial approximation of a black-box malware classifier in the most challenging setting, where the adversary's knowledge is limited to a prediction label for a given input. Beginning with a limited input set for the black-box classifier, we leverage feature representation mapping and cross-domain transferability to approximate a black-box malware classifier by locally training a substitute. Our approach approximates the target model with different feature types for the target and the substitute model while also using non-overlapping data for training the target, training the substitute, and the comparison of the two. We evaluate the effectiveness of our approach against two black-box classifiers trained on Windows Portable Executables (PEs). Against a Convolutional Neural Network (CNN) trained on raw byte sequences of PEs, our approach achieves a 92 accurate substitute (trained on pixel representations of PEs), and nearly 90 prediction agreement between the target and the substitute model. Against a 97.8 our 91 suggesting the strength of our purely black-box approximation.

READ FULL TEXT

page 8

page 14

page 22

page 23

research
04/23/2018

Low Resource Black-Box End-to-End Attack Against State of the Art API Call Based Malware Classifiers

In this paper, we present a black-box attack against API call based mach...
research
03/30/2020

Efficient Black-box Optimization of Adversarial Windows Malware with Constrained Manipulations

Windows malware detectors based on machine learning are vulnerable to ad...
research
07/19/2017

Generic Black-Box End-to-End Attack Against State of the Art API Call Based Malware Classifiers

In this paper, we present a black-box attack against API call based mach...
research
04/05/2022

A Set Membership Approach to Discovering Feature Relevance and Explaining Neural Classifier Decisions

Neural classifiers are non linear systems providing decisions on the cla...
research
06/05/2021

Extracting Weighted Automata for Approximate Minimization in Language Modelling

In this paper we study the approximate minimization problem for language...
research
06/14/2018

Copycat CNN: Stealing Knowledge by Persuading Confession with Random Non-Labeled Data

In the past few years, Convolutional Neural Networks (CNNs) have been ac...
research
06/11/2018

Learning to Speed Up Structured Output Prediction

Predicting structured outputs can be computationally onerous due to the ...

Please sign up or login with your details

Forgot password? Click here to reset