PhilaeX: Explaining the Failure and Success of AI Models in Malware Detection

07/02/2022
by   Zhi Lu, et al.
0

The explanation to an AI model's prediction used to support decision making in cyber security, is of critical importance. It is especially so when the model's incorrect prediction can lead to severe damages or even losses to lives and critical assets. However, most existing AI models lack the ability to provide explanations on their prediction results, despite their strong performance in most scenarios. In this work, we propose a novel explainable AI method, called PhilaeX, that provides the heuristic means to identify the optimized subset of features to form the complete explanations of AI models' predictions. It identifies the features that lead to the model's borderline prediction, and those with positive individual contributions are extracted. The feature attributions are then quantified through the optimization of a Ridge regression model. We verify the explanation fidelity through two experiments. First, we assess our method's capability in correctly identifying the activated features in the adversarial samples of Android malwares, through the features attribution values from PhilaeX. Second, the deduction and augmentation tests, are used to assess the fidelity of the explanations. The results show that PhilaeX is able to explain different types of classifiers correctly, with higher fidelity explanations, compared to the state-of-the-arts methods such as LIME and SHAP.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2021

"How Does It Detect A Malicious App?" Explaining the Predictions of AI-based Android Malware Detector

AI methods have been proven to yield impressive performance on Android m...
research
01/18/2023

Understanding the Role of Human Intuition on Reliance in Human-AI Decision-Making with Explanations

AI explanations are often mentioned as a way to improve human-AI decisio...
research
08/10/2023

FINER: Enhancing State-of-the-art Classifiers with Feature Attribution to Facilitate Security Analysis

Deep learning classifiers achieve state-of-the-art performance in variou...
research
03/12/2021

Explainable AI by BAPC – Before and After correction Parameter Comparison

By means of a local surrogate approach, an analytical method to yield ex...
research
07/05/2022

"Even if ..." – Diverse Semifactual Explanations of Reject

Machine learning based decision making systems applied in safety critica...
research
08/30/2023

Calibrated Explanations for Regression

Artificial Intelligence (AI) is often an integral part of modern decisio...
research
08/08/2019

Measurable Counterfactual Local Explanations for Any Classifier

We propose a novel method for explaining the predictions of any classifi...

Please sign up or login with your details

Forgot password? Click here to reset