Impact of Feature Encoding on Malware Classification Explainability

07/10/2023
by   Elyes Manai, et al.
0

This paper investigates the impact of feature encoding techniques on the explainability of XAI (Explainable Artificial Intelligence) algorithms. Using a malware classification dataset, we trained an XGBoost model and compared the performance of two feature encoding methods: Label Encoding (LE) and One Hot Encoding (OHE). Our findings reveal a marginal performance loss when using OHE instead of LE. However, the more detailed explanations provided by OHE compensated for this loss. We observed that OHE enables deeper exploration of details in both global and local contexts, facilitating more comprehensive answers. Additionally, we observed that using OHE resulted in smaller explanation files and reduced analysis time for human analysts. These findings emphasize the significance of considering feature encoding techniques in XAI research and suggest potential for further exploration by incorporating additional encoding methods and innovative visualization approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/07/2021

Counterfactuals and Causability in Explainable Artificial Intelligence: Theory, Algorithms, and Applications

There has been a growing interest in model-agnostic methods that can mak...
research
07/18/2023

What's meant by explainable model: A Scoping Review

We often see the term explainable in the titles of papers that describe ...
research
07/26/2023

A New Perspective on Evaluation Methods for Explainable Artificial Intelligence (XAI)

Within the field of Requirements Engineering (RE), the increasing signif...
research
03/30/2022

Does Configuration Encoding Matter in Learning Software Performance? An Empirical Study on Encoding Schemes

Learning and predicting the performance of a configurable software syste...

Please sign up or login with your details

Forgot password? Click here to reset