Test Selection for Deep Learning Systems

04/30/2019
by   Wei Ma, et al.
0

Testing of deep learning models is challenging due to the excessive number and complexity of computations involved. As a result, test data selection is performed manually and in an ad hoc way. This raises the question of how we can automatically select candidate test data to test deep learning models. Recent research has focused on adapting test selection metrics from code-based software testing (such as coverage) to deep learning. However, deep learning models have different attributes from code such as spread of computations across the entire network reflecting training data properties, balance of neuron weights and redundancy (use of many more neurons than needed). Such differences make code-based metrics inappropriate to select data that can challenge the models (can trigger misclassification). We thus propose a set of test selection metrics based on the notion of model uncertainty (model confidence on specific inputs). Intuitively, the more uncertain we are about a candidate sample, the more likely it is that this sample triggers a misclassification. Similarly, the samples for which we are the most uncertain, are the most informative and should be used to improve the model by retraining. We evaluate these metrics on two widely-used image classification problems involving real and artificial (adversarial) data. We show that uncertainty-based metrics have a strong ability to select data that are misclassified and lead to major improvement in classification accuracy during retraining: up to 80 state-of-the-art metrics on one dataset and up to 29

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/08/2022

Guiding the retraining of convolutional neural networks against adversarial inputs

Background: When using deep learning models, there are many possible vul...
research
12/29/2020

Uncertainty-Wizard: Fast and User-Friendly Neural Network Uncertainty Quantification

Uncertainty and confidence have been shown to be useful metrics in a wid...
research
07/29/2023

Evaluating the Robustness of Test Selection Methods for Deep Neural Networks

Testing deep learning-based systems is crucial but challenging due to th...
research
06/19/2023

AdaSelection: Accelerating Deep Learning Training through Data Subsampling

In this paper, we introduce AdaSelection, an adaptive sub-sampling metho...
research
06/17/2021

CoCoFuzzing: Testing Neural Code Models with Coverage-Guided Fuzzing

Deep learning-based code processing models have shown good performance f...
research
01/28/2022

Systematic Training and Testing for Machine Learning Using Combinatorial Interaction Testing

This paper demonstrates the systematic use of combinatorial coverage for...
research
04/27/2021

One Backward from Ten Forward, Subsampling for Large-Scale Deep Learning

Deep learning models in large-scale machine learning systems are often c...

Please sign up or login with your details

Forgot password? Click here to reset