Preserving Order of Data When Validating Defect Prediction Models

09/05/2018
by   Davide Falessi, et al.
0

[Context] The use of defect prediction models, such as classifiers, can support testing resource allocations by using data of the previous releases of the same project for predicting which software components are likely to be defective. A validation technique, hereinafter technique defines a specific way to split available data in training and test sets to measure a classifier accuracy. Time-series techniques have the unique ability to preserve the temporal order of data; i.e., preventing the testing set to have data antecedent to the training set. [Aim] The aim of this paper is twofold: first we check if there is a difference in the classifiers accuracy measured by time-series versus non-time-series techniques. Afterward, we check for a possible reason for this difference, i.e., if defect rates change across releases of a project. [Method] Our method consists of measuring the accuracy, i.e., AUC, of 10 classifiers on 13 open and two closed projects by using three validation techniques, namely cross validation, bootstrap, and walk-forward, where only the latter is a time-series technique. [Results] We find that the AUC of the same classifier used on the same project and measured by 10-fold varies compared to when measured by walk-forward in the range [-0.20, 0.22], and it is statistically different in 45 measured by bootstrap varies compared to when measured by walk-forward in the range [-0.17, 0.43], and it is statistically different in 56 [Conclusions] We recommend choosing the technique to be used by carefully considering the conclusions to draw, the property of the available datasets, and the level of realism with the classifier usage scenario.

READ FULL TEXT
research
05/28/2019

Evaluating time series forecasting models: An empirical study on performance estimation methods

Performance estimation aims at estimating the loss that a predictive mod...
research
01/31/2018

The Impact of Class Rebalancing Techniques on the Performance and Interpretation of Defect Prediction Models

Defect prediction models that are trained on class imbalanced datasets (...
research
02/27/2020

Complexity Measures and Features for Times Series classification

Classification of time series is a growing problem in different discipli...
research
10/21/2019

Generalised learning of time-series: Ornstein-Uhlenbeck processes

In machine learning, statistics, econometrics and statistical physics, k...
research
04/12/2022

Hold-out estimates of prediction models for Markov processes

We consider the selection of prediction models for Markovian time series...
research
01/24/2019

Transfer-Learning Oriented Class Imbalance Learning for Cross-Project Defect Prediction

Cross-project defect prediction (CPDP) aims to predict defects of projec...
research
09/11/2019

Iterative versus Exhaustive Data Selection for Cross Project Defect Prediction: An Extended Replication Study

Context: The effectiveness of data selection approaches in improving the...

Please sign up or login with your details

Forgot password? Click here to reset