Article ID Journal Published Year Pages File Type
386000 Expert Systems with Applications 2011 6 Pages PDF
Abstract

Predicting the fault-proneness labels of software program modules is an emerging software quality assurance activity and the quality of datasets collected from previous software version affects the performance of fault prediction models. In this paper, we propose an outlier detection approach using metrics thresholds and class labels to identify class outliers. We evaluate our approach on public NASA datasets from PROMISE repository. Experiments reveal that this novel outlier detection method improves the performance of robust software fault prediction models based on Naive Bayes and Random Forests machine learning algorithms.

Research highlights► The quality of datasets affect the performance of fault prediction models. ► When outliers are removed, the performance of fault prediction models increase. ► Software metrics thresholds based outlier detection approach is an effective technique to identify outliers.

Related Topics
Physical Sciences and Engineering Computer Science Artificial Intelligence
Authors
, ,