@article {Wahono:2014:1936-6612:1945, title = "A Comparison Framework of Classification Models for Software Defect Prediction", journal = "Advanced Science Letters", parent_itemid = "infobike://asp/asl", publishercode ="asp", year = "2014", volume = "20", number = "10-11", publication date ="2014-10-01T00:00:00", pages = "1945-1950", itemtype = "ARTICLE", issn = "1936-6612", url = "https://www.ingentaconnect.com/content/asp/asl/2014/00000020/f0030010/art00045", doi = "doi:10.1166/asl.2014.5640", author = "Wahono, Romi Satria and Herman, Nanna Suryana and Ahmad, Sabrina", abstract = "Software defects are expensive in quality and cost. The accurate prediction of defect-prone software modules can help direct test effort, reduce costs, and improve the quality of software. Machine learning classification algorithms is a popular approach for predicting software defect. Various types of classification algorithms have been applied for software defect prediction. However, no clear consensus on which algorithm perform best when individual studies are looked at separately. In this research, a comparison framework is proposed, which aims to benchmark the performance of a wide range of classification models within the field of software defect prediction. For the purpose of this study, 10 classifiers are selected and applied to build classification models and test their performance in 9 NASA MDP datasets. Area under curve (AUC) is employed as an accuracy indicator in our framework to evaluate the performance of classifiers. Friedman and Nemenyi post hoc tests are used to test for significance of AUC differences between classifiers. The results show that the logistic regression perform best in most NASA MDP datasets. Naive bayes, neural network, support vector machine and k* classifiers also perform well. Decision tree based classifiers tend to underperform, as well as linear discriminant analysis and k-nearest neighbor.", }