Publications

Stats

View publication

Title Evaluating Defect Prediction Approaches: A Benchmark and an Extensive Comparison
Authors Marco D'Ambros, Michele Lanza, Romain Robbes
Publication date August 2012
Abstract Reliably predicting software defects is one of the holy
grails of
software engineering. Researchers have devised and implemented a plethora of
defect/bug prediction approaches varying in terms of accuracy, complexity
and the input data they require. However, the absence of an established
benchmark makes it hard, if not impossible, to compare approaches. We
present a benchmark for defect prediction, in the form of a publicly
available dataset consisting of several software systems, and provide an
extensive comparison of well-known bug prediction approaches, together with
novel approaches we devised. We evaluate the performance of the approaches
using different performance indicators: classification of entities as
defect-prone or not, ranking of the entities, with and without taking into
account the effort to review an entity. We performed three sets of
experiments aimed at (1) comparing the approaches across different systems,
(2) testing whether the differences in performance are statistically
significant, and (3) investigating the stability of approaches across
different learners. Our results indicate that, while some approaches perform
better than others in a statistically significant manner, external validity
in defect prediction is still an open problem, as generalizing results to
different contexts/learners proved to be a partially unsuccessful
endeavor.
Downloaded 6 times
Pages 531-577
Volume 17
Journal name Empirical Software Engineering
Publisher Springer-Verlag (Berlin/Heidelberg, Germany)
PDF View PDF
Reference URL View reference page