Nothing Special   »   [go: up one dir, main page]

skip to main content
article

DConfusion: a technique to allow cross study performance evaluation of fault prediction studies

Published: 01 April 2014 Publication History

Abstract

There are many hundreds of fault prediction models published in the literature. The predictive performance of these models is often reported using a variety of different measures. Most performance measures are not directly comparable. This lack of comparability means that it is often difficult to evaluate the performance of one model against another. Our aim is to present an approach that allows other researchers and practitioners to transform many performance measures back into a confusion matrix. Once performance is expressed in a confusion matrix alternative preferred performance measures can then be derived. Our approach has enabled us to compare the performance of 600 models published in 42 studies. We demonstrate the application of our approach on 8 case studies, and discuss the advantages and implications of doing this.

Cited By

View all
  • (2023)DexBERT: Effective, Task-Agnostic and Fine-Grained Representation Learning of Android BytecodeIEEE Transactions on Software Engineering10.1109/TSE.2023.331087449:10(4691-4706)Online publication date: 1-Oct-2023
  • (2022)A Survey of Different Approaches for the Class Imbalance Problem in Software Defect PredictionInternational Journal of Software Science and Computational Intelligence10.4018/IJSSCI.30126814:1(1-26)Online publication date: 3-Jun-2022
  • (2020)Evaluation of Sampling-Based Ensembles of Classifiers on Imbalanced Data for Software Defect Prediction ProblemsSN Computer Science10.1007/s42979-020-0119-41:2Online publication date: 30-Mar-2020
  • Show More Cited By
  1. DConfusion: a technique to allow cross study performance evaluation of fault prediction studies

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Automated Software Engineering
    Automated Software Engineering  Volume 21, Issue 2
    April 2014
    168 pages

    Publisher

    Kluwer Academic Publishers

    United States

    Publication History

    Published: 01 April 2014

    Author Tags

    1. Confusion matrix
    2. Fault
    3. Machine learning

    Qualifiers

    • Article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 04 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2023)DexBERT: Effective, Task-Agnostic and Fine-Grained Representation Learning of Android BytecodeIEEE Transactions on Software Engineering10.1109/TSE.2023.331087449:10(4691-4706)Online publication date: 1-Oct-2023
    • (2022)A Survey of Different Approaches for the Class Imbalance Problem in Software Defect PredictionInternational Journal of Software Science and Computational Intelligence10.4018/IJSSCI.30126814:1(1-26)Online publication date: 3-Jun-2022
    • (2020)Evaluation of Sampling-Based Ensembles of Classifiers on Imbalanced Data for Software Defect Prediction ProblemsSN Computer Science10.1007/s42979-020-0119-41:2Online publication date: 30-Mar-2020
    • (2019)The Prevalence of Errors in Machine Learning ExperimentsIntelligent Data Engineering and Automated Learning – IDEAL 201910.1007/978-3-030-33607-3_12(102-109)Online publication date: 14-Nov-2019
    • (2018)Reproducibility and replicability of software defect prediction studiesInformation and Software Technology10.1016/j.infsof.2018.02.00399:C(148-163)Online publication date: 1-Jul-2018
    • (2016)So You Need More Method Level Datasets for Your Software Defect Prediction?Proceedings of the 10th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement10.1145/2961111.2962620(1-6)Online publication date: 8-Sep-2016
    • (2015)Different Classifiers Find Different Defects Although With Different Level of ConsistencyProceedings of the 11th International Conference on Predictive Models and Data Analytics in Software Engineering10.1145/2810146.2810149(1-10)Online publication date: 21-Oct-2015
    • (2014)A proposed method to evaluate and compare fault predictions across studiesProceedings of the 10th International Conference on Predictive Models in Software Engineering10.1145/2639490.2639504(2-11)Online publication date: 17-Sep-2014

    View Options

    View options

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media