Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/823456.823988guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Using Simulation to Evaluate Prediction Techniques

Published: 04 April 2001 Publication History

Abstract

The need for accurate software prediction systems increases as software becomes much larger and more complex. A variety of techniques have been proposed, however, none has proved consistently accurate and there is still much uncertainty as to what technique suits which type of prediction problem. We believe that the underlying characteristics - size, number of features, type of distribution, etc. - of the dataset influence the choice of the prediction system to be used. In previous work, it has proved difficult to obtain significant results over small datasets. Consequently we required large validation datasets, moreover, we wished to control the characteristics of such datasets in order to systematically explore the relationship between accuracy, choice of prediction system and dataset characteristic. Our solution has been to simulate data allowing both control and the possibility of large (1000) validation cases. In this paper we compared regression, rule induction and nearest neighbour (a form of case based reasoning). The results suggest that there are significant differences depending upon the characteristics of the dataset. Consequently researchers should consider prediction context when evaluating competing prediction systems. We also observed that the more "messy" the data and the more complex the relationship with the dependent variable the more variability in the results. This became apparent since we sampled two different training sets from each simulated population of data. In the more complex cases we observed significantly different results depending upon the training set. This suggests that researchers will need to exercise caution when comparing different approaches and utilise procedures such as bootstrapping in order to generate multiple samples for training purposes.

Cited By

View all
  • (2018)Linear Programming as a Baseline for Software Effort EstimationACM Transactions on Software Engineering and Methodology10.1145/323494027:3(1-28)Online publication date: 17-Sep-2018
  • (2016)Multi-objective software effort estimationProceedings of the 38th International Conference on Software Engineering10.1145/2884781.2884830(619-630)Online publication date: 14-May-2016
  • (2013)Tuning of cost drivers by significance occurrences and their calibration with novel software effort estimation methodAdvances in Software Engineering10.1155/2013/3519132013(6-6)Online publication date: 1-Jan-2013
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
METRICS '01: Proceedings of the 7th International Symposium on Software Metrics
April 2001
ISBN:0769510434

Publisher

IEEE Computer Society

United States

Publication History

Published: 04 April 2001

Author Tags

  1. dataset characteristic
  2. prediction system
  3. simulation

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 21 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2018)Linear Programming as a Baseline for Software Effort EstimationACM Transactions on Software Engineering and Methodology10.1145/323494027:3(1-28)Online publication date: 17-Sep-2018
  • (2016)Multi-objective software effort estimationProceedings of the 38th International Conference on Software Engineering10.1145/2884781.2884830(619-630)Online publication date: 14-May-2016
  • (2013)Tuning of cost drivers by significance occurrences and their calibration with novel software effort estimation methodAdvances in Software Engineering10.1155/2013/3519132013(6-6)Online publication date: 1-Jan-2013
  • (2013)Software development cost estimation using similarity difference between software attributesProceedings of the 2013 International Conference on Information Systems and Design of Communication10.1145/2503859.2503860(1-6)Online publication date: 11-Jul-2013
  • (2013)Using CBR and CART to predict maintainability of relational database-driven software applicationsProceedings of the 17th International Conference on Evaluation and Assessment in Software Engineering10.1145/2460999.2461019(132-143)Online publication date: 14-Apr-2013
  • (2012)Web effort estimationProceedings of the 8th International Conference on Predictive Models in Software Engineering10.1145/2365324.2365330(29-38)Online publication date: 21-Sep-2012
  • (2008)Using process simulation to assess the test design effort reduction of a model-based testing approachProceedings of the Software process, 2008 international conference on Making globally distributed software development a success story10.5555/1789757.1789788(282-293)Online publication date: 10-May-2008
  • (2008)A constrained regression technique for cocomo calibrationProceedings of the Second ACM-IEEE international symposium on Empirical software engineering and measurement10.1145/1414004.1414040(213-222)Online publication date: 9-Oct-2008
  • (2008)Comparative studies of the model evaluation criterions mmre and pred in software cost estimation researchProceedings of the Second ACM-IEEE international symposium on Empirical software engineering and measurement10.1145/1414004.1414015(51-60)Online publication date: 9-Oct-2008
  • (2008)Confidence in software cost estimation results based on MMRE and PREDProceedings of the 4th international workshop on Predictor models in software engineering10.1145/1370788.1370804(63-70)Online publication date: 12-May-2008
  • Show More Cited By

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media