Techniques for evaluating fault prediction models
Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicabil...
Ausführliche Beschreibung
Autor*in: |
Jiang, Yue [verfasserIn] |
---|
Format: |
Artikel |
---|---|
Sprache: |
Englisch |
Erschienen: |
2008 |
---|
Schlagwörter: |
---|
Anmerkung: |
© Springer Science+Business Media, LLC 2008 |
---|
Übergeordnetes Werk: |
Enthalten in: Empirical software engineering - Springer US, 1996, 13(2008), 5 vom: 12. Aug., Seite 561-595 |
---|---|
Übergeordnetes Werk: |
volume:13 ; year:2008 ; number:5 ; day:12 ; month:08 ; pages:561-595 |
Links: |
---|
DOI / URN: |
10.1007/s10664-008-9079-3 |
---|
Katalog-ID: |
OLC2071660064 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | OLC2071660064 | ||
003 | DE-627 | ||
005 | 20230503052507.0 | ||
007 | tu | ||
008 | 200819s2008 xx ||||| 00| ||eng c | ||
024 | 7 | |a 10.1007/s10664-008-9079-3 |2 doi | |
035 | |a (DE-627)OLC2071660064 | ||
035 | |a (DE-He213)s10664-008-9079-3-p | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
082 | 0 | 4 | |a 004 |q VZ |
100 | 1 | |a Jiang, Yue |e verfasserin |4 aut | |
245 | 1 | 0 | |a Techniques for evaluating fault prediction models |
264 | 1 | |c 2008 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ohne Hilfsmittel zu benutzen |b n |2 rdamedia | ||
338 | |a Band |b nc |2 rdacarrier | ||
500 | |a © Springer Science+Business Media, LLC 2008 | ||
520 | |a Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. | ||
650 | 4 | |a Fault-prediction models | |
650 | 4 | |a Model evaluation | |
650 | 4 | |a Predictive models in software engineering | |
650 | 4 | |a Empirical studies | |
700 | 1 | |a Cukic, Bojan |4 aut | |
700 | 1 | |a Ma, Yan |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Empirical software engineering |d Springer US, 1996 |g 13(2008), 5 vom: 12. Aug., Seite 561-595 |w (DE-627)235946516 |w (DE-600)1401304-6 |w (DE-576)102432406 |x 1382-3256 |7 nnns |
773 | 1 | 8 | |g volume:13 |g year:2008 |g number:5 |g day:12 |g month:08 |g pages:561-595 |
856 | 4 | 1 | |u https://doi.org/10.1007/s10664-008-9079-3 |z lizenzpflichtig |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a SYSFLAG_A | ||
912 | |a GBV_OLC | ||
912 | |a SSG-OLC-MAT | ||
912 | |a GBV_ILN_70 | ||
912 | |a GBV_ILN_100 | ||
951 | |a AR | ||
952 | |d 13 |j 2008 |e 5 |b 12 |c 08 |h 561-595 |
author_variant |
y j yj b c bc y m ym |
---|---|
matchkey_str |
article:13823256:2008----::ehiusoeautnfutrd |
hierarchy_sort_str |
2008 |
publishDate |
2008 |
allfields |
10.1007/s10664-008-9079-3 doi (DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p DE-627 ger DE-627 rakwb eng 004 VZ Jiang, Yue verfasserin aut Techniques for evaluating fault prediction models 2008 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © Springer Science+Business Media, LLC 2008 Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies Cukic, Bojan aut Ma, Yan aut Enthalten in Empirical software engineering Springer US, 1996 13(2008), 5 vom: 12. Aug., Seite 561-595 (DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 1382-3256 nnns volume:13 year:2008 number:5 day:12 month:08 pages:561-595 https://doi.org/10.1007/s10664-008-9079-3 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 AR 13 2008 5 12 08 561-595 |
spelling |
10.1007/s10664-008-9079-3 doi (DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p DE-627 ger DE-627 rakwb eng 004 VZ Jiang, Yue verfasserin aut Techniques for evaluating fault prediction models 2008 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © Springer Science+Business Media, LLC 2008 Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies Cukic, Bojan aut Ma, Yan aut Enthalten in Empirical software engineering Springer US, 1996 13(2008), 5 vom: 12. Aug., Seite 561-595 (DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 1382-3256 nnns volume:13 year:2008 number:5 day:12 month:08 pages:561-595 https://doi.org/10.1007/s10664-008-9079-3 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 AR 13 2008 5 12 08 561-595 |
allfields_unstemmed |
10.1007/s10664-008-9079-3 doi (DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p DE-627 ger DE-627 rakwb eng 004 VZ Jiang, Yue verfasserin aut Techniques for evaluating fault prediction models 2008 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © Springer Science+Business Media, LLC 2008 Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies Cukic, Bojan aut Ma, Yan aut Enthalten in Empirical software engineering Springer US, 1996 13(2008), 5 vom: 12. Aug., Seite 561-595 (DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 1382-3256 nnns volume:13 year:2008 number:5 day:12 month:08 pages:561-595 https://doi.org/10.1007/s10664-008-9079-3 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 AR 13 2008 5 12 08 561-595 |
allfieldsGer |
10.1007/s10664-008-9079-3 doi (DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p DE-627 ger DE-627 rakwb eng 004 VZ Jiang, Yue verfasserin aut Techniques for evaluating fault prediction models 2008 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © Springer Science+Business Media, LLC 2008 Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies Cukic, Bojan aut Ma, Yan aut Enthalten in Empirical software engineering Springer US, 1996 13(2008), 5 vom: 12. Aug., Seite 561-595 (DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 1382-3256 nnns volume:13 year:2008 number:5 day:12 month:08 pages:561-595 https://doi.org/10.1007/s10664-008-9079-3 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 AR 13 2008 5 12 08 561-595 |
allfieldsSound |
10.1007/s10664-008-9079-3 doi (DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p DE-627 ger DE-627 rakwb eng 004 VZ Jiang, Yue verfasserin aut Techniques for evaluating fault prediction models 2008 Text txt rdacontent ohne Hilfsmittel zu benutzen n rdamedia Band nc rdacarrier © Springer Science+Business Media, LLC 2008 Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies Cukic, Bojan aut Ma, Yan aut Enthalten in Empirical software engineering Springer US, 1996 13(2008), 5 vom: 12. Aug., Seite 561-595 (DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 1382-3256 nnns volume:13 year:2008 number:5 day:12 month:08 pages:561-595 https://doi.org/10.1007/s10664-008-9079-3 lizenzpflichtig Volltext GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 AR 13 2008 5 12 08 561-595 |
language |
English |
source |
Enthalten in Empirical software engineering 13(2008), 5 vom: 12. Aug., Seite 561-595 volume:13 year:2008 number:5 day:12 month:08 pages:561-595 |
sourceStr |
Enthalten in Empirical software engineering 13(2008), 5 vom: 12. Aug., Seite 561-595 volume:13 year:2008 number:5 day:12 month:08 pages:561-595 |
format_phy_str_mv |
Article |
institution |
findex.gbv.de |
topic_facet |
Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies |
dewey-raw |
004 |
isfreeaccess_bool |
false |
container_title |
Empirical software engineering |
authorswithroles_txt_mv |
Jiang, Yue @@aut@@ Cukic, Bojan @@aut@@ Ma, Yan @@aut@@ |
publishDateDaySort_date |
2008-08-12T00:00:00Z |
hierarchy_top_id |
235946516 |
dewey-sort |
14 |
id |
OLC2071660064 |
language_de |
englisch |
fullrecord |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">OLC2071660064</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230503052507.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">200819s2008 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s10664-008-9079-3</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2071660064</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s10664-008-9079-3-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Jiang, Yue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Techniques for evaluating fault prediction models</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2008</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© Springer Science+Business Media, LLC 2008</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Fault-prediction models</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Model evaluation</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Predictive models in software engineering</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Empirical studies</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Cukic, Bojan</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Ma, Yan</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Empirical software engineering</subfield><subfield code="d">Springer US, 1996</subfield><subfield code="g">13(2008), 5 vom: 12. Aug., Seite 561-595</subfield><subfield code="w">(DE-627)235946516</subfield><subfield code="w">(DE-600)1401304-6</subfield><subfield code="w">(DE-576)102432406</subfield><subfield code="x">1382-3256</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:13</subfield><subfield code="g">year:2008</subfield><subfield code="g">number:5</subfield><subfield code="g">day:12</subfield><subfield code="g">month:08</subfield><subfield code="g">pages:561-595</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s10664-008-9079-3</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_70</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_100</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">13</subfield><subfield code="j">2008</subfield><subfield code="e">5</subfield><subfield code="b">12</subfield><subfield code="c">08</subfield><subfield code="h">561-595</subfield></datafield></record></collection>
|
author |
Jiang, Yue |
spellingShingle |
Jiang, Yue ddc 004 misc Fault-prediction models misc Model evaluation misc Predictive models in software engineering misc Empirical studies Techniques for evaluating fault prediction models |
authorStr |
Jiang, Yue |
ppnlink_with_tag_str_mv |
@@773@@(DE-627)235946516 |
format |
Article |
dewey-ones |
004 - Data processing & computer science |
delete_txt_mv |
keep |
author_role |
aut aut aut |
collection |
OLC |
remote_str |
false |
illustrated |
Not Illustrated |
issn |
1382-3256 |
topic_title |
004 VZ Techniques for evaluating fault prediction models Fault-prediction models Model evaluation Predictive models in software engineering Empirical studies |
topic |
ddc 004 misc Fault-prediction models misc Model evaluation misc Predictive models in software engineering misc Empirical studies |
topic_unstemmed |
ddc 004 misc Fault-prediction models misc Model evaluation misc Predictive models in software engineering misc Empirical studies |
topic_browse |
ddc 004 misc Fault-prediction models misc Model evaluation misc Predictive models in software engineering misc Empirical studies |
format_facet |
Aufsätze Gedruckte Aufsätze |
format_main_str_mv |
Text Zeitschrift/Artikel |
carriertype_str_mv |
nc |
hierarchy_parent_title |
Empirical software engineering |
hierarchy_parent_id |
235946516 |
dewey-tens |
000 - Computer science, knowledge & systems |
hierarchy_top_title |
Empirical software engineering |
isfreeaccess_txt |
false |
familylinks_str_mv |
(DE-627)235946516 (DE-600)1401304-6 (DE-576)102432406 |
title |
Techniques for evaluating fault prediction models |
ctrlnum |
(DE-627)OLC2071660064 (DE-He213)s10664-008-9079-3-p |
title_full |
Techniques for evaluating fault prediction models |
author_sort |
Jiang, Yue |
journal |
Empirical software engineering |
journalStr |
Empirical software engineering |
lang_code |
eng |
isOA_bool |
false |
dewey-hundreds |
000 - Computer science, information & general works |
recordtype |
marc |
publishDateSort |
2008 |
contenttype_str_mv |
txt |
container_start_page |
561 |
author_browse |
Jiang, Yue Cukic, Bojan Ma, Yan |
container_volume |
13 |
class |
004 VZ |
format_se |
Aufsätze |
author-letter |
Jiang, Yue |
doi_str_mv |
10.1007/s10664-008-9079-3 |
dewey-full |
004 |
title_sort |
techniques for evaluating fault prediction models |
title_auth |
Techniques for evaluating fault prediction models |
abstract |
Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. © Springer Science+Business Media, LLC 2008 |
abstractGer |
Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. © Springer Science+Business Media, LLC 2008 |
abstract_unstemmed |
Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment. © Springer Science+Business Media, LLC 2008 |
collection_details |
GBV_USEFLAG_A SYSFLAG_A GBV_OLC SSG-OLC-MAT GBV_ILN_70 GBV_ILN_100 |
container_issue |
5 |
title_short |
Techniques for evaluating fault prediction models |
url |
https://doi.org/10.1007/s10664-008-9079-3 |
remote_bool |
false |
author2 |
Cukic, Bojan Ma, Yan |
author2Str |
Cukic, Bojan Ma, Yan |
ppnlink |
235946516 |
mediatype_str_mv |
n |
isOA_txt |
false |
hochschulschrift_bool |
false |
doi_str |
10.1007/s10664-008-9079-3 |
up_date |
2024-07-04T03:56:30.941Z |
_version_ |
1803619287373447168 |
fullrecord_marcxml |
<?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01000caa a22002652 4500</leader><controlfield tag="001">OLC2071660064</controlfield><controlfield tag="003">DE-627</controlfield><controlfield tag="005">20230503052507.0</controlfield><controlfield tag="007">tu</controlfield><controlfield tag="008">200819s2008 xx ||||| 00| ||eng c</controlfield><datafield tag="024" ind1="7" ind2=" "><subfield code="a">10.1007/s10664-008-9079-3</subfield><subfield code="2">doi</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-627)OLC2071660064</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-He213)s10664-008-9079-3-p</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-627</subfield><subfield code="b">ger</subfield><subfield code="c">DE-627</subfield><subfield code="e">rakwb</subfield></datafield><datafield tag="041" ind1=" " ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="082" ind1="0" ind2="4"><subfield code="a">004</subfield><subfield code="q">VZ</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Jiang, Yue</subfield><subfield code="e">verfasserin</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Techniques for evaluating fault prediction models</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="c">2008</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">Text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">ohne Hilfsmittel zu benutzen</subfield><subfield code="b">n</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">Band</subfield><subfield code="b">nc</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">© Springer Science+Business Media, LLC 2008</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Abstract Many statistical techniques have been proposed to predict fault-proneness of program modules in software engineering. Choosing the “best” candidate among many available models involves performance assessment and detailed comparison, but these comparisons are not simple due to the applicability of varying performance measures. Classifying a software module as fault-prone implies the application of some verification activities, thus adding to the development cost. Misclassifying a module as fault free carries the risk of system failure, also associated with cost implications. Methodologies for precise evaluation of fault prediction models should be at the core of empirical software engineering research, but have attracted sporadic attention. In this paper, we overview model evaluation techniques. In addition to many techniques that have been used in software engineering studies before, we introduce and discuss the merits of cost curves. Using the data from a public repository, our study demonstrates the strengths and weaknesses of performance evaluation techniques and points to a conclusion that the selection of the “best” model cannot be made without considering project cost characteristics, which are specific in each development environment.</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Fault-prediction models</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Model evaluation</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Predictive models in software engineering</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Empirical studies</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Cukic, Bojan</subfield><subfield code="4">aut</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Ma, Yan</subfield><subfield code="4">aut</subfield></datafield><datafield tag="773" ind1="0" ind2="8"><subfield code="i">Enthalten in</subfield><subfield code="t">Empirical software engineering</subfield><subfield code="d">Springer US, 1996</subfield><subfield code="g">13(2008), 5 vom: 12. Aug., Seite 561-595</subfield><subfield code="w">(DE-627)235946516</subfield><subfield code="w">(DE-600)1401304-6</subfield><subfield code="w">(DE-576)102432406</subfield><subfield code="x">1382-3256</subfield><subfield code="7">nnns</subfield></datafield><datafield tag="773" ind1="1" ind2="8"><subfield code="g">volume:13</subfield><subfield code="g">year:2008</subfield><subfield code="g">number:5</subfield><subfield code="g">day:12</subfield><subfield code="g">month:08</subfield><subfield code="g">pages:561-595</subfield></datafield><datafield tag="856" ind1="4" ind2="1"><subfield code="u">https://doi.org/10.1007/s10664-008-9079-3</subfield><subfield code="z">lizenzpflichtig</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_USEFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SYSFLAG_A</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_OLC</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">SSG-OLC-MAT</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_70</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">GBV_ILN_100</subfield></datafield><datafield tag="951" ind1=" " ind2=" "><subfield code="a">AR</subfield></datafield><datafield tag="952" ind1=" " ind2=" "><subfield code="d">13</subfield><subfield code="j">2008</subfield><subfield code="e">5</subfield><subfield code="b">12</subfield><subfield code="c">08</subfield><subfield code="h">561-595</subfield></datafield></record></collection>
|
score |
7.4007034 |