Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers

Today's software complexity makes developing defect-free software almost impossible. Consequently, developing classifiers to classify software modules into defective and non-defective before software releases have attracted great interest in academia and software industry alike. Although many c...

Full description

Bibliographic Details
Main Authors: Hussam Ghunaim, Julius Dichter
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8710236/
_version_ 1819179171588341760
author Hussam Ghunaim
Julius Dichter
author_facet Hussam Ghunaim
Julius Dichter
author_sort Hussam Ghunaim
collection DOAJ
description Today's software complexity makes developing defect-free software almost impossible. Consequently, developing classifiers to classify software modules into defective and non-defective before software releases have attracted great interest in academia and software industry alike. Although many classifiers have been proposed, no one has been proven superior over others. The major reason is that while a research shows that classifier A is better than classifier B, we can find other research that shows the opposite. These conflicts are usually triggered when researchers report results using their preferable performance evaluation measures such as, recall and precision. Although this approach is valid, it does not examine all possible facets of classifiers performance characteristics. Thus, the performance evaluation might improve or deteriorate if researchers choose other performance measures. As a result, software developers usually struggle to select the most suitable classifier to use in their projects. The goal of this paper is to apply the fuzzy analytical hierarchy process (FAHP) as a popular multicriteria decision-making technique to reliably evaluate classifiers' performance. This evaluation framework incorporates a wider spectrum of performance measures to evaluate classifiers performance rather than relying on selected preferable measures. The results show that this approach will increase software developers' confidence in research outcomes and help them in avoiding false conclusions and infer reasonable boundaries for them. We exploited 22 popular performance measures and 11 software defect classifiers. The analysis was carried out using KNIME data mining platform and 12 software defect data sets provided by the NASA metrics data program (MDP) repository.
first_indexed 2024-12-22T21:54:12Z
format Article
id doaj.art-373e29e9472b4a4193e8e40917e6286e
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-22T21:54:12Z
publishDate 2019-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-373e29e9472b4a4193e8e40917e6286e2022-12-21T18:11:16ZengIEEEIEEE Access2169-35362019-01-017627946280410.1109/ACCESS.2019.29159648710236Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect ClassifiersHussam Ghunaim0https://orcid.org/0000-0002-2859-2189Julius Dichter1Computer Science and Engineering Department, School of Engineering, University of Bridgeport, Bridgeport, CT, USAComputer Science and Engineering Department, School of Engineering, University of Bridgeport, Bridgeport, CT, USAToday's software complexity makes developing defect-free software almost impossible. Consequently, developing classifiers to classify software modules into defective and non-defective before software releases have attracted great interest in academia and software industry alike. Although many classifiers have been proposed, no one has been proven superior over others. The major reason is that while a research shows that classifier A is better than classifier B, we can find other research that shows the opposite. These conflicts are usually triggered when researchers report results using their preferable performance evaluation measures such as, recall and precision. Although this approach is valid, it does not examine all possible facets of classifiers performance characteristics. Thus, the performance evaluation might improve or deteriorate if researchers choose other performance measures. As a result, software developers usually struggle to select the most suitable classifier to use in their projects. The goal of this paper is to apply the fuzzy analytical hierarchy process (FAHP) as a popular multicriteria decision-making technique to reliably evaluate classifiers' performance. This evaluation framework incorporates a wider spectrum of performance measures to evaluate classifiers performance rather than relying on selected preferable measures. The results show that this approach will increase software developers' confidence in research outcomes and help them in avoiding false conclusions and infer reasonable boundaries for them. We exploited 22 popular performance measures and 11 software defect classifiers. The analysis was carried out using KNIME data mining platform and 12 software defect data sets provided by the NASA metrics data program (MDP) repository.https://ieeexplore.ieee.org/document/8710236/Classifiersdata miningempirical software engineering (ESE)FAHPKNIMEperformance evaluation
spellingShingle Hussam Ghunaim
Julius Dichter
Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
IEEE Access
Classifiers
data mining
empirical software engineering (ESE)
FAHP
KNIME
performance evaluation
title Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
title_full Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
title_fullStr Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
title_full_unstemmed Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
title_short Applying the FAHP to Improve the Performance Evaluation Reliability of Software Defect Classifiers
title_sort applying the fahp to improve the performance evaluation reliability of software defect classifiers
topic Classifiers
data mining
empirical software engineering (ESE)
FAHP
KNIME
performance evaluation
url https://ieeexplore.ieee.org/document/8710236/
work_keys_str_mv AT hussamghunaim applyingthefahptoimprovetheperformanceevaluationreliabilityofsoftwaredefectclassifiers
AT juliusdichter applyingthefahptoimprovetheperformanceevaluationreliabilityofsoftwaredefectclassifiers