helped to achieve the highest accuracy among its
contenders. However, the presence of outliers and
imbalance problems in data might result in the
degraded performance of ADSD. Moreover, when the
number of samples grows very high ADSD cannot
deal with it due to the huge computational complexity
of kNN. To address this, we can take advantage of the
sampling method which will be addressed in future.
ACKNOWLEDGEMENTS
This research is supported by the fellowship from
ICT Division, Ministry of Posts, Telecommunications
and Information Technology, Bangladesh. No -
56.00.0000.028.33.006.20-84; Dated 13.04.2021.
REFERENCES
Agarwal, S., Tomar, D., et al. (2014). Prediction of software
defects using twin support vector machine. In 2014
International Conference on Information Systems and
Computer Networks (ISCON), pages 128–132. IEEE.
Akhter, S., Sharmin, S., Ahmed, S., Sajib, A. A., and
Shoyaib, M. (2021). mrelief: A reward penalty based
feature subset selection considering data overlapping
problem. In Computational Science–ICCS 2021:
21st International Conference, Krakow, Poland, June
16–18, 2021, Proceedings, Part I, pages 278–292.
Springer.
Akintola, A. G., Balogun, A. O., Lafenwa-Balogun, F. B.,
and Mojeed, H. A. (2018). Comparative analysis
of selected heterogeneous classifiers for software
defects prediction using filter-based feature selection
methods. FUOYE Journal of Engineering and
Technology, 3(1):134–137.
Erlikh, L. (2000). Leveraging legacy system dollars for e-
business. IT professional, 2(3):17–23.
Gao, W., Hu, L., and Zhang, P. (2020). Feature redundancy
term variation for mutual information-based feature
selection. Applied Intelligence, 50(4):1272–1288.
Goh, L., Song, Q., and Kasabov, N. (2004). A novel
feature selection method to improve classification of
gene expression data. In Proceedings of the second
conference on Asia-Pacific bioinformatics-Volume 29,
pages 161–166. Australian Computer Society, Inc.
Greene, C. S., Penrod, N. M., Kiralis, J., and Moore,
J. H. (2009a). Spatially uniform relieff (surf)
for computationally-efficient filtering of gene-gene
interactions. BioData mining, 2(1):1–9.
Greene, C. S., Penrod, N. M., Kiralis, J., and Moore,
J. H. (2009b). Spatially uniform relieff (surf)
for computationally-efficient filtering of gene-gene
interactions. BioData mining, 2(1):1–9.
Ibrahim, D. R., Ghnemat, R., and Hudaib, A. (2017).
Software defect prediction using feature selection
and random forest algorithm. In 2017 International
Conference on New Trends in Computing Sciences
(ICTCS), pages 252–257. IEEE.
Igor, K. (1994). Estimating attributes: analysis and
extensions of relief. European conference on machine
learning. Springer, Berlin, Heidelberg.
Jimoh, R., Balogun, A., Bajeh, A., and Ajayi, S. (2018).
A promethee based evaluation of software defect
predictors. Journal of Computer Science and Its
Application, 25(1):106–119.
Jo, I., Lee, S., and Oh, S. (2019). Improved measures of
redundancy and relevance for mrmr feature selection.
Computers, 8(2):42.
Khan, M. S. H., Roy, P., Khanam, F., Hera, F. H.,
and Das, A. K. (2019). An efficient resource
allocation mechanism for time-sensitive data in dew
computing. In 2019 International Conference of
Artificial Intelligence and Information Technology
(ICAIIT), pages 506–510. IEEE.
Kira, Kenji, and Rendell, L. A. (1992). The feature selection
problem: traditional method and a new algorithm.
AAAI, 2:129–134.
Lamkanfi, A., Demeyer, S., Giger, E., and Goethals, B.
(2010). Predicting the severity of a reported bug.
In 2010 7th IEEE Working Conference on Mining
Software Repositories (MSR 2010), pages 1–10. IEEE.
Laradji, I. H., Alshayeb, M., and Ghouti, L. (2015).
Software defect prediction using ensemble learning
on selected features. Information and Software
Technology, 58:388–402.
Lessmann, S., Baesens, B., Mues, C., and Pietsch, S.
(2008). Benchmarking classification models for
software defect prediction: A proposed framework
and novel findings. IEEE Transactions on Software
Engineering, 34(4):485–496.
Lewis, D. D. (1992). Feature selection and feature
extraction for text categorization. In Speech and
Natural Language: Proceedings of a Workshop Held
at Harriman, New York, February 23-26, 1992.
Menzies, T., Greenwald, J., and Frank, A. (2006). Data
mining static code attributes to learn defect predictors.
IEEE transactions on software engineering, 33(1):2–
13.
Moody, J. and Yang, H. (1999). Data visualization and
feature selection: New algorithms for nongaussian
data. Advances in neural information processing
systems, 12:687–693.
Nguyen Xuan Vinh, Shuo Zhou, J. C. and Bailey, J.
(2016). Can high-order dependencies improve mutual
information based feature selection? Pattern
Recognition, 53:46—-58.
Peng, H., Long, F., and Ding, C. (2005). Feature
selection based on mutual information criteria of max-
dependency, max-relevance, and min-redundancy.
IEEE Transactions on pattern analysis and machine
intelligence, 27(8):1226–1238.
Pisheh, F. and Vilalta, R. (2019). Filter-based information-
theoretic feature selection. Proceedings of the
2019 3rd International Conference on Advances in
Artificial Intelligence.
ENASE 2023 - 18th International Conference on Evaluation of Novel Approaches to Software Engineering
156