As future work, we intend to explore different rele-
vance and redundancy measures for supervised and
unsupervised problems as well as fine tuning of the
threshold used by the algorithm. We also intend to
devise a strategy to lower the time taken to compute
the redundancy between features, which is the most
time consuming part of the proposed method.
REFERENCES
Alcal
´
a-Fdez, J., Fernandez, A., Luengo, J., Derrac, J.,
Garc
´
ıa, S., S
´
anchez, L., and Herrera, F. (2011). KEEL
data-mining software tool: data set repository, integra-
tion of algorithms and experimental analysis frame-
work. Journal of Multiple-Valued Logic and Soft
Computing, 17(2-3):255–287.
Bishop, C. (1995). Neural Networks for Pattern Recogni-
tion. Oxford University Press.
Bishop, C. (2007). Pattern Recognition and Machine
Learning. Springer.
Chandrashekar, G. and Sahin, F. (2014). A survey on fea-
ture selection methods. Computers & Electrical En-
gineering, 40(1):16 – 28. 40th-year commemorative
issue.
Cover, T. and Thomas, J. (2006). Elements of information
theory. John Wiley & Sons, second edition.
Dua, D. and Graff, C. (2019). UCI machine learning repos-
itory.
Duda, R., Hart, P., and Stork, D. (2001). Pattern classifica-
tion. John Wiley & Sons, second edition.
Fayyad, U. and Irani, K. (1993). Multi-interval discretiza-
tion of continuous-valued attributes for classification
learning. In Proceedings of the International Joint
Conference on Uncertainty in AI, pages 1022–1027.
Ferreira, A. and Figueiredo, M. (2012). Efficient feature
selection filters for high-dimensional data. Pattern
Recognition Letters, 33(13):1794 – 1804.
Frank, E., Hall, M., and Witten, I. (2016). The weka
workbench. online appendix for ”data mining: Prac-
tical machine learning tools and techniques”. Morgan
Kaufmann.
Garcia, S., Luengo, J., and Herrera, F. (2016). Tutorial on
practical tips of the most influential data preprocess-
ing algorithms in data mining. Knowledge-Based Sys-
tems, 98:1 – 29.
Garcia, S., Luengo, J., Saez, J., Lopez, V., and Herrera, F.
(2013). A survey of discretization techniques: tax-
onomy and empirical analysis in supervised learning.
IEEE Transactions on Knowledge and Data Engineer-
ing, 25(4):734–750.
Guyon, I. and Elisseeff, A. (2003). An introduction to vari-
able and feature selection. Journal of Machine Learn-
ing Research (JMLR), 3:1157–1182.
Guyon, I., Gunn, S., Nikravesh, M., and Zadeh (Editors), L.
(2006). Feature extraction, foundations and applica-
tions. Springer.
Hemada, B. and Lakshmi, K. (2013). A study on discretiza-
tion techniques. International journal of engineering
research and technology, 2(8).
John, G., Kohavi, R. ., and Pfleger, K. (1994). Irrelevant fea-
tures and the subset selection problem. In Proceedings
of the International Conference on Machine Learning
(ICML), pages 121–129. Morgan Kaufmann.
Miao, J. and Niu, L. (2016). A survey on feature selection.
Procedia Computer Science, 91:919 – 926. Promot-
ing Business Analytics and Quantitative Management
of Technology: 4th International Conference on In-
formation Technology and Quantitative Management
(ITQM 2016).
Ooi, C., Chetty, M., and Teng, S. (2005). Relevance, re-
dundancy and differential prioritization in feature se-
lection for multi-class gene expression data. In Pro-
ceedings of the International Conference on Biologi-
cal and Medical Data Analysis (ISBMDA), pages 367–
378, Berlin, Heidelberg. Springer-Verlag.
Peng, H., Long, F., and Ding, C. (2005). Feature selec-
tion based on mutual information: criteria of max-
dependency, max-relevance, and min-redundancy.
IEEE Transactions on Pattern Analysis and Machine
Intelligence (PAMI), 27(8):1226–1238.
Witten, I., Frank, E., Hall, M., and Pal, C. (2016). Data min-
ing: practical machine learning tools and techniques.
Morgan Kauffmann, fourth edition.
Yu, L. and Liu, H. (2003). Feature selection for high-
dimensional data: a fast correlation-based filter solu-
tion. In Proceedings of the International Conference
on Machine Learning (ICML), pages 856–863.
Yu, L. and Liu, H. (2004). Efficient feature selection via
analysis of relevance and redundancy. Journal of Ma-
chine Learning Research (JMLR), 5:1205–1224.
Zhang, L., Li, Z., Chen, H., and Wen, J. (2006). Minimum
redundancy gene selection based on grey relational
analysis. In Proceedings of the IEEE International
Conference on Data Mining - Workshops (ICDMW),
pages 120–124, Washington, DC, USA. IEEE Com-
puter Society.
Zhao, Z., Morstatter, F., Sharma, S., Alelyani, S., Anand,
A., and Liu, H. (2010). Advancing feature selection
research - ASU feature selection repository. Techni-
cal report, Computer Science & Engineering, Arizona
State University.
ICPRAM 2021 - 10th International Conference on Pattern Recognition Applications and Methods
372