tion method using wrapper model is proposed. In the
method, the features are selected in dynamic mode,
i.e. the set of selected features can be different for dif-
ferent test objects in contrast to the static mode, where
the selected set of features is the same for all test ob-
jects. In the selection procedure, we formulate the
optimal feature selection problem adopting the sum
of relevance of features and diversity of feature en-
semble as an optimality criterion. Since this problem
can not be directly solved using analytical ways, we
propose to apply genetic algorithm (GA).
The performance of proposed feature selection
method (DFS) was experimentally verified using 7
real benchmark data sets. The DFS method out-
performed the six state-of-art feature selection algo-
rithms in terms of the quality of the feature subset and
the classification accuracy.
There are some avenues for future research. First,
we can consider the cost associated with each fea-
ture, which in the optimization problem (16) can play
the role of constraints. It means, that feature selec-
tion method should maximize the sum of relevance
of features and diversity of feature ensemble in dy-
namic fashion, and simultaneously should keep the
cost of measure of member features on an acceptable
level. Second, we can apply for solving optimization
problem (16) other heuristic optimization procedures,
e.g. the simulated annealing (SA) algorithm. As it
results from the authors’ earlier experience (Lysiak
et al., 2014), the SA method is faster than the GA
algorithm, which can have great practical importance.
ACKNOWLEDGEMENTS
We would like to thank the anonymous reviewers for
their constructive comments and helpful suggestions.
This work was financed from the National Science
Center resources in 2012-2014 as a research project
No ST6/06168 and supported by the statutory funds
of the Department of Systems and Computer Net-
works, Wroclaw University of Technology.
REFERENCES
Bache, K. and Lichman, M. (2013). UCI machine learning
repository.
Bolon-Canedo, V., Sanchez-Marono, N., and Alonzo-
Betandos, A. (2012). A review of feature selection
methods on synthetic data. Knowledge Information
System, 34:483–519.
Chandrashekar, G. and Sahin, F. (2014). A survey on fea-
ture selection methods. Computers and Electrical En-
gineering, 40:16–28.
Demˇsar, J. (2006). Statistical comparisons of classifiers
over multiple data sets. The Journal of Machine
Learning Research, 7:1–30.
Dietterich, T. (1998). Approximate statistical tests for com-
paring supervised classification learning algorithms.
Neural Computation, 10:1895–1923.
Duda, R., Hart, P., and Stork, D. (2012). Pattern Classifica-
tion. Wiley Interscience, New York.
Duin, R., Juszczak, P., Pekalska, E., and et al. (2007). A
matlab toolbox for pattern recognition.
Goldberg, D. (1989). Genetic Algorithms in Search, Op-
timization and Machine learning. Addison-Wesley
Publishing Company, London.
Gu, Q., Li, Z., and Han, J. (2012). Generalized fisher score
for feature selection. CoRR, abs/1202.3725.
Guyon, I. and Elisseeff, A. (2003). An introduction to vari-
able and feature selection. Journal on Machine Learn-
ing Research, 3:1157–1182.
Kuncheva, L. (2004a). Combining Pattern Classifier: Meth-
ods and Algorithms. Wiley-Interscience, London.
Kuncheva, L. (2004b). Ludmila kuncheva collection.
Lysiak, R., Kurzynski, M., and Woloszynski, T. (2014). Op-
timal selection of ensemble classifiers using measures
of competence and diversity of base classifiers. Neu-
rocomputing, 126:29–35.
Saeys, Y., Abeel, T., and Van de Peer, Y. (2008). Ro-
bust feature selection using ensemble feature selec-
tion techiques. Lecture Notes in Artificial Intelligence,
5212:313–325.
Saeys, Y., Inza, I., and Larranaga, P. (2007). A review of
feature selection techniques in bioinformatics. Bioin-
formatics, 23:2507–2517.
Wang, H., Khoshgoftaar, T., and Napolitano, A. (2010).
A comparative study of ensemble feature selection
techniques for software defect prediction. In 2010
Ninth Int. Conf. on Machine Learning and Applica-
tions, pages 135–140. IEEE Computer Society.
Woloszynski, T. (2013). Classifier competence based on
probabilistic modeling (ccprmod.m) at matlab central
file exchange.
Woloszynski, T. and Kurzynski, M. (2010). A measure
of competence based on randomized reference classi-
fier for dynamic ensemble selection. In 2010 Twenti-
eth International Conference on Pattern Recognition,,
pages 4194–4197. Int. Association on Pattern Recog-
nition.
Woloszynski, T. and Kurzynski, M. (2011). A probabilistic
model of classifier competence for dynamic ensemble
selection. Pattern Recognition, 44(10-11):2656–2668.
Woloszynski, T., Kurzynski, M., Podsiadlo, P., and Sta-
chowiak, G. (2012). A measure of competence based
on random classification for dynamic ensemble selec-
tion. Information Fusion, 13:207–213.
Wolpert, D. H. (1992). Stacked generalization. Neural Net-
works, 5:214–259.
Zafra, A., Pechenizkiy, M., and Ventura, S. (2010). Reduc-
ing dimensionality in multiple instance learning with
a filter method. Lecture Notes in Computer Science,
6077:35–44.