5 CONCLUSIONS
We have tested five types of neural networks and
three different methods of diversity enhancement.
We have proposed the one-cycle learning method for
neural networks, same as the method of diversity
enhancement which we call input filtering. Based on
the experimental study results we can formulate the
following outcomes:
Neural networks in general look suitable as
the base algorithms for classifiers ensembles.
The method of one-cycle learning looks
suitable for ensembles building too.
Filtering gave surprisingly good results as the
method of diversity enhancement.
Doubling of patterns gave surprisingly well
results too. We expected that this method
would lead to over-fitting, but this assumption
did not prove correct.
We expected more from shuffling of patterns.
But as the results show, doubling of patterns is
more permissible.
Boosting results are shown in Table 6 looking at it,
we can pronounce the following:
The best performance (train error 0.062%, test
error 2.89%) has been reached with the Back
Propagation network with 50 hidden neurons.
The worst performance has been reached with
the Hebbian network.
Green colour represents our results in Figure
6, which are promising by comparison with
other approaches.
Adaboost, neural networks and input filters look
as a very promising combination. Although we have
used only random filters, the performance of the
combined classifier was satisfactory. We have
proved the positive influence of input filters.
Nevertheless the random method of input filters
selecting makes the adaptation process very time
consuming. We have to look for more sophisticated
methods of detecting problematic areas in the
patterns. Once such areas are found, we will able to
design and possibly generalize some method of the
bespoke input filter construction.
ACKNOWLEDGEMENTS
The research steps described here have been
financially supported by the University of Ostrava
grant SGS16/PrF/2014.
REFERENCES
Breiman, L. (1996). Bagging predictors. In Machine
Learning. (pp. 123-140).
Davidson, I., and Fan, W. (2006). When efficient model
averaging out-performs boosting and bagging.
Knowledge Discovery in Databases, 478-486.
Dietterich, T. G. (2000). An experimental comparison of
three methods for constructing ensembles of decision
trees: Bagging, boosting, and randomization. Machine
learning, 139-157.
Drucker, H., Schapire, R., and Simard, P. (1993). Boosting
performance in neural networks. Int. Journ. of Pattern
Recognition. and Artificial Intelligence, 705-719.
Fausett, L. V. (1994). Fundamentals of Neural Networks.
Englewood Cliffs, New Jersey: Prentice-Hall, Inc.
Freund, Y. (1995). Boosting a weak learning algorithm by
majority. Information and Computation, 256-285.
Freund, Y., and Schapire, R. (1996). Experiments with a
New Boosting Algorithm. ICML, 148-156.
Freund, Y., and Schapire, R. E. (1997). A decision-
theoretic generalization of on-line learning and an
application to boosting. J. of Comp. and System
Sciences, 119–139.
Freund, Y., and Schapire, R. (1999). A short introduction
to boosting. J. Japan. Soc. for Artif. Intel., 771-780.
Iwakura, T., Okamoto, S., and Asakawa, K. (2010). An
AdaBoost Using a Weak-Learner Generating Several
Weak-Hypotheses for Large Training Data of Natural
Language Processing. IEEJ Transactions on
Electronics, Information and Systems, 83-91.
Kocian, V., Volná, E., Janošek, M., and Kotyrba, M.
(2011). Optimizatinon of training sets for Hebbian-
learningbased classifiers. In Proc. of the 17th Int.
Conf. on Soft Computing, Mendel 2011, pp. 185-190.
Kocian, V., and Volná, E. (2012). Ensembles of neural-
networks-based classifiers. In Proc. of the 18th Int.
Conf. on Soft Computing, Mendel 2012, pp. 256-261.
LeCun, Y. Bottou, L. Bengio, Y. and Haffner, P. (1998)
Gradient-based learning applied to document
recognition. Proceedings of the IEEE, 86(11):2278-
2324, November.
LeCun, Y., Cortes, C., and Burges, C. (2014). Te MNIST
Database. Retrieved from http://yann.lecun.com/
exdb/mnist/
Quinlan, J. R. (1996). Bagging, boosting, and C4.5.
Thirteenth National Conference on Artificial
Intelligence, (pp. 725-730).
Schapire, R. E. (1990). The strength of weak learnability.
Machine Learning, 197-227.
Schapire, R. E. (1999). A brief introduction to boosting.
Sixteenth International Joint Conference on Artificial
Intelligence IJCAI (pp. 1401–1406). Morgan
Kaufmann Publishers Inc.
Valiant, L. G. (1984). A theory of the learnable.
Communications of the ACM, 1134-1142.
BoostingofNeuralNetworksoverMNISTData
263