also brings up missclassifications consistent across
all individual models, which are interesting as they
showcase model, training method or dataset limita-
tions. We show all such examples for models trained
on original GTSRB images in Figure 4.
These experiments provide valuable insight into
how committees boost model performance scores and
help us with assumptions on what they can and cannot
do.
5 CONCLUSION
In this paper and work of others we observe some
consistencies in results achieved using committees
of base deep models. For considered problems that
are not saturated, even smaller committees improve
recognition rates by a value close to 2%. However,
when room for improvement is much smaller, com-
mittees need to be much larger or built with greater
care to be reliable, as smaller committees could have
a significant amount of wrongly classified examples
when individual models make similar errors. We
show statistics for committees of various sizes on two
datasets, trained on original or preprocessed images,
as well as hybrid committees. When using a single
preprocessing method to build committees, the in-
crease achieved is similar and the final recognition
rate depends largely on average performance of in-
dividual models. Hybrid committees prove more of
a challenge, since the right choice of preprocessing
method combinations can boost or reduce results de-
pending on whether the preprocessing methods prove
compatible for that dataset and model.
We also looked into performance metrics specific
for committees since they can only improve results on
examples that individual models do not consistently
classify. Defining base correct classification rate as
the examples all individual models classify correctly,
we calculated true improvement as the increase of
correct classifications above the base. Results of this
metric showed an ∼ 117% increase on GTSRB and
∼ 110% increase on CIFAR-10, giving a much better
insight on how much committees help rather than just
the increase in recognition rate. Overall, we brought
to light intricacies of a much used but not elaborated
approach to boost final model performance.
ACKNOWLEDGEMENT
This work has been supported by the Croatian Science
Foundation under the project UIP-11-2013-1544.
REFERENCES
Ciresan, D. C., Meier, U., Masci, J., and Schmidhuber,
J. (2011). A committee of neural networks for traf-
fic sign classification. In IJCNN, pages 1918–1921.
IEEE.
Ciresan, D. C., Meier, U., and Schmidhuber, J. (2012).
Multi-column deep neural networks for image classi-
fication. In CVPR, pages 3642–3649. IEEE.
Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long,
J., Girshick, R. B., Guadarrama, S., and Darrell, T.
(2014). Caffe: Convolutional architecture for fast fea-
ture embedding. CoRR, abs/1408.5093.
Jin, J., Fu, K., and Zhang, C. (2014). Traffic sign recog-
nition with hinge loss trained convolutional neural
networks. Intelligent Transportation Systems, IEEE
Transactions on, PP(99):1–10.
Krizhevsky, A. (2009). Learning multiple layers of features
from tiny images.
Krizhevsky, A. (2014). One weird trick for parallelizing
convolutional neural networks. CoRR, abs/1404.5997.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Im-
agenet classification with deep convolutional neural
networks. In Advances in Neural Information Pro-
cessing Systems 25, pages 1097–1105. Curran Asso-
ciates, Inc.
Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S.,
Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bern-
stein, M. S., Berg, A. C., and Fei-Fei, L. (2014). Ima-
genet large scale visual recognition challenge. CoRR,
abs/1409.0575.
Simonyan, K. and Zisserman, A. (2014). Very deep con-
volutional networks for large-scale image recognition.
CoRR.
Stallkamp, J., Schlipsing, M., Salmen, J., and Igel, C.
(2011). The German Traffic Sign Recognition Bench-
mark: A multi-class classification competition. In
IEEE International Joint Conference on Neural Net-
works, pages 1453–1460.
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.,
Anguelov, D., Erhan, D., Vanhoucke, V., and Rabi-
novich, A. (2015). Going deeper with convolutions.
In The IEEE Conference on Computer Vision and Pat-
tern Recognition (CVPR).
Zeiler, M. D. and Fergus, R. (2014). Visualizing and under-
standing convolutional networks. In Computer Vision
- ECCV 2014 - 13th European Conference, Zurich,
pages 818–833.
Evaluating the Effects of Convolutional Neural Network Committees
565