an augmentation of the dataset. Also, when com-
paring prediction times for a single sample versus
multiple samples, we saw no differences at all, giv-
ing the possibility of parallel classification of multi-
ple signs per video frame, instead of just one sign.
In all cases, the data collected is hardware dependent,
as demonstrated, but with the current state-of-the-art
GPUs, real-time capability is indeed achievable, even
though, from our experiments, some of the deepest
models did not show fast enough results, especially
when we consider the fact that we are simply talk-
ing about classification time and not account for the
actual traffic sign localization within the video frame
captured. Several different processors ran six differ-
ent CNN Deep Learning models. The models were
trained and verified to identify German traffic sign
images with varying orientations and augmentations
from the GTSRB benchmark dataset with a high ac-
curacy. With some variations to the dataset, identifi-
cation of images occurred in less than one millisec-
ond on one of the processors. In others cases, with
deeper networks, more specifically with multiple con-
volutional layers, the results were not as good in com-
parison with shallower networks along with a depen-
dency on the processor used. Our model 6 has just
three convolutional layers and it gave the best accu-
racy. After a certain depth, the features that the net-
work looks for may not make sense for the classifi-
cation purposes and we believe this could possibly be
one of the reasons for the multiple convolutional lay-
ers not performing better. We aim to further improve
prevention of gradient vanishing and use deep resid-
ual learning with deeper networks to observe their be-
havior. As future work, we can modify the dataset
by converting the images to gray-scale, known to be
helpful to improve accuracy as it was the case with the
winner in the GTSRB competition. Another aspect to
consider is the distribution of the images within the
classes which are not homogeneous. Most common
traffic signs have up to 3 thousand images, while the
least common ones have only a few hundred. In this
case, a reduced version of the dataset, by removing
classes where the number of images is less than 1000,
could give us a more homogeneous dataset, and hope-
fully an increase in overall accuracy, by not relying
on classes with too few trained samples. And yet an-
other aspect to be considered is to fine-tune the mod-
els used in the experiment, in terms of activation and
loss functions and its parameters. In terms of hard-
ware, this experiment was limited to desktop or note-
book PCs, and so, it did not consider some of the em-
bedded platforms that are intended to real-time appli-
cations like traffic sign classification. We intend to
implement these networks on NVDIA Tegra X1 de-
velopment board and achieve great increment in the
classification speed for real-time applications.
REFERENCES
Bojarski, M., Del Testa, D., Dworakowski, D., Firner, B.,
Flepp, B., Goyal, P., Jackel, L. D., Monfort, M.,
Muller, U., and Zhang, J. e. a. (2017). End to end
learning for self-driving cars. Arxiv.org.
Chen, L., Li, Q., Li, M., Zhang, L., and Mao, Q. (2012). De-
sign of a multi-sensor cooperation travel environment
perception system for autonomous vehicle. Sensors,
12(12):12386–12404.
Ciresan, D., Meier, U., Masci, J., and Schmidhuber, J.
(2011). A committee of neural networks for traffic
sign classification. The 2011 International Joint Con-
ference on Neural Networks.
Greenhalgh, JackMirmehdi, M. (2012). Real-time de-
tection and recognition of road traffic signs. IEEE
Transactions on Intelligent Transportation Systems,
13(4):1498–1506.
Igel, C. German traffic sign recognition
benchmark gtsrb dataset available at:
http://benchmark.ini.rub.de/?section=home.
Jin, J., Fu, K., and Zhang, C. (2014). Traffic sign recogni-
tion with hinge loss trained convolutional neural net-
works. IEEE Transactions on Intelligent Transporta-
tion Systems, 15(5):1991–2000.
Le, T. T., Tran, S. T., Mita, S., and Nguyen, T. D. (2010).
Real time traffic sign detection using color and shape-
based features. Intelligent Information and Database
Systems, pages 268–278.
Mao, X., Hijazi, S., Casas, R., Kaul, P., Kumar, R., and
Rowen, C. (2016). Hierarchical cnn for traffic sign
recognition. 2016 IEEE Intelligent Vehicles Sympo-
sium (IV).
Menge, Anheuser-Busch, O. (2016). Complete first self-
driving truck delivery of beer. Neural Networks.
Sermanet, PierreLeCun, Y. (2011). Traffic sign recognition
with multi-scale convolutional networks. The 2011
International Joint Conference on Neural Networks.
Stallkamp, J., Schlipsing, M., Salmen, J., and Igel, C.
(2012). Man vs. computer: Benchmarking machine
learning algorithms for traffic sign recognition. Neu-
ral Networks, 32:323–332.
Trivedi, M. M. Lisa dataset. available at:
http://cvrr.ucsd.edu. Cvrr.ucsd.edu.
Zeng, Y., Xu, X., Fang, Y., and Zhao, K. (2015). Traffic
sign recognition using deep convolutional networks
and extreme learning machine. Lecture Notes in Com-
puter Science, pages 272–280.
Empirical Evaluation of Convolutional Neural Networks Prediction Time in Classifying German Traffic Signs
267