network. To this end, we generated 15 noisy im-
ages for each image in the training set with different
signal-to-noise ratios. Then, the ConvNets are trained
using the new noisy training sets. Finally, we evaluate
the ConvNets using the noisy test sets. Table 6 illus-
trates the result. Because of space limitation we could
not include the results from the CIFAR10 and the
MNIST datasets. However, the complete results are
available at deim.urv.cat/ rivi/cnn-noise-tolerance/.
Result: While augmenting the training set with
noisy images improves the performance, however, we
observe that the ConvNets are still sensitive to noise.
For instance, the scatter plots beside Table 6 shows
that even after training with noisy training set, it is
still possible to generate a Gaussian noise with σ = 1
in order to incorrectly classify the images.
5 CONCLUSIONS
In this paper we studied the degree of which Con-
vNets are tolerant against noise. For this purpose,
we first proposed a method for finding the minimum
noisy image close to the decision boundary that is
misclassified by the ConvNet. We applied our method
on the ConvNets trained on the CIFAR10, the MNIST
and the GTSRB datasets and showed that it is possi-
ble to generate low magnitude noises which are hardly
perceivable by human eyes but they alter the classifi-
cation score of the ConvNets. Then, we carried out
several experiments to study different aspects of sta-
bility. First, we randomly generated many noisy im-
ages with various signal-to-noise ratios and classified
them using the three ConvNets. We found out that the
three ConvNets makes mistakes even with very low
magnitude noisy images. This can be explained by the
fact that the inter-class margin of the feature vectors
computed by ConvNets might be very small. Another
possibility is that because ConvNets are highly non-
linear functions, a small change in the input causes a
significant change in the output. For these two rea-
sons, images may fall into wrong classes when they
are degraded by a low magnitude noise. Second, we
examined the effect of ensemble of ConvNets and
found that although ensembles improve the classifica-
tion accuracy but they are still very vulnerable to low
magnitude noises. Third, we investigated the effect
of augmenting the training datasets with many noisy
images on the stability. Results reveal that even Con-
vNets trained on noisy datasets are not stable against
noise and they easily make mistakes by low magni-
tude noises.
ACKNOWLEDGEMENTS
The authors are grateful for the support granted by
Generalitat de Catalunya’s Ag
`
ecia de Gesti
´
o d’Ajuts
Universitaris i de Recerca (AGAUR) through FI-DGR
2015 fellowship.
REFERENCES
Aghdam, H. H., Heravi, E. J., and Puig, D. (2015). Rec-
ognizing Traffic Signs using a Practical Deep Neu-
ral Network. In Second Iberian Robotics Conference,
Lisbon. Springer.
Ba, L. and Caurana, R. (2013). Do Deep Nets Really Need
to be Deep ? arXiv preprint arXiv:1312.6184, pages
1–6.
Cirean, D., Meier, U., Masci, J., and Schmidhuber, J.
(2012). Multi-column deep neural network for traf-
fic sign classification. Neural Networks, 32:333–338.
Coates, A. and Ng, A. (2011). Selecting Receptive Fields
in Deep Networks. Nips, (i):1–9.
Dosovitskiy, A. and Brox, T. (2015). Inverting Convolu-
tional Networks with Convolutional Networks. pages
1–15.
Girshick, R., Donahue, J., Darrell, T., Berkeley, U. C., and
Malik, J. (2014). Rich feature hierarchies for accurate
object detection and semantic segmentation. Cvpr’14,
pages 2–9.
Glorot, X. and Bengio, Y. (2010). Understanding the
difficulty of training deep feedforward neural net-
works. Proceedings of the 13th International Con-
ference on Artificial Intelligence and Statistics (AIS-
TATS), 9:249–256.
Goodfellow, I., Mirza, M., Da, X., Courville, A., and
Bengio, Y. (2013). An Empirical Investigation of
Catastrophic Forgeting in Gradient-Based Neural Net-
works. arXiv preprint arXiv: . . . .
He, K., Zhang, X., Ren, S., and Sun, J. Delving Deep into
Rectifiers: Surpassing Human-Level Performance on
ImageNet Classification.
Jin, J., Fu, K., and Zhang, C. (2014). Traffic Sign Recog-
nition With Hinge Loss Trained Convolutional Neural
Networks. IEEE Transactions on Intelligent Trans-
portation Systems, 15(5):1991–2000.
Krizhevsky, A. (2009). Learning Multiple Layers of Fea-
tures from Tiny Images. pages 1–60.
Krizhevsky, a., Sutskever, I., and Hinton, G. (2012). Im-
agenet classification with deep convolutional neural
networks. Advances in neural information processing
systems, pages 1097–1105.
LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998).
Gradient-based learning applied to document recogni-
tion. Proceedings of the IEEE, 86(11):2278–2323.
Mahendran, A. and Vedaldi, A. (2014). Understanding
Deep Image Representations by Inverting Them.
Nguyen, a., Yosinski, J., and Clune, J. (2015). Deep Neural
Networks are Easily Fooled: High Confidence Predic-
tions for Unrecognizable Images. Cvpr 2015.
Analyzing the Stability of Convolutional Neural Networks against Image Degradation
381