tos of different environments.
As a future work, we consider the application
of different feature extraction methods, such as co-
occurrence matrix and SURF, and different models of
CNNs to the classification task to improve the results.
Furthermore, we intend to design a wearable device
to incorporate the software.
ACKNOWLEDGEMENTS
The authors would like to thank all the people who
contributed with images for the elaboration of this da-
taset, the Brazilian Federal Agency for Support and
Evaluation of Graduate Education within the Mini-
stry of Education of Brazil (Capes) and the Natio-
nal Council for Scientific and Technological Develop-
ment (CNPq) for its financial support on this work.
REFERENCES
Bashiri, F. S., LaRose, E., Peissig, P., and Tafti, A. P. (2018).
Mcindoor20000: A fully-labeled image dataset to ad-
vance indoor objects detection. Data in Brief, 17:71 –
75.
Bay, H., Ess, A., Tuytelaars, T., and Gool, L. V. (2008).
Speeded-up robust features (surf). Computer Vision
and Image Understanding, 110(3):346 – 359. Simila-
rity Matching in Computer Vision and Multimedia.
Costa, Y., Oliveira, L., Koerich, A., Gouyon, F., and Mar-
tins, J. (2012). Music genre classification using lbp
textural features. Signal Processing, 92(11):2723 –
2737.
Dalal, N. and Triggs, B. (2005). Histograms of oriented
gradients for human detection. In 2005 IEEE Compu-
ter Society Conference on Computer Vision and Pat-
tern Recognition (CVPR’05), volume 1, pages 886–
893 vol. 1.
Glorot, X. and Bengio, Y. (2010). Understanding the dif-
ficulty of training deep feedforward neural networks.
In Proceedings of the Thirteenth International Con-
ference on Artificial Intelligence and Statistics, vo-
lume 9 of Proceedings of Machine Learning Research,
pages 249–256, Chia Laguna Resort, Sardinia, Italy.
PMLR.
Gonzalez, R. C. and Woods, R. E. (2006). Digital Image
Processing (3rd Edition). Prentice-Hall, Inc., Upper
Saddle River, NJ, USA.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning. MIT Press.
He, K., Zhang, X., Ren, S., and Sun, J. (2015). Delving
deep into rectifiers: Surpassing human-level perfor-
mance on imagenet classification. In 2015 IEEE In-
ternational Conference on Computer Vision (ICCV),
pages 1026–1034.
Jolliffe, I. (2005). Principal Component Analysis. Ameri-
can Cancer Society.
Kingma, D. P. and Ba, J. (2015). Adam: A method for
stochastic optimization. International Conference on
Learning Representations.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012).
Imagenet classification with deep convolutional neu-
ral networks. In Proceedings of the 25th Internatio-
nal Conference on Neural Information Processing Sy-
stems - Volume 1, NIPS’12, pages 1097–1105, USA.
Curran Associates Inc.
Kunene, D., Vadapalli, H., and Cronje, J. (2016). Indoor
sign recognition for the blind. In Proceedings of the
Annual Conference of the South African Institute of
Computer Scientists and Information Technologists,
SAICSIT ’16, pages 19:1–19:9, New York, NY, USA.
ACM.
Lecun, Y. (1989). Generalization and network design stra-
tegies. Elsevier, Zurich, Switzerland.
Lowe, D. G. (2004). Distinctive image features from scale-
invariant keypoints. International Journal of Compu-
ter Vision, 60(2):91–110.
Maas, A. L., Hannun, A. Y., and Ng, A. Y. (2013). Rectifier
nonlinearities improve neural network acoustic mo-
dels. In ICML Workshop on Deep Learning for Audio,
Speech and Language Processing.
Mitchell, T. M. (1997). Machine Learning. McGraw-Hill,
Inc., New York, NY, USA, 1 edition.
Ni, Z., Fu, S., Tang, B., He, H., and Huang, X. (2014).
Experimental studies on indoor sign recognition and
classification. In 2014 IEEE Symposium on Compu-
tational Intelligence and Data Mining (CIDM), pages
489–494.
Ojala, T., Pietik
¨
ainen, M., and Harwood, D. (1996). A com-
parative study of texture measures with classification
based on featured distributions. Pattern Recognition,
29(1):51 – 59.
Prajapati, R. and Shah, P. (2016). Design and testing algo-
rithm for real time text images: Rehabilitation aid for
blind. International Journal of Science Technology &
Engineering, 2(11):275 – 278.
Simonyan, K. and Zisserman, A. (2015). Very deep con-
volutional networks for large-scale image recognition.
International Conference on Learning Representati-
ons.
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I.,
and Salakhutdinov, R. (2014). Dropout: A simple way
to prevent neural networks from overfitting. Journal
of Machine Learning Research, 15:1929–1958.
Tola, E., Lepetit, V., and Fua, P. (2010). Daisy: An effi-
cient dense descriptor applied to wide-baseline stereo.
IEEE Transactions on Pattern Analysis and Machine
Intelligence, 32(5):815–830.
Vapnik, V. N. (1995). The Nature of Statistical Learning
Theory. Springer-Verlag New York, Inc., New York,
NY, USA.
Wang, S., Yang, X., and Tian, Y. (2013). Detecting sig-
nage and doors for blind navigation and wayfinding.
Network Modeling Analysis in Health Informatics and
Bioinformatics, 2(2):81–93.
VISAPP 2019 - 14th International Conference on Computer Vision Theory and Applications
512