Table 2: Mean accuracy and execution times (s) for Com-
puter 1.
Data
Database 1 Database 2 Database 3
CPU Time 5.79 ± 0.29 26.13 ± 0.61 715.63 ± 4.31
CPU Acc. 56.07 ± 0.65 70.91 ± 0.73 90.98 ± 0.12
GPU Time
1.91 ± 0.05 3.05 ± 0.04 30.56 ± 0.33
GPU Acc. 55.97 ± 0.63 70.91 ± 0.73 90.99 ± 0.12
GPU Gain 3.03x 8.57x 23.42x
Table 3: Mean accuracy and execution times (s) for Com-
puter 2.
Data
Database 1 Database 2 Database 3
CPU Time 3.03 ± 0.03 17.46 ± 0.12 613.30 ± 2.74
CPU Acc. 56.02 ± 0.77 70.92 ± 0.62 91.08 ± 0.10
GPU Time
1.18 ± 0.01 1.75 ± 0.01 13.19 ± 0.04
GPU Acc. 56.00 ± 0.79 70.92 ± 0.62 91.08 ± 0.10
GPU Gain 2.57x 9.98x 46.50x
very interesting results with the proposed approach,
which can be better observed in large datasets.
5 CONCLUSIONS
This work presented a massive parallel approach for
OPF testing phase, A parallel operation called ma-
trix association, which can be seen as a generalization
of a matrix multiplication procedure, has been pro-
posed to consider this kind of data structure in order
to maximize the gain of GPUs. Experimental results
have shown the performance gain of the proposed ap-
proach in 3 out 4 databases, being the worst result in
the smaller database, which highlights the main usage
of GPU-based algorithms in applications that require
a large volume data.
It is worthy noting that GPU 1 configuration can
be found at $130 by the time this article was submit-
ted, meaning that to develop parallel pattern recogni-
tion application is not exclusive only for those with
expensive equipments, even with large datasets.
ACKNOWLEDGEMENTS
The authors are grateful to FAPESP grants
#2009/16206-1, #2010/12697-8 and #2011/08348-
0, and CNPq grants #470571/2013-6 and
#303182/2011-3.
REFERENCES
All`ene, C., Audibert, J. Y., Couprie, M., Cousty, J., and
Keriven, R. (2007). Some links between min-cuts,
optimal spanning forests and watersheds. In Proceed-
ings of the International Symposium on Mathematical
Morphology, pages 253–264. MCT/INPE.
Catanzaro, B., Sundaram, N., and Keutzer, K. (2008). Fast
support vector machine training and classification on
graphics processors. In Proceedings of the 25th inter-
national conference on Machine learning, pages 104–
111, New York, NY, USA. ACM.
Cormen, T. H., Leiserson, C. E., Rivest, R. L., and Stein, C.
(2001). Introduction to Algorithms. The MIT Press, 2
edition.
Duda, R. O., Hart, P. E., and Stork, D. G. (2000). Pattern
Classification (2nd Edition). Wiley-Interscience.
Gainaru, A., Slusanschi, E., and Trausan-Matu, S. (2011).
Mapping data mining algorithms on a GPU architec-
ture: a study. In Proceedings of the 19th interna-
tional conference on Foundations of intelligent sys-
tems, ISMIS’11, pages 102–112, Berlin, Heidelberg.
Springer-Verlag.
Harris, M. (2010). Optimizing parallel reduction in CUDA.
Jang, H., Park, A., and Jung, K. (2008). Neural network im-
plementation using CUDA and OpenMP. In DICTA
’08: Proceedings of the 2008 Digital Image Com-
puting: Techniques and Applications, pages 155–161,
Washington, DC, USA. IEEE Computer Society.
Kaewpijit, S., Moigne, J., and El-Ghazawi, T. (2003).
Automatic reduction of hyperspectral imagery using
wavelet spectral analysis. IEEE Transactions on Geo-
science and Remote Sensing, 41(4):863–871.
Landgrebe, D. (2005). Signal Theory Methods in Multi-
spectral Remote Sensing. Wiley, Newark, NJ.
Oh, K. and Jung, K. (2004). GPU implementation of neural
networks. Pattern Recognition, 37(6):1311–1314.
Papa, J. P., Falc˜ao, A. X., Albuquerque, V. H. C., and
Tavares, J. M. R. S. (2012). Efficient supervised
optimum-path forest classification for large datasets.
Pattern Recognition, 45(1):512–520.
Papa, J. P., Falc˜ao, A. X., and Suzuki, C. T. N. (2009). Su-
pervised pattern classification based on optimum-path
forest. International Journal of Imaging Systems and
Technology, 19(2):120–131.
FastOptimum-PathForestClassificationonGraphicsProcessors
631