3.3 Post-processing
From the previous steps, 3 ~ 10 candidates regions
are remained. To decide correct digit regions, we
tried to OCR verification
3.3.1 OCR Verification
Our OCR (Optical character Recognition) system
based on the (Kye Kyung Kim, 2002) that consist of
MLP(multi layer perception) with 198 input neurons,
100 hidden neurons and 10 output neurons. All
candidates regions will be recognized by this MLP,
and selected one or two regions to correct digits.
4 EXPERIMENTAL RESULTS
Our experiment environment consist of Intel
Pentium 2G-Hz, 1G Ram Notebook, Visual C++6.0
under the Windows XP OS. From the system
configuration in figure 2, we captured and tested a
lot of video scenes. We can get the high Exit digit
recognition rate over 90%.
5 CONCLUSIONS
This paper presents an approach for detecting the
Exit number to enhance the safety and mobility of
blind people while walking around subway station.
An image-based technique has been developed to
detect the isolated number pattern at the crossing
roads. The presences of exit numbers are inferred by
careful analysis of numeral width, height, rate,
number of numerals, as well as bandwidth trend. If
we have several candidates of numerals, we adapt to
the OCR function. It was found that the proposed
technique performed with good accuracy. Future
work will focus on new methods for extracting and
all kinds of text characters with higher accuracy and
on the development of a full demonstration system.
ACKNOWLEDGEMENTS
This research was supported by the Conversing
Research Center Program through the National
Research Foundation of Korea(NRF) funded by the
Ministry of Education, Science and Technology
(2009-0082293).
REFERENCES
Andreas Hub, Tim Hartter, Thomas Ertl, “Interactive
Tracking of Movable Objects for the Blind on the
Basis of Environment Models and Perception-Oriented
Object Recognition Methods”, 2006 .
A. Zandifar, R. Duraiswami, A. Chahine, and L. Davis, “A
Video Based Interface to Textual Information for the
Visually Impaired”, IEEE 4th ICMI, 2002, pp.325-330.
B. Thylefors, A. D. Negrel, R. Pararajasegaram, and K. Y.
Dadzie, “Global data on blindness,” Bull. WHO, vol.
73, no. 1, pp. 115–121, Jan. 1995.
“Blindness and visual disability: Seeing ahead projections
into the next century,” WHO Fact Sheet No. 146, 1997.
C. A. Shingledecker and E. Foulke, “A human factor
approach to the assessment of mobility of blind
pedestrians,” Hum. Factor, vol. 20, no. 3, pp. 273–286,
Jun. 1978.
Hub, A., Diepstraten, J., Ertl, T. “Design and
Development of an Indoor Navigation and Object
Identification System for the Blind”. Proceedings of
the ACM SIGACCESS conference on Computers and
accessibility, Atlanta, GA, USA, Designing for
accessibility, 147-152, 2004.
K. Matsuo, K.Ueda and M.Umeda, “Extraction of
Character String from Scene Image by Binarizing
Local Target Area”, T-IEE Japan, Vol. 122-C(2), 2002,
pp.232-241.
Kye Kyung Kim; Yun Koo Chung; Suen, C.Y, “Post-
processing scheme for improving recognition
performance of touching handwritten numeral
strings,” 16th International Conference on Pattern
Recognition, Volume 3, 2002 Page(s):327 – 330
L. Gu, N. Tanaka, T. Kaneko and R.M. Haralick, “The
Extraction of Characters from Cover Images Using
Mathematical Morphology”, IEICE Japan, D-II, Vol.
J80, No.10, 1997, pp. 2696-2704.
“OpenCV 1.0, Open Source Computer Vision Library,”
http://www.intel.com/technology/ computing/opencv/,
2006
R. H. Whitestock, L. Frank, and R. Haneline, “Dog
guides,” in Foundations of Orientation and Mobility,
B. B. Blasch and W. R. Weiner, Eds. New York:
Amer. Foundation for the Blind, 1997.
T. Yamaguchi, Y. Nakano, M. Maruyama, H. Miyao and
T.Hananoi, “Digit Classification on Signboards for
Telephone Number Recognition”, Proc.of the ICDAR,
2003, pp.359-363.
Y. Liu, T. Yamamura, N. Ohnishi and N. Sugie,
“Extraction of Character String Regions from a Scene
Image”, IEICE Japan, D-II, Vol. J81, No.4, 1998,
pp.641-650.
VISAPP 2010 - International Conference on Computer Vision Theory and Applications
546