including success with FP) makes our method suitable
to achieve our goals in this paper. There is, however,
still room for improvement. Specifically, the 13.95%
FP rate might result in extended human review time
over the results to filter the wrong classification. The
most likely explanation for these classification errors
is that the CNN that we used had not been trained
to include all possible kanji variants. We checked that
the kanji most often confused with ”ima” did not have
its own class properly defined. Consequently, extend-
ing our kanji image library will likely result in a re-
duced false positive misclassification rate. The failure
rate of 6.97% is acceptable in the context of our ap-
plication and can be attributed in almost equal parts to
failure in classification and detection. Regarding the
failure due to the misclassificatio of a correctly de-
tected ”ima” kanji, a possible issue is that the ground
truth present in the ETL database (ETL, 2018) cor-
responds to modern writing of the kanji which differ
slightly from some of the ways in which they were
written in the Edo period.
4 CONCLUSIONS AND FUTURE
WORK
We have presented what is, to the best of our knowl-
edge the first work using computer vision tools and
deep learning for the analysis of Wasan documents.
We have presented a tool to select an important struc-
tural element, the ”ima” character that indicates the
position of the start of the textual problem and lies
beside or underneath the graphical description of the
problem. The application presented in this paper rep-
resents, thus, the first step in the construction of a
data base of Wasan documents that can be search-
able in terms of the geometric properties of each prob-
lem. The results presented in this work illustrate how
our algorithms are able to overcome the image quality
problems present in the images (mainly noise in data,
low resolution and orientation tilt) and locate the oc-
currences of the ”ima” character with a high success
rate of 93.02%. Among the issues that remain is the
presence of a 13.95% of false positive detections and
the 6.97% of cases in which the detection failed.
Both issues will be addressed in future work.
In order to reduce False positives, we will widen
our Kanji database so that it includes more than the
present 876 kanji classes. Regarding the cases in what
the algorithm failed, roughly half where due to detec-
tion failure and the other half to classification failure.
The first issue will be addressed by combining the re-
sults of the three blob detectors obtaining best results
in Experiment1 (LoG,DoG,DoH). while the second
can be addressed by tailoring the ground truth for the
”ima” kanji so it is closer to the writing peculiarities
of the EDO period. The experiments included in this
paper also indicate that modifying the code used to
produce the experiments presented to obtain the full
text of the description of each Wasan problems is fea-
sible with minor modifications of the current code.
REFERENCES
Agrawal, M. and Doermann, D. (2013). Clutter noise re-
moval in binary document images. IJDAR, 16(4):351–
369.
Arnia, F., Fardian, Muchallil, S., and Munadi, K. (2015).
Noise characterization in ancient document images
based on DCT coefficient distribution. In ICDAR
2015, pages 971–975.
Barna, N. H., Erana, T. I., Ahmed, S., and Heickal, H.
(2018). Segmentation of heterogeneous documents
into homogeneous components using morphological
operations. In 17th IEEE/ACIS 2018, Singapore, Sin-
gapore, June 6-8, 2018, pages 513–518.
Bradski, G. (2000). The OpenCV Library. Dr. Dobb’s Jour-
nal of Software Tools.
Cires¸an, D. C., Meier, U., Gambardella, L. M., and Schmid-
huber, J. (2010). Deep, big, simple neural nets for
handwritten digit recognition. Neural Computation,
22(12):3207–3220. PMID: 20858131.
ETL (2018). Etl character database. http://etlcdb.db.
aist.go.jp/. Accessed: 2018-11-20.
Fernandes, L. A. and Oliveira, M. M. (2008). Real-time
line detection through an improved hough transform
voting scheme. Pattern Recognition, 41(1):299 – 314.
Goyal, B., Dogra, A., Agrawal, S., and Sohi, B. S. (2018).
Two-dimensional gray scale image denoising via mor-
phological operations in NSST domain & bitonic fil-
tering. Future Generation Comp. Syst., 82:158–175.
Gre¸bowiec, M. and Protasiewicz, J. (2018). A neural frame-
work for online recognition of handwritten kanji char-
acters. In 2018 FedCSIS, pages 479–483.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Im-
agenet classification with deep convolutional neural
networks. In Proceedings of the 25th International
Conference on Neural Information Processing Sys-
tems - Volume 1, NIPS’12, pages 1097–1105, USA.
Curran Associates Inc.
Lindeberg, T. (2015). Image matching using generalized
scale-space interest points. Journal of Mathematical
Imaging and Vision, 52(1):3–36.
Marr, D. and Hildreth, E. (1980). Theory of Edge Detec-
tion. Proceedings of the Royal Society of London Se-
ries B, 207:187–217.
Martzloff, J.-C. (1990). A survey of japanese publica-
tions on the history of japanese traditional mathemat-
ics (wasan) from the last 30 years. Historia Mathe-
matica, 17(4):366 – 373.
ICPRAM 2019 - 8th International Conference on Pattern Recognition Applications and Methods
764