classification (Pre = 79.8% and Rec = 86.7%). Fifth,
the inclusion of the M class mostly deteriorates the
performance of the H classification (e.g. for CNN:
13.3% confusion). Moreover, the M class is mostly
confused as H (33.9% for CNN), something that is
observed for all methods. Hence, the M class was the
most difficult one to recognize. A quantitative
measure of independent reviewers’ agreement on the
annotation of the M class would help evaluating the
difficulty of this task, or even whether the 3-class
classification scheme is indeed appropriate for our
application. In the future we aim to elaborate further
on this issue.
Given that this study investigates a novel
application in the area of computer assisted surgery,
there are still open issues for further research. First,
the results in this study are based on the ground truth
assessment
provided by a single expert. The
recruitment of additional experts is essential in order
to evaluate their level of agreement and most
importantly to establish the most appropriate
vascularity annotation scheme. Moreover, we aim to
expand our dataset by including more images from
additional LC operations. Second, the results are
based on classification of patches extracted from GB
images. Hence, it is important to extend the CNN
model to predict the vascular pattern of the entire GB
region in the laparoscopic image. A potential solution
would be to sequentially extract patches from a user-
specified GB region and then aggregate the CNN’s
patch predictions. The investigation of more
advanced CNN models, alternative loss functions to
penalize misclassifications of extreme classes, and
color preprocessing techniques for visual
enhancement of the GB wall vessels, are also major
topics of interest for future research work.
ACKNOWLEDGEMENTS
The author thanks Special Account for Research
Grants and National and Kapodistrian University of
Athens for funding to attend the meeting.
REFERENCES
Bishop, C.M., 2006. Pattern recognition and machine
learning. New York: Springer-Verlag New York, Inc.
Dalal, N. and Triggs, B., 2005. Histograms of Oriented
Gradients for Human Detection, In IEEE Computer
Society Conference on Computer Vision and Pattern
Recognition (CVPR’05), pp. 886–893.
Huang J. et al., 1997. Image indexing using color
correlograms, In Proceedings of IEEE Computer
Society Conference on Computer Vision and Pattern
Recognition. IEEE Comput. Soc, pp. 762–768.
Ioffe, S. and Szegedy, C., 2015. Batch Normalization:
accelerating deep network training by reducing internal
covariate shift, In Proceedings of the 32nd
International Conference on International Conference
on Machine Learning (ICML ’15), pp. 448–456.
Jin, A. et al., 2018. Tool detection and operative skill
assessment in surgical videos using region-based
convolutional neural networks, In IEEE Winter
Conference on Applications of Computer Vision
(WACV). Lake Tahoe, NV, USA, pp. 691–699.
Jin, Y. et al., 2019. Multi-task recurrent convolutional
network with correlation loss for surgical video
analysis. arXiv preprint. Available at:
http://arxiv.org/abs/1907.06099.
Kingma, D.P. and Ba, J, 2014. Adam: a method for
stochastic optimization. arXiv preprint. Available at:
http://arxiv.org/abs/1412.6980.
Loukas, C. et al., 2016. Shot boundary detection in
endoscopic surgery videos using a variational Bayesian
framework, International Journal of Computer Assisted
Radiology and Surgery, 11(11), pp. 1937–1949.
Loukas, C. et al., 2018. Keyframe extraction from
laparoscopic videos based on visual saliency detection,
Computer Methods and Programs in Biomedicine, 165,
pp. 13–23.
Loukas, C., 2018. Video content analysis of surgical
procedures, Surgical Endoscopy, 32(2), pp. 553–568.
Loukas, C. and Georgiou, E., 2013. Surgical workflow
analysis with Gaussian mixture multivariate
autoregressive (GMMAR) models: a simulation study,
Computer Aided Surgery, 18(3–4), pp. 47–62.
Loukas, C. and Georgiou, E., 2015. Smoke detection in
endoscopic surgery videos: a first step towards retrieval
of semantic events, International Journal of Medical
Robotics and Computer Assisted Surgery, 11(1), pp.
80–94.
Loukas, C. et al., 2011. The contribution of simulation
training in enhancing key components of laparoscopic
competence, The American Surgeon, 77(6), pp. 708-
715.
Lux, M. and Marques, O., 2013. Visual information
retrieval using Java and LIRE, Synthesis Lectures on
Information Concepts, Retrieval, and Services. Edited
by G. Marchionini. Morgan & Claypool.
Pass, G. et al., 1996. Comparing images using color
coherence vectors, In
Proceedings of the fourth ACM
international conference on Multimedia-
MULTIMEDIA ’96. New York, New York, USA: ACM
Press, pp. 65–73.
Petscharnig, S and Schöffmann, K., 2018. Binary
convolutional neural network features off-the-shelf for
image to video linking in endoscopic multimedia
databases, Multimedia Tools and Applications, 77(21),
pp. 28817–28842.