Figure 12: Summary of the Pratt coefficient obtained using
method based on watershed transform (Osma-Ruiz et al.,
2008).
4 CONCLUSIONS
The present work proposes an alternative to the ex-
istent methods for the glottis segmentation in laryn-
geal images. Despite of the poor illumination in most
images, this methodology provided good results in
the majority of the tested images. Only 11 images
had Pratt coefficients lower than 0.5. The errors in
the segmentation process are attributable to the pre-
processing stage that causes the glottis to lose de-
tails and to be confused with the background. This
in turn complicates the work of the snake that only
segmented part of the glottis which was not affected.
To resolve this inconvenience is necessary to ad-
just the parameters involved in the pre-processing for
each image that presented errors in the segmentation.
Other inconvenience presented in the method pro-
posed, its the hard dependence of the pre-processing.
All of the subsequent stages are closely related with
it. Therefore a wrong setting in the parameters in the
first stage could affect the remaining.
One of the most important achievements reached
is the fact that we do not need to incur in heuristic cri-
teria as the mentioned in the previous work such as:
“the glottis is the darkest object in the image” or “the
glottis is always centered in the image”. We avoid
them, based on the fact that the glottis is always sur-
rounded by grey tones. Taking this account, we can
even out the pixels that belong to the background and
highlight the pixels that belong to the glottis. Lastly,
but not least important is the fact that the snake can be
used for tracking, whereupon the algorithm proposed
could be extended to real time videos.
The solution proposed is very promising even
more if we consider that can be extended to tracking
of the vocal fold in real time; however this algorithm
need to be tested in more different conditions in order
to ensure its generalization capabilities.
ACKNOWLEDGEMENTS
This research work has been financed by the Span-
ish government through the project grant TEC2009-
14123-C04-02.
The authors would also thank the ENT service of the
Gregorio Mara
˜
non Hospital for the acquisition of the
images.
REFERENCES
Abdou, I. E. and Pratt, W. K. (1979). Quantitative design
and evaluation of enhancement/thresholding edge de-
tectors. Proceedings of The IEEE, 67:753–763.
Acton, S. T. and Ray, N. (2009). Biomedical image anal-
ysis: Segmentation. Synthesis Lectures on Image,
Video, and Multimedia Processing, 4(1):1–108.
Guti
´
errez-Arriola, J., Osma-Ruiz, V., Godino-Llorente, J.,
S
´
aenz-Lech
´
on, N., Fraile, R., and no, J. A.-L. (2010).
Preprocesado avanzado de im
´
agenes lar
´
ıngeas para
mejorar la segmentaci
´
on del
´
Area glotal. In 1er Work-
shop de Tecnolog
´
ıas Multibiom
´
etricas para la iden-
tificaci
´
on de Personas, Las Palmas de Gran Canaria,
Espa
˜
na.
Kass, M., Witkin, A., and Terzopoulos, D. (1988). Snakes:
Active contour models. INTERNATIONAL JOURNAL
OF COMPUTER VISION, 1(4):321–331.
Marendic, B., Galatsanos, N., and Bless, D. (2001). New
active contour algorithm for tracking vibrating vocal
folds. In Image Processing, 2001. Proceedings. 2001
International Conference on, volume 1, pages 397 –
400 vol.1.
Mehta, D. D., Deliyski, D. D., Quatieri, T. F., and Hill-
man, R. E. (2011). Automated measurement of vocal
fold vibratory asymmetry from high-speed videoen-
doscopy recordings. Speech, Language and Hearing
Research, 54(1):47 – 54.
Osma-Ruiz, V., Godino-Llorente, J. I., Senz-Lechn, N., and
Fraile, R. (2008). Segmentation of the glottal space
from laryngeal images using the watershed trans-
form. Computerized Medical Imaging and Graphics,
32(3):193 – 201.
Perona, P. and Malik, J. (1990). Scale-space and edge detec-
tion using anisotropic diffusion. IEEE Trans. Pattern
Anal. Mach. Intell., 12(7):629–639.
Russ, J. C. (2002). Image Processing Handbook, Fourth
Edition. CRC Press, Inc., Boca Raton, FL, USA, 4th
edition.
Xu, C. and Prince, J. L. (1998). Snakes, shapes, and gradi-
ent vector flow. IEEE TRANSACTIONS ON IMAGE
PROCESSING, 7(3):359–369.
Yan, Y., Chen, X., and Bless, D. (2006). Automatic trac-
ing of vocal-fold motion from high-speed digital im-
ages. Biomedical Engineering, IEEE Transactions on,
53(7):1394 –1400.
BIOSIGNALS2013-InternationalConferenceonBio-inspiredSystemsandSignalProcessing
322