A FRAMEWORK FOR WEBCAM-BASED HAND REHABILITATION EXERCISES

Rui Liu, Burkhard C. Wünsche, Christof Lutteroth, Patrice Delmas

Abstract

Applications for home-based care are rapidly increasing in importance due to spiraling health and elderly care costs. An important aspect of home-based care is exercises for rehabilitation and improving general health. In this paper we present a framework for demonstrating and monitoring hand exercises. The three main components are a 3D hand model, a high-level animation framework which facilitates the task of specifying hand exercises via skeletal animation, and a hand tracking program to monitor and evaluate users’ performance. Our hand tracking solution has no calibration stage and is easily set-up. Segmentation is performed using a perception-based colour space, and hand tracking and motion estimate are obtained using novel variations to a CAMSHIFT and contour analysis algorithms. The results indicate that the robust tracking along with the demonstration and reconstruction of hand exercises provide an effective platform for hand rehabilitation.

References

  1. Boian, R., Sharma, A., Han, C., Merians, A., Burdea, G., Adamovich, S., Recce, M., Tremaine, M., and Poizner, H. (2002). Virtual reality-based post-stroke hand rehabilitation. Studies in Health Technology and Informatics, 85:64-70.
  2. Bradski, D. G. R. and Kaehler, A. (2008). OpenCV, 1st edition. O'Reilly Media, Inc.
  3. Bradski, G. R. (1998). Real time face and object tracking as a component of a perceptual user interface. In WACV 7898: Proceedings of the 4th IEEE Workshop on Applications of Computer Vision (WACV'98), page 214, Washington, DC, USA. IEEE Computer Society.
  4. Chen, Q., Georganas, N., and Petriu, E. (2007). Real-time vision-based hand gesture recognition using haar-like features. In Instrumentation and Measurement Technology Conference Proceedings, 2007. IMTC 2007. IEEE, pages 1-6.
  5. Chong, H. Y., Gortler, S. J., and Zickler, T. (2008). A perception-based color space for illuminationinvariant image processing. ACM Trans. Graph., 27(3):1-7.
  6. Douglas, D. H. and Peucker, T. K. (1973). Algorithm for the reduction of the number of points required to represent a digitized line or its caricature. Cartographica, 10:112-122.
  7. Fischer, H. C., Stubblefield, K., Kline, T., Luo, X., Kenyon, R. V., and Kamper, D. G. (2007). Hand rehabilitation following stroke: A pilot study of assisted finger extension training in a virtual environment. Topics in Stroke Rehabilitation, Volume 14:1-12.
  8. Handexercise.org (2010). Hand exercise: A resource for hand exercising. http://www.handexercise.org/.
  9. Health Information Translations (2010). Active hand exercises. http://www.healthinfotranslations.org/ pdfDocs/Active Hand Exercises.pdf.
  10. Homma, K. and Takenaka, E.-I. (1985). An image processing method for feature extraction of space-occupying lesions. J Nucl Med, 26(12):1472-1477.
  11. Jack, D., Boian, R., Merians, A. S., Tremaine, M., Burdea, G. C., Adamovich, S. V., Recce, M., and Poizner, H. (2001). Virtual reality-enhanced stroke rehabilitation. IEEE transactions on neural systems and rehabilitation engineering, 9(3):308-318.
  12. Kakumanu, P., Makrogiannis, S., and Bourbakis, N. (2007). A survey of skin-color modeling and detection methods. Pattern Recogn., 40(3):1106-1122.
  13. Kass, M., Witkin, A., and Terzopoulos, D. (1988). Snakes: Active Contour Models. International Journal of Computer Vision, 1(4):321-331.
  14. Liu, R. (2010). A framework for webcam-based hand rehabilitation exercises. BSc Honours Dissertation, Graphics Group, Department of Computer Science, University of Auckland, New Zealand.
  15. Mahmoudi, F. and Parviz, M. (2006). Visual hand tracking algorithms. Geometric Modeling and Imaging-New Trends, 0:228-232.
  16. MHTeam (2010). Make human open source tool for making 3d characters. http://www.makehuman.org/.
  17. Stenger, B., Mendona, P. R. S., and Cipolla, R. (2001). Model-based 3d tracking of an articulated hand. Computer Vision and Pattern Recognition, IEEE Computer Society Conference on, 2:310.
  18. Stenger, B., Thayananthan, A., Torr, P. H. S., and Cipolla, R. (2006). Model-based hand tracking using a hierarchical bayesian filter. IEEE Trans. Pattern Anal. Mach. Intell., 28(9):1372-1384.
  19. Vassili, V. V., Sazonov, V., and Andreeva, A. (2003). A survey on pixel-based skin color detection techniques. In Proc. Graphicon, pages 85-92.
  20. Wessel, J. (2004). The effectiveness of hand exercises for persons with rheumatoid arthritis: A systematic review. Journal of Hand Therapy, 17(2):174-180.
Download


Paper Citation


in Harvard Style

Liu R., C. Wünsche B., Lutteroth C. and Delmas P. (2011). A FRAMEWORK FOR WEBCAM-BASED HAND REHABILITATION EXERCISES . In Proceedings of the International Conference on Computer Vision Theory and Applications - Volume 1: VISAPP, (VISIGRAPP 2011) ISBN 978-989-8425-47-8, pages 626-631. DOI: 10.5220/0003365206260631


in Bibtex Style

@conference{visapp11,
author={Rui Liu and Burkhard C. Wünsche and Christof Lutteroth and Patrice Delmas},
title={A FRAMEWORK FOR WEBCAM-BASED HAND REHABILITATION EXERCISES},
booktitle={Proceedings of the International Conference on Computer Vision Theory and Applications - Volume 1: VISAPP, (VISIGRAPP 2011)},
year={2011},
pages={626-631},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0003365206260631},
isbn={978-989-8425-47-8},
}


in EndNote Style

TY - CONF
JO - Proceedings of the International Conference on Computer Vision Theory and Applications - Volume 1: VISAPP, (VISIGRAPP 2011)
TI - A FRAMEWORK FOR WEBCAM-BASED HAND REHABILITATION EXERCISES
SN - 978-989-8425-47-8
AU - Liu R.
AU - C. Wünsche B.
AU - Lutteroth C.
AU - Delmas P.
PY - 2011
SP - 626
EP - 631
DO - 10.5220/0003365206260631