sian mixtures, there are still some open issues in fea-
ture representation. Our next step is to apply this
work to other benchmark databases with richer mo-
tion variations and more information to be modeled
by a Gaussian mixture where more Gaussian compo-
nents would be necessary. Moreover, an extension
of the action classification method is envisioned in
order to integrate it into a complete scheme consist-
ing of motion detection, background subtraction, and
action recognition in natural and cluttered environ-
ments, which is a difficult and more challenging topic.
REFERENCES
Aggarwal, J. K. and Ryoo, M. S. (2011). Human activ-
ity analysis: A review. ACM Compututing Surveys,
43(3):1–43.
Bishop, C. M. (2006). Pattern Recognition and Machine
Learning. Springer.
Blank, M., Gorelick, L., Shechtman, E., Irani, M., and
Basri, R. (2005). Actions as space-time shapes. In
Proc. 10
th
IEEE International Conference on Com-
puter Vision, pages 1395–1402, Beiging, China.
Chaudhry, R., Ravichandran, A., Hager, G. D., and Vidal,
R. (2009). Histograms of oriented optical flow and
Binet-Cauchy kernels on nonlinear dynamical sys-
tems for the recognition of human actions. In Proc.
IEEE Computer Society Conference on Computer Vi-
sion and Pattern Recognition, pages 1932–1939, Mi-
ami, Florida, USA.
Doll´ar, P., Rabaud, V., Cottrell, G., and Belongie, S. (2005).
Behavior recognition via sparse spatio-temporal fea-
tures. In Proc. 14
th
International Conference on Com-
puter Communications and Networks, pages 65–72,
Beijing, China.
Efros, A. A., Berg, A. C., Mori, G., and Malik, J. (2003).
Recognizing action at a distance. In Proc. 9
th
IEEE
International Conference on Computer Vision, vol-
ume 2, pages 726–733, Nice, France.
Fathi, A. and Mori, G. (2008). Action recognition by learn-
ing mid-level motion features. In Proc. IEEE Com-
puter Society Conference on Computer Vision and
Pattern Recognition, pages 1–8, Anchorage, Alaska,
USA.
Gorelick, L., Blank, M., Shechtman, E., Irani, M., and
Basri, R. (2007). Actions as space-time shapes. IEEE
Transactions on Pattern Analysis and Machine Intel-
ligence, 29(12):2247–2253.
Jhuang, H., Serre, T., Wolf, L., and Poggio, T. (2007). A
biologically inspired system for action recognition. In
Proc. IEEE International Conference on Computer Vi-
sion, pages 1–8, Rio de Janeiro, Brazil.
Lin, Z., Jiang, Z., and Davis, L. S. (2009). Recognizing ac-
tions by shape-motion prototype trees. In Proc. IEEE
International Conference on Computer Vision, pages
444–451, Miami, Florida, USA.
Lucas, B. D. and Kanade, T. (1981). An iterative image
registration technique with an application to stereo vi-
sion. In Proc. 7
th
International Joint Conference on
Artificial Intelligence, pages 674–679, Nice, France.
Morris, B. T. and Trivedi, M. M. (2011). Trajectory learn-
ing for activity understanding: Unsupervised, multi-
level, and long-term adaptive approach. IEEE Trans-
actions on Pattern Analysis and Machine Intelligence,
33(11):2287–2301.
Niebles, J. C., Wang, H., and Fei-Fei, L. (2008). Un-
supervised learning of human action categories us-
ing spatial-temporal words. International Journal of
Computer Vision, 79(3):299–318.
Poppe, R. (2010). A survey on vision-based human action
recognition. Image and Vision Computing, 28(6):976–
990.
Schindler, K. and Gool, L. V. (2008). Action snippets: How
many frames does human action recognition require?
pages 1–8, Anchorage, Alaska, USA.
Schuldt, C., Laptev, I., and Caputo, B. (2004). Recogniz-
ing human actions: A local SVM approach. In Proc.
17
th
International Conference on Pattern Recognition,
pages 32–36, Cambridge, UK.
Seo, H. J. and Milanfar, P. (2011). Action recognition from
one example. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 33(5):867–882.
Theodoridis, S. and Koutroumbas, K. (2008). Pattern
Recognition. Academic Press, 4th edition.
Vlachos, M., Gunopoulos, D., and Kollios, G. (2002). Dis-
covering similar multidimensional trajectories. In
Proc. 18
th
International Conference on Data Engi-
neering, pages 673–682, San Jose, California, USA.
Wang, Y. and Mori, G. (2011). Hidden part models for
human action recognition: Probabilistic versus max
margin. IEEE Transactions on Pattern Analysis and
Machine Intelligence, 33(7):1310–1323.
Yan, X. and Luo, Y. (2012). Recognizing human actions us-
ing a new descriptor based on spatial-temporal interest
points and weighted-output classifier. Neurocomput-
ing, 87:51–61.
Yao, A., Gall, J., and Gool, L. V. (2010). A Hough
transform-based voting framework for action recog-
nition. In Proc. IEEE Computer Society Conference
on Computer Vision and Pattern Recognition, pages
2061–2068, San Francisco, CA, USA.
ActionRecognitionbyMatchingClusteredTrajectoriesofMotionVectors
117