Authors:
Hemerson Tacon
1
;
André de Souza Brito
1
;
Hugo de Lima Chaves
1
;
Marcelo Bernardes Vieira
1
;
Saulo Moraes Villela
1
;
Helena de Almeida Maia
2
;
Darwin Ttito Concha
2
and
Helio Pedrini
2
Affiliations:
1
Department of Computer Science, Federal University of Juiz de Fora (UFJF), Juiz de Fora, MG, Brazil
;
2
Institute of Computing, University of Campinas (UNICAMP), Campinas, SP, Brazil
Keyword(s):
Deep Learning, Human Action Recognition, Data Augmentation, Visual Rhythm, Video Analysis.
Abstract:
Despite the significant progress of Deep Learning models on the image classification task, it still needs enhancements for the Human Action Recognition task. In this work, we propose to extract horizontal and vertical Visual Rhythms as well as their data augmentations as video features. The data augmentation is driven by crops extracted from the symmetric extension of the time dimension, preserving the video frame rate, which is essential to keep motion patterns. The crops provide a 2D representation of the video volume matching the fixed input size of a 2D Convolutional Neural Network. In addition, multiple crops with stride guarantee coverage of the entire video. We verified that the combination of horizontal and vertical directions leads do better results than previous methods. A multi-stream strategy combining RGB and Optical Flow information is modified to include the additional spatiotemporal streams: one for the horizontal Symmetrically Extended Visual Rhythm (SEVR), and anoth
er for the vertical one. Results show that our method achieves accuracy rates close to the state of the art on the challenging UCF101 and HMDB51 datasets. Furthermore, we assessed the impact of data augmentations methods for Human Action Recognition and verified an increase of 10% for the UCF101 dataset.
(More)