loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Paper Unlock

Authors: Marc Moreaux 1 ; Natalia Lyubova 2 ; Isabelle Ferrané 3 and Frédéric Lerasle 3

Affiliations: 1 Softbank Robotics Europe, and Univ. de Toulouse, France ; 2 Softbank Robotics Europe and, France ; 3 Univ. de Toulouse, France

Keyword(s): Semi-supervised Class Localization, Image Classification, Class Saliency, Global Average Pooling.

Related Ontology Subjects/Areas/Topics: Computer Vision, Visualization and Computer Graphics ; Image and Video Analysis ; Visual Attention and Image Saliency

Abstract: This work addresses the issue of image classification and localization of human actions based on visual data acquired from RGB sensors. Our approach is inspired by the success of deep learning in image classification. In this paper, we describe our method and how the concept of Global Average Pooling (GAP) applies in the context of semi-supervised class localization. We benchmark it with respect to Class Activation Mapping initiated in (Zhou et al., 2016), propose a regularization over the GAP maps to enhance the results, and study whether a combination of these two ideas can result in a better classification accuracy. The models are trained and tested on the Stanford 40 Action dataset (Yao et al., 2011) describing people performing 40 different actions such as drinking, cooking or watching TV. Compared to the aforementioned baseline, our model improves the classification accuracy by 5.3 percent points, achieves a localization accuracy of 50.3%, and drastically diminishes th e computation needed to retrieve the class saliency from the base convolutional model. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 3.145.47.193

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Moreaux, M.; Lyubova, N.; Ferrané, I. and Lerasle, F. (2018). Mind the Regularized GAP, for Human Action Classification and Semi-supervised Localization based on Visual Saliency. In Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018) - Volume 4: VISAPP; ISBN 978-989-758-290-5; ISSN 2184-4321, SciTePress, pages 307-314. DOI: 10.5220/0006548303070314

@conference{visapp18,
author={Marc Moreaux. and Natalia Lyubova. and Isabelle Ferrané. and Frédéric Lerasle.},
title={Mind the Regularized GAP, for Human Action Classification and Semi-supervised Localization based on Visual Saliency},
booktitle={Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018) - Volume 4: VISAPP},
year={2018},
pages={307-314},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0006548303070314},
isbn={978-989-758-290-5},
issn={2184-4321},
}

TY - CONF

JO - Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018) - Volume 4: VISAPP
TI - Mind the Regularized GAP, for Human Action Classification and Semi-supervised Localization based on Visual Saliency
SN - 978-989-758-290-5
IS - 2184-4321
AU - Moreaux, M.
AU - Lyubova, N.
AU - Ferrané, I.
AU - Lerasle, F.
PY - 2018
SP - 307
EP - 314
DO - 10.5220/0006548303070314
PB - SciTePress