首页> 外文会议>International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications >Mind the Regularized GAP, for Human Action Classification and Semi-supervised Localization based on Visual Saliency
【24h】

Mind the Regularized GAP, for Human Action Classification and Semi-supervised Localization based on Visual Saliency

机译:介绍正规化的差距,用于人类行动分类和基于视觉显着性的半监督本地化

获取原文

摘要

This work addresses the issue of image classification and localization of human actions based on visual data acquired from RGB sensors. Our approach is inspired by the success of deep learning in image classification. In this paper, we describe our method and how the concept of Global Average Pooling (GAP) applies in the context of semi-supervised class localization. We benchmark it with respect to Class Activation Mapping initiated in (Zhou et al., 2016), propose a regularization over the GAP maps to enhance the results, and study whether a combination of these two ideas can result in a better classification accuracy. The models are trained and tested on the Stanford 40 Action dataset (Yao et al., 2011) describing people performing 40 different actions such as drinking, cooking or watching TV. Compared to the aforementioned baseline, our model improves the classification accuracy by 5.3 percent points, achieves a localization accuracy of 50.3%, and drastically diminishes the computation needed to retrieve the class saliency from the base convolutional model.
机译:这项工作解决了基于从RGB传感器获取的视觉数据的图像分类和人类动作本地化问题。我们的方法受到图像分类中深度学习的成功的启发。在本文中,我们描述了我们的方法以及全局平均汇总(GAP)的概念在半监督课程本地化的背景下适用。我们在(周等人,2016年)启动的类激活映射方面基准测试,提出了差距地图的正则化以增强结果,并研究这两个想法的组合是否可以导致更好的分类准确性。这些模型在斯坦福40行动数据集(Yao等,2011)上培训并测试了,描述了执行40种不同行动,如饮酒,烹饪或看电视。与上述基线相比,我们的模型将分类精度提高了5.3%的分数,实现了50.3%的本地化准确性,并大幅度减少了从基础卷积模型中检索班级显着性所需的计算。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号