首页> 外文会议>IEEE Sensors Applications Symposium >Learning Action Images Using Deep Convolutional Neural Networks For 3D Action Recognition
【24h】

Learning Action Images Using Deep Convolutional Neural Networks For 3D Action Recognition

机译:使用深卷积神经网络进行3D动作识别的学习动作图像

获取原文

摘要

Recently, 3D action recognition has received more attention of research and industrial communities thanks to the popularity of depth sensors and the efficiency of skeleton estimation algorithms. Accordingly, a large number of methods have been studied by using either handcrafted features with traditional classifiers or recurrent neural networks. However, they cannot learn high-level spatial and temporal features of a whole skeleton sequence exhaustively. In this paper, we proposed a novel encoding technique to transform the pose features of joint-joint distance and joint-joint orientation to color pixels. By concatenating the features of all frames in a sequence, the spatial joint correlations and temporal pose dynamics of action appearance are depicted by a color image. For learning action models, we adopt the strategy of end-to-end fine-tuning a pre-trained deep convolutional neural networks to completely capture multiple high-level features at multi-scale action representation. The proposed method achieves the state-of-the-art performance on NTU RGB+D, the largest and most challenging 3D action recognition dataset, for both the cross-subject and cross-view evaluation protocols.
机译:最近,由于深度传感器的普及和骨架估计算法的效率,3D动作识别获得了研究和工业社区的更多关注。因此,通过使用传统分类器或经常性神经网络的手工特征来研究大量方法。然而,他们无法详尽地学习整个骨架序列的高级空间和时间特征。在本文中,我们提出了一种新颖的编码技术,用于将关节关节距离的姿态特征变换为彩色像素。通过在序列中连接所有帧的特征,通过彩色图像描绘出动作外观的空间关节相关和时间姿势动态。对于学习动作模型,我们采用端到端精细调整的策略进行预先培训的深度卷积神经网络,以在多尺度动作表示下完全捕获多个高级功能。所提出的方法实现了对NTU RGB + D,最大和最具挑战性的3D动作识别数据集的最先进的性能,适用于交叉对象和跨视图评估协议。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号