首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network
【2h】

Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network

机译:使用直接非循环图来增强基于骨架的动作识别用线性映射卷积神经网络

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Research on the human activity recognition could be utilized for the monitoring of elderly people living alone to reduce the cost of home care. Video sensors can be easily deployed in the different zones of houses to achieve monitoring. The goal of this study is to employ a linear-map convolutional neural network (CNN) to perform action recognition with RGB videos. To reduce the amount of the training data, the posture information is represented by skeleton data extracted from the 300 frames of one film. The two-stream method was applied to increase the accuracy of recognition by using the spatial and motion features of skeleton sequences. The relations of adjacent skeletal joints were employed to build the direct acyclic graph (DAG) matrices, source matrix, and target matrix. Two features were transferred by DAG matrices and expanded as color texture images. The linear-map CNN had a two-dimensional linear map at the beginning of each layer to adjust the number of channels. A two-dimensional CNN was used to recognize the actions. We applied the RGB videos from the action recognition datasets of the NTU RGB+D database, which was established by the Rapid-Rich Object Search Lab, to execute model training and performance evaluation. The experimental results show that the obtained precision, recall, specificity, F1-score, and accuracy were 86.9%, 86.1%, 99.9%, 86.3%, and 99.5%, respectively, in the cross-subject source, and 94.8%, 94.7%, 99.9%, 94.7%, and 99.9%, respectively, in the cross-view source. An important contribution of this work is that by using the skeleton sequences to produce the spatial and motion features and the DAG matrix to enhance the relation of adjacent skeletal joints, the computation speed was faster than the traditional schemes that utilize single frame image convolution. Therefore, this work exhibits the practical potential of real-life action recognition.
机译:人类活动识别的研究可用于监测独立的老年人,以降低家庭护理的成本。视频传感器可以轻松部署在房屋的不同区域以实现监控。本研究的目标是使用线性地图卷积神经网络(CNN)来执行与RGB视频的动作识别。为了减少训练数据的量,姿势信息由从一部电影的300帧提取的骨架数据表示。应用了两流方法以通过使用骨架序列的空间和运动特征来提高识别的准确性。采用相邻骨架关节的关系来构建直接非循环图(DAG)矩阵,源矩阵和目标矩阵。两个特征由DAG矩阵转移并作为颜色纹理图像扩展。线性映射CNN在每层开始时具有二维线性图,以调整通道的数量。使用二维CNN来识别动作。我们从NTU RGB + D数据库的动作识别数据集应用RGB视频,该数据集由快速丰富的对象搜索实验室建立,以执行模型培训和性能评估。实验结果表明,所获得的精确度,召回,特异性,F1-得分,和准确性分别为86.9%,86.1%,99.9%,86.3%,和99.5%,横受试者源中,以及94.8%,94.7分别在横视源中分别为99.9%,94.7%和99.9%。这项工作的重要贡献是,通过使用骨架序列来产生空间和运动特征和DAG矩阵来增强相邻骨骼关节的关系,计算速度比利用单帧图像卷积的传统方案更快。因此,这项工作展现了现实生活行动识别的实际潜力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号