...
首页> 外文期刊>Image and Vision Computing >Ordered trajectories for human action recognition with large number of classes
【24h】

Ordered trajectories for human action recognition with large number of classes

机译:用于人类动作识别的有序轨迹,种类繁多

获取原文
获取原文并翻译 | 示例
           

摘要

Recently, a video representation based on dense trajectories has been shown to outperform other human action recognition methods on several benchmark datasets. The trajectories capture the motion characteristics of different moving objects in space and temporal dimensions. In dense trajectories, points are sampled at uniform intervals in space and time and then tracked using a dense optical flow field over a fixed length of L frames (optimally 15) spread overlapping over the entire video. However, among these base (dense) trajectories, a few may continue for longer than duration L, capturing motion characteristics of objects that may be more valuable than the information from the base trajectories. Thus, we propose a technique that searches for trajectories with a longer duration and refer to these as 'ordered trajectories'. Experimental results show that ordered trajectories perform much better than the base trajectories, both standalone and when combined. Moreover, the uniform sampling of dense trajectories does not discriminate objects of interest from the background or other objects. Consequently, a lot of information is accumulated, which actually may not be useful. This can especially escalate when there is more data due to an increase in the number of action classes. We observe that our proposed trajectories remove some background clutter, too. We use a Bag-of-Words framework to conduct experiments on the benchmark HMDB51, UCF50 and UCF101 datasets containing the largest number of action classes to date. Further, we also evaluate three state-of-the art feature encoding techniques to study their performance on a common platform. (C) 2015 Elsevier B.V. All rights reserved.
机译:最近,在多个基准数据集上,基于密集轨迹的视频表示已优于其他人类动作识别方法。轨迹捕获了空间和时间维度上不同移动物体的运动特征。在密集的轨迹中,在空间和时间上以均匀的间隔对点进行采样,然后使用固定长度的L帧(最佳为15个)在整个视频中重叠分布的密集光流场跟踪点。但是,在这些基本(密集)轨迹中,一些轨迹的持续时间可能比持续时间L长,从而捕获了比来自基本轨迹的信息更有价值的对象的运动特征。因此,我们提出了一种搜索持续时间较长的轨迹并将其称为“有序轨迹”的技术。实验结果表明,无论是独立的还是组合的,有序轨迹的性能都比基本轨迹好得多。此外,密集轨迹的均匀采样不会将感兴趣的对象与背景或其他对象区分开。因此,积累了很多信息,实际上可能没有用。当由于动作类的数量增加而有更多数据时,这尤其可能升级。我们观察到,我们提出的轨迹也消除了一些背景杂波。我们使用“词袋”框架对基准HMDB51,UCF50和UCF101数据集进行实验,该数据集包含迄今为止最多的动作类。此外,我们还评估了三种最新的特征编码技术,以研究它们在通用平台上的性能。 (C)2015 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号