首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition >Predicting the Where and What of actors and actions through Online Action Localization
【24h】

Predicting the Where and What of actors and actions through Online Action Localization

机译:通过在线行动本地化预测行动者和行动的何处和行动

获取原文

摘要

This paper proposes a novel approach to tackle the challenging problem of 'online action localization' which entails predicting actions and their locations as they happen in a video. Typically, action localization or recognition is performed in an offline manner where all the frames in the video are processed together and action labels are not predicted for the future. This disallows timely localization of actions - an important consideration for surveillance tasks. In our approach, given a batch of frames from the immediate past in a video, we estimate pose and over-segment the current frame into superpixels. Next, we discriminatively train an actor foreground model on the superpixels using the pose bounding boxes. A Conditional Random Field with superpixels as nodes, and edges connecting spatio-temporal neighbors is used to obtain action segments. The action confidence is predicted using dynamic programming on SVM scores obtained on short segments of the video, thereby capturing sequential information of the actions. The issue of visual drift is handled by updating the appearance model and pose refinement in an online manner. Lastly, we introduce a new measure to quantify the performance of action prediction (i.e. online action localization), which analyzes how the prediction accuracy varies as a function of observed portion of the video. Our experiments suggest that despite using only a few frames to localize actions at each time instant, we are able to predict the action and obtain competitive results to state-of-the-art offline methods.
机译:本文提出了一种解决“在线行动定位”挑战性问题的新方法,这需要在视频中发生预测行动及其位置。通常,以离线方式执行动作定位或识别,其中视频中的所有帧都被处理在一起,并且未对未来预测动作标签。这不允许及时本地化行动 - 对监督任务的重要考虑因素。在我们的方法中,给出了一批来自视频中的直接过去的帧,我们估计了将当前帧的姿势和过度分段为超像素。接下来,我们使用姿势边界盒差异地在超像素上训练演员前景模型。使用SuperPixels作为节点的条件随机字段,以及连接时空邻居的边缘来获得动作段。使用动态编程在视频的短段上获得的SVM分数上进行动作置信度,从而捕获动作的顺序信息。通过以在线方式更新外观模型和姿势细化来处理视觉漂移问题。最后,我们介绍了一种新的度量来量化动作预测的性能(即在线行动定位),其分析了预测精度如何随着视频部分的函数而变化。我们的实验表明,尽管只使用了几个帧来定位在每次即时的行动,我们都能够预测竞争结果,并将最先进的离线方法获得竞争结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号