首页> 外文会议>Computer Vision (ICCV), 2011 IEEE International Conference on >Manhattan scene understanding using monocular, stereo, and 3D features
【24h】

Manhattan scene understanding using monocular, stereo, and 3D features

机译:使用单眼,立体和3D功能了解曼哈顿场景

获取原文
获取原文并翻译 | 示例

摘要

This paper addresses scene understanding in the context of a moving camera, integrating semantic reasoning ideas from monocular vision with 3D information available through structure-from-motion. We combine geometric and photometric cues in a Bayesian framework, building on recent successes leveraging the indoor Manhattan assumption in monocular vision. We focus on indoor environments and show how to extract key boundaries while ignoring clutter and decorations. To achieve this we present a graphical model that relates photometric cues learned from labeled data, stereo photo-consistency across multiple views, and depth cues derived from structure-from-motion point clouds. We show how to solve MAP inference using dynamic programming, allowing exact, global inference in ∼100 ms (in addition to feature computation of under one second) without using specialized hardware. Experiments show our system out-performing the state-of-the-art.
机译:本文探讨了在移动相机环境中的场景理解,将单眼视觉的语义推理思想与可通过运动构造获得的3D信息相结合。我们结合贝叶斯框架中的几何和光度学线索,以单眼视觉中曼哈顿室内假设为基础的最新成功。我们专注于室内环境,并展示如何在忽略杂乱和装饰的同时提取关键边界。为了实现这一点,我们提出了一个图形模型,该模型将从标记数据中获悉的光度学线索,跨多个视图的立体光一致性以及从运动点云结构得出的深度线索联系起来。我们展示了如何使用动态编程解决MAP推理,如何在不使用专门硬件的情况下,在约100 ms内(除了不到一秒钟的特征计算)实现精确的全局推理。实验表明,我们的系统性能优于最新技术。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号