首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >An End-to-End Trainable Multi-Column CNN for Scene Recognition in Extremely Changing Environment
【2h】

An End-to-End Trainable Multi-Column CNN for Scene Recognition in Extremely Changing Environment

机译:端到端可训练的多列CNN用于在瞬息万变的环境中进行场景识别

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Scene recognition is an essential part in the vision-based robot navigation domain. The successful application of deep learning technology has triggered more extensive preliminary studies on scene recognition, which all use extracted features from networks that are trained for recognition tasks. In the paper, we interpret scene recognition as a region-based image retrieval problem and present a novel approach for scene recognition with an end-to-end trainable Multi-column convolutional neural network (MCNN) architecture. The proposed MCNN utilizes filters with receptive fields of different sizes to have Multi-level and Multi-layer image perception, and consists of three components: front-end, middle-end and back-end. The first seven layers VGG16 are taken as front-end for two-dimensional feature extraction, Inception-A is taken as the middle-end for deeper learning feature representation, and Large-Margin Softmax Loss (L-Softmax) is taken as the back-end for enhancing intra-class compactness and inter-class-separability. Extensive experiments have been conducted to evaluate the performance according to compare our proposed network to existing state-of-the-art methods. Experimental results on three popular datasets demonstrate the robustness and accuracy of our approach. To the best of our knowledge, the presented approach has not been applied for the scene recognition in literature.
机译:场景识别是基于视觉的机器人导航领域的重要组成部分。深度学习技术的成功应用引发了对场景识别的更广泛的初步研究,这些研究都使用了从训练有素的识别任务的网络中提取的功能。在本文中,我们将场景识别解释为基于区域的图像检索问题,并提出了一种具有端到端可训练多列卷积神经网络(MCNN)架构的场景识别新方法。提出的MCNN利用具有不同大小的接收场的滤波器来实现多级和多层图像感知,并由前端,中间端和后端三个部分组成。前七个层VGG16被用作二维特征提取的前端,Inception-A被用作深度学习特征表示的中间,而大余量的Softmax损失(L-Softmax)被作为后端-end用于增强类内部的紧凑性和类间的可分离性。通过将我们建议的网络与现有的最新方法进行比较,已经进行了广泛的实验以评估性能。在三个流行的数据集上的实验结果证明了我们方法的鲁棒性和准确性。就我们所知,所提出的方法尚未应用于文学中的场景识别。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号