首页> 外文会议>International Conference on Electrical and Electronics Engineering >A New Image Captioning Approach for Visually Impaired People
【24h】

A New Image Captioning Approach for Visually Impaired People

机译:一种新的视觉受损人物的形象标题方法

获取原文

摘要

Automatic caption generation in natural language to describe the visual content of an image has attracted an increasing amount of attention in the last decade due to its potential applications. It is a challenging task to generate captions with proper linguistics properties as it requires an advanced level of image understanding that goes far beyond image classification and object detection. In this paper, we propose to use the Stanford CoreNLP model to generate a caption after images are trained using VGG16 deep learning architecture. The visual attributes of images are extracted with the VGG16, which conveys richer content, and then they are fed into the Stanford model for caption generation. Experimental results on the MSCOCO dataset show that the proposed model significantly outperforms the state-of-the-art approaches consistently across different evaluation metrics.
机译:自然语言中的自动标题生成描述图像的视觉内容由于其潜在应用而在过去十年中引起了越来越大的关注。有一个具有挑战性的任务,可以生成具有正确语言学属性的标题,因为它需要高级图像理解,远远超出图像分类和对象检测。在本文中,我们建议使用斯坦福Corenlp模型在使用VGG16深度学习架构进行图像训练后生成标题。使用VGG16提取图像的视觉属性,该VGG16传达更丰富的内容,然后将它们馈入以用于字幕生成的斯坦福模型。 MSCOCO数据集上的实验结果表明,拟议的模型在不同的评估指标上一致地优于最先进的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号