【24h】

An Analysis of Deep Contextual Word Embeddings and Neural Architectures for Toponym Mention Detection in Scientific Publications

机译:用于科学出版物中地名提及检测的深层上下文词嵌入和神经体系结构分析

获取原文

摘要

Toponym detection in scientific papers is an open task and a key first step in place entity enrichment of documents. We examine three common neural architectures in NLP: 1) convolutional neural network, 2) multi-layer perceptron (both applied in a sliding window context) and 3) bi-directional LSTM and apply contextual and non-contextual word embedding layers to these models. We find that deep contextual word embeddings improve the performance of the bi-LSTM with CRF neural architecture achieving the best performance when multiple layers of deep contextual embeddings are concatenated. Our best performing model achieves an average Fl of 0.910 when evaluated on overlap macro exceeding previous state-of-the-art models in the toponym detection task.
机译:科学论文中的地名检测是一项艰巨的任务,也是实现文档实体丰富的关键的第一步。我们研究了NLP中的三种常见神经体系结构:1)卷积神经网络,2)多层感知器(均在滑动窗口上下文中应用)和3)双向LSTM,并将上下文和非上下文词嵌入层应用于这些模型。我们发现,当多层深层上下文嵌入被串联在一起时,深层上下文词嵌入通过CRF神经体系结构提高了bi-LSTM的性能,从而获得了最佳性能。当对重叠宏进行评估时,我们的最佳性能模型获得的平均Fl为0.910,超过了地名检测任务中以前的最新模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号