首页> 外文会议>Pacific Asia Conference on Language, Information and Computation >Attention-based BLSTM-CRF Architecture for Mongolian Named Entity Recognition
【24h】

Attention-based BLSTM-CRF Architecture for Mongolian Named Entity Recognition

机译:基于关注的BLSTM-CRF架构,用于蒙古人名为实体识别

获取原文

摘要

Adding external information from unlabeled data to Named Entity Recognition task is the direction of research for solving the scarcity of labeled data. In this issue, adding pre-trained context embeddings as external information from language model is the state-of-the-art technology. However, it has combined external information to the model with simple concatenation, which assumed the same contribution of both internal and external information. While with the use of pre-trained language models, model parameters were not be updated during the iterative process of named entity recognition model. The language model cannot fully capture the semantic and syntactic roles of the contextual words in the labeled data. In view of that case, this paper proposed an improved approach by using pre-trained context embeddings based on attention mechanism. The attention mechanism layer can dynamically balance the difference between the internal information learned from bidirectional Long Short-term Memory model and external information from neural language model. Experimental results have shown that our model has achieved substantial improvement over previous one in Mongolian Named Entity Recognition task.
机译:将未标记数据添加到命名实体识别任务的外部信息是解决标记数据稀缺的研究方向。在此问题中,将预先训练的上下文嵌入作为来自语言模型的外部信息是最先进的技术。但是,它将外部信息组合到具有简单连接的模型,这假设了内部和外部信息的相同贡献。虽然使用预先训练的语言模型,但在命名实体识别模型的迭代过程中未更新模型参数。语言模型无法完全捕获标记数据中的上下文单词的语义和语法角色。鉴于这种情况,本文通过基于注意机制使用预先训练的上下文嵌入来提出了一种改进的方法。注意机制层可以动态地平衡从双向短期内存模型和来自神经语言模型的外部信息学习的内部信息之间的差异。实验结果表明,我们的模型在蒙古人命名实体识别任务方面取得了大幅改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号