首页> 外文期刊>ETRI journal >Zero‐anaphora resolution in Korean based on deep language representation model: BERT
【24h】

Zero‐anaphora resolution in Korean based on deep language representation model: BERT

机译:基于深语陈述模型的韩国零安息日分辨率:伯特

获取原文
           

摘要

It is necessary to achieve high performance in the task of zero anaphora resolution (ZAR) for completely understanding the texts in Korean, Japanese, Chinese, and various other languages. Deep‐learning‐based models are being employed for building ZAR systems, owing to the success of deep learning in the recent years. However, the objective of building a high‐quality ZAR system is far from being achieved even using these models. To enhance the current ZAR techniques, we fine‐tuned a pre‐trained bidirectional encoder representations from transformers (BERT). Notably, BERT is a general language representation model that enables systems to utilize deep bidirectional contextual information in a natural language text. It extensively exploits the attention mechanism based upon the sequence‐transduction model Transformer. In our model, classification is simultaneously performed for all the words in the input word sequence to decide whether each word can be an antecedent. We seek end‐to‐end learning by disallowing any use of hand‐crafted or dependency‐parsing features. Experimental results show that compared with other models, our approach can significantly improve the performance of ZAR.
机译:有必要在零安差拉分辨率(ZAR)任务中实现高性能,以完全了解韩语,日语,中文和各种其他语言的文本。由于近年来深度学习的成功,基于深度学习的模型正在用于建立扎尔系统。然而,建立高质量的ZAR系统的目的即使使用这些模型也远未实现。为了增强当前的ZAR技术,我们可以精细调整来自变压器(BERT)的预先训练的双向编码器表示。值得注意的是,BERT是一种通用语言表示模型,其使系统能够利用自然语言文本中的深度双向上下文信息。它广泛利用了基于序列转换模型变压器的注意机制。在我们的模型中,同时对输入字序列中的所有单词同时执行分类,以确定每个单词是否可以是先发分子。我们通过禁止任何使用手工制作或依赖性解析功能来寻求端到端的学习。实验结果表明,与其他模型相比,我们的方法可以显着提高ZAR的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号