首页> 外文会议>Conference of the European Chapter of the Association for Computational Linguistics >T-NER: An Ail-Round Python Library for Transformer-based Named Entity Recognition
【24h】

T-NER: An Ail-Round Python Library for Transformer-based Named Entity Recognition

机译:T-ner:用于基于变换器的命名实体识别的AIL级Python库

获取原文

摘要

Language model (LM) pretraining has led to consistent improvements in many NLP downstream tasks, including named entity recognition (NER). In this paper, we present T-NER (Transformer-based Named Entity Recognition), a Python library for NER LM finetuning. In addition to its practical utility, T-NER facilitates the study and investigation of the cross-domain and cross-lingual generalization ability of LMs finetuned on NER. Our library also provides a web app where users can get model predictions interactively for arbitrary text, which facilitates qualitative model evaluation for non-expert programmers. We show the potential of the library by compiling nine public NER datasets into a unified format and evaluating the cross-domain and cross-lingual performance across the datasets. The results from our initial experiments show that in-domain performance is generally competitive across datasets. However, cross-domain generalization is challenging even with a large pretrained LM, which has nevertheless capacity to learn domain-specific features if fine-tuned on a combined dataset. To facilitate future research, we also release all our LM checkpoints via the Hugging Face model hub.
机译:语言模型(LM)预先曝光导致许多NLP下游任务中的一致性改进,包括命名实体识别(ner)。在本文中,我们呈现T-NER(基于变换器的命名实体识别),是NER LM FFETUNING的PYTHON库。除了其实用效用之外,T-NER还促进了对NER跨越LMS跨域和交叉语言泛化能力的研究和调查。我们的图书馆还提供了一个Web应用程序,用户可以在任意文本中交互以交互方式获得模型预测,这促进了非专家程序员的定性模型评估。我们通过将九个公共ner数据集编译为统一格式并评估数据集中的跨域和交叉语言性能来显示库的潜力。来自我们初始实验的结果表明,域中的性能通常在数据集中竞争。然而,即使具有大型预磨损的LM,跨域泛化也是具有挑战性的,但是,如果在组合的数据集上进行微调,则迄今为止学习特定于域特征的能力。为了促进未来的研究,我们还通过拥抱面部模型中心释放所有LM检查站。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号