首页> 外文会议>International Conference on Control Systems and Computer Science >Experiments with Contextualized Word Embeddings for Keyphrase Extraction
【24h】

Experiments with Contextualized Word Embeddings for Keyphrase Extraction

机译:基于语境化单词嵌入的实验,用于关键肾上腺瓶提取

获取原文

摘要

This paper explores the use of contextualized word embeddings for keyphrase extraction, framed as a sequence labeling task. It compares multiple configurations of neural network architectures that build on top of a transformer encoder. The baseline model is obtained by fine-tuning the BERT/SciBERT transformers. Four architectures are presented and evaluated. Each of them is using a different configuration of LSTM and/or multi-head attention modules. We propose a custom loss function and use it to train all architectures with the exception of the baseline. The paper also documents the differences observed when building on top of either the original word embeddings (obtained from training on a language modelling task) or the fine-tuned embeddings of the baseline model. Finally, we present a case study on the behaviour of the attention mechanisms during inference, highlighting the most interesting findings.
机译:本文探讨了对基调提取的上下文化单词嵌入,框架作为序列标记任务。 它比较了在变压器编码器顶部构建的神经网络架构的多种配置。 基线模型是通过微调伯特/斯敏的变压器来获得的。 呈现和评估了四种架构。 它们中的每一个都使用LSTM和/或多主题模块的不同配置。 我们提出了一种自定义损失函数,并使用它以除基线除外所有架构。 本文还记录了在原始单词嵌入物的顶部构建时观察到的差异(从语言建模任务的培训获得)或基线模型的微调嵌入。 最后,我们在推论期间提出了注意力机制的行为,突出了最有趣的调查结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号