首页> 外文期刊>BMC Medical Informatics and Decision Making >Distributed representation and one-hot representation fusion with gated network for clinical semantic textual similarity
【24h】

Distributed representation and one-hot representation fusion with gated network for clinical semantic textual similarity

机译:具有门控网络的分布式表示和单热表示融合,用于临床语义文本相似性

获取原文
           

摘要

Semantic textual similarity (STS) is a fundamental natural language processing (NLP) task which can be widely used in many NLP applications such as Question Answer (QA), Information Retrieval (IR), etc. It is a typical regression problem, and almost all STS systems either use distributed representation or one-hot representation to model sentence pairs. In this paper, we proposed a novel framework based on a gated network to fuse distributed representation and one-hot representation of sentence pairs. Some current state-of-the-art distributed representation methods, including Convolutional Neural Network (CNN), Bi-directional Long Short Term Memory networks (Bi-LSTM) and Bidirectional Encoder Representations from Transformers (BERT), were used in our framework, and a system based on this framework was developed for a shared task regarding clinical STS organized by BioCreative/OHNLP in 2018. Compared with the systems only using distributed representation or one-hot representation, our method achieved much higher Pearson correlation. Among all distributed representations, BERT performed best. The highest Person correlation of our system was 0.8541, higher than the best official one of the BioCreative/OHNLP clinical STS shared task in 2018 (0.8328) by 0.0213. Distributed representation and one-hot representation are complementary to each other and can be fused by gated network.
机译:语义文本相似性(STS)是一种基本的自然语言处理(NLP)任务,可以广泛用于许多NLP应用程序,例如问题答案(QA),信息检索(IR)等。它是一个典型的回归问题,但几乎所有STS系统都使用分布式表示或单热表示来模型句子对。在本文中,我们提出了一种基于门控网络的新颖框架,以保险丝分布式表示和句子对的单热表示。在我们的框架中使用了一些当前最先进的分布式表示方法,包括卷积神经网络(CNN),双向长短短期存储网络(Bi-LSTM)和来自变压器(BERT)的双向编码器表示,而基于此框架的系统是为2018年通过生物成像/ OHNLP组织的临床STS的共享任务开发的系统。与仅使用分布式表示或单热表示的系统相比,我们的方法实现了更高的Pearson相关性。在所有分布式表示中,BERT都表现最佳。我们的系统的最高关联是0.8541,高于2018年共享任务的最佳官方/ OHNLP临床STS共享任务(0.8328)0.0213。分布式表示和单热表示彼此互补,可以通过门控网络融合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号