首页> 美国卫生研究院文献>BMC Bioinformatics >Fast and scalable neural embedding models for biomedical sentence classification
【2h】

Fast and scalable neural embedding models for biomedical sentence classification

机译:快速可扩展的神经嵌入模型用于生物医学句子分类

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

BackgroundBiomedical literature is expanding rapidly, and tools that help locate information of interest are needed. To this end, a multitude of different approaches for classifying sentences in biomedical publications according to their coarse semantic and rhetoric categories (e.g., Background, Methods, Results, Conclusions) have been devised, with recent state-of-the-art results reported for a complex deep learning model. Recent evidence showed that shallow and wide neural models such as fastText can provide results that are competitive or superior to complex deep learning models while requiring drastically lower training times and having better scalability. We analyze the efficacy of the fastText model in the classification of biomedical sentences in the PubMed 200k RCT benchmark, and introduce a simple pre-processing step that enables the application of fastText on sentence sequences. Furthermore, we explore the utility of two unsupervised pre-training approaches in scenarios where labeled training data are limited.
机译:背景技术生物医学文献正在迅速发展,需要帮助查找感兴趣信息的工具。为此,已经设计了多种不同的方法来根据生物医学出版物中的句子的粗略语义和修辞类别(例如,背景,方法,结果,结论)对句子进行分类,并报告了最新的最新结果。复杂的深度学习模型。最新证据表明,诸如fastText之类的浅层和宽泛神经模型可以提供与复杂的深度学习模型相比更具竞争性或更好的结果,同时所需的训练时间大大减少,并且具有更好的可伸缩性。在PubMed 200k RCT基准测试中,我们分析了FastText模型在生物医学句子分类中的功效,并介绍了一个简单的预处理步骤,该步骤可将fastText应用到句子序列上。此外,我们探索了在标签训练数据有限的情况下两种无监督训练方法的实用性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号