首页> 外文会议>International Conference for Emerging Technology >Fine-tuning the BERTSUMEXT model for Clinical Report Summarization
【24h】

Fine-tuning the BERTSUMEXT model for Clinical Report Summarization

机译:调整BERTSUMEXT模型以进行临床报告汇总

获取原文

摘要

Background: Medical personnel are expected to parse through scores of reports each day, covering the medical history of their patients. This reading task is crucial to the effectiveness of the healthcare provided. However, it has been noticed that doctors often have to spend a lot of time going through these documents, in order to get a concise gist of the most medically relevant details. This could even affect the amount of time left for doctor-patient interaction. It is in this scenario, that the potential usefulness of an automatic clinical report summarization tool becomes apparent. Such a system would save a lot of effort for the doctor, and make a lot of time available for quality patient-doctor interaction. The focus of this paper is on extractive summarization.Method: Due to its vast pre-training, BERT (Bidirectional Encoder Representations from Transformers) is one of the most knowledgeable NLP (Natural Language Processing) models currently available- making it one of the best choices for a task like summarization. BERTSUM is the BERT version fine-tuned for summarization, BERTSUMEXT being the extractive summarization variant. The BERTSUMEXT architecture has previously been used to create a model that has been extensively pre-trained on the CNN/DailyMail dataset of news articles and corresponding summaries. It was noticed through testing that this pre-trained version of BERTSUMEXT does not perform very well on clinical reports and therefore needs to be improved to be employed in a clinical report summarization system. The method adopted here is to further train the BERTSUMEXT model using different training strategies on a clinical report summarization dataset and assess the performance improvement. The idea is to expand BERTSUMEXT’s knowledge to give it a ‘medical edge’ that it lacks.Results: The training strategy that modifies the parameter values of the extractive summarization layers of the BERTSUMEXT architecture shows a clear improvement on all nine parameters of the ROUGE (Recall Oriented Understudy for Gisting Evalution) automatic evaluation metric and the human evaluation paradigm. The ROUGE metric evaluates summary quality by measuring the overlap between the reference gold summary and the candidate summary generated by the model. The Human Evaluation Paradigm is a method where we obtain a professional doctor’s opinion on the summary quality produced by the model.
机译:背景:医务人员每天都需要分析数十份报告,涵盖其患者的病史。此阅读任务对于所提供医疗保健的有效性至关重要。然而,已经注意到,为了获得最医学上最相关细节的简明扼要,医生通常不得不花费大量时间阅读这些文件。这甚至可能会影响医患互动的时间。在这种情况下,自动临床报告摘要工具的潜在用途变得显而易见。这样的系统将为医生节省很多精力,并为高质量的医患互动提供了大量时间。方法:由于其大量的预训练,BERT(来自变压器的双向编码器表示)是当前可用的最知识丰富的NLP(自然语言处理)模型之一,使其成为最佳的模型之一。总结等任务的选择。 BERTSUM是为汇总进行了微调的BERT版本,BERTSUMEXT是提取汇总的变体。 BERTSUMEXT体系结构以前已用于创建一个模型,该模型已在新闻文章的CNN / DailyMail数据集和相应摘要上进行了广泛的预训练。通过测试发现,该预训练版本的BERTSUMEXT在临床报告中表现不佳,因此需要改进以在临床报告摘要系统中使用。此处采用的方法是在临床报告摘要数据集上使用不同的训练策略进一步训练BERTSUMEXT模型,并评估性能改进。结果是:修改BERTSUMEXT体系结构的提取摘要层的参数值的训练策略显示出对ROUGE的所有九个参数都有明显的改进(回顾针对指导性评估的基础研究)自动评估指标和人工评估范式。 ROUGE度量标准通过测量参考金摘要与模型生成的候选摘要之间的重叠来评估摘要质量。人工评估范式是一种方法,通过这种方法,我们可以获得专业医生对模型产生的摘要质量的意见。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号