首页> 外文会议>International Computer Conference, Computer Society of Iran >The Effect of Using Masked Language Models in Random Textual Data Augmentation
【24h】

The Effect of Using Masked Language Models in Random Textual Data Augmentation

机译:在随机文本数据增强中使用屏蔽语言模型的效果

获取原文

摘要

Powerful yet simple augmentation techniques have significantly helped modern deep learning-based text classifiers to become more robust in recent years. Although these augmentation methods have proven to be effective, they often utilize random or non-contextualized operations to generate new data. In this work, we modify a specific augmentation method called Easy Data Augmentation or EDA with more sophisticated text editing operations powered by masked language models such as BERT and RoBERTa to analyze the benefits or setbacks of creating more linguistically meaningful and hopefully higher quality augmentations. Our analysis demonstrates that using a masked language model for word insertion almost always achieves better results than the initial method but it comes at a cost of more time and resources which can be comparatively remedied by deploying a lighter and smaller language model like DistilBERT.
机译:强大而简单的增强技术显着帮助现代基于深度学习的文本分类器近年来变得更加强劲。 虽然这些增强方法已被证明是有效的,但它们通常利用随机或非内容化的操作来生成新数据。 在这项工作中,我们修改了一个特定的增强方法,称为简单的数据增强或EDA,具有更复杂的文本编辑操作,由屏蔽语言模型(如BERT和Roberta),以分析创建更多语言有意义和希望更高质量的增强的好处或挫折。 我们的分析表明,使用用于Word插入的屏蔽语言模型几乎总是始终实现比初始方法更好的结果,但它以更高的时间和资源来实现,这些时间和资源可以通过部署较浅和更小的语言模型来竞争地进行ZeriLerber。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号