首页> 外文会议>International Conference on Computational Science >Conditional BERT Contextual Augmentation
【24h】

Conditional BERT Contextual Augmentation

机译:条件BERT上下文增强

获取原文

摘要

Data augmentation methods are often applied to prevent overfitting and improve generalization of deep neural network models. Recently proposed contextual augmentation augments labeled sentences by randomly replacing words with more varied substitutions predicted by language model. Bidirectional Encoder Representations from Transformers (BERT) demonstrates that a deep bidirectional language model is more powerful than either an unidirectional language model or the shallow concatenation of a forward and backward model. We propose a novel data augmentation method for labeled sentences called conditional BERT contextual augmentation. We retrofit BERT to conditional BERT by introducing a new conditional masked language model (The term "conditional masked language model" appeared once in original BERT paper, which indicates context-conditional, is equivalent to term "masked language model". In our paper, "conditional masked language model" indicates we apply extra label-conditional constraint to the "masked language model".) task. The well trained conditional BERT can be applied to enhance contextual augmentation. Experiments on six various different text classification tasks show that our method can be easily applied to both convolutional or recurrent neural networks classifier to obtain improvement.
机译:数据增强方法通常用于防止过度拟合并改善深度神经网络模型的通用性。最近提出的上下文增强通过将单词随机替换为语言模型预测的更多变化来增强标记的句子。变压器的双向编码器表示(BERT)证明,深度双向语言模型比单向语言模型或向前和向后模型的浅层级联更强大。我们提出了一种新的针对带标签句子的数据增强方法,称为条件BERT上下文增强。我们通过引入新的条件屏蔽语言模型将BERT改造为条件BERT(术语“条件屏蔽语言模型”在原始BERT论文中曾经出现一次,表示上下文条件,等同于术语“屏蔽语言模型”。在本文中, “条件屏蔽语言模型”表示我们将额外的标签条件约束应用于“屏蔽语言模型”。)任务。训练有素的条件BERT可以用于增强上下文增强。对六种不同的文本分类任务进行的实验表明,我们的方法可以轻松地应用于卷积神经网络分类器或递归神经网络分类器,以获得改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号