首页> 外文会议>Workshop on Knowledge Extraction and Integration for Deep Learning Architectures >Commonsense Statements Identification and Explanation with Transformer-based Encoders
【24h】

Commonsense Statements Identification and Explanation with Transformer-based Encoders

机译:型号陈述与基于变压器的编码器的识别和解释

获取原文

摘要

In this work, we present our empirical attempt to identify the proper strategy of using Transformer Language Models to identify sentences consistent with commonsense. We tackle the first two tasks from the ComVE (Wang et al., 2020a) competition. The starting point for our work is the BERT assumption according to which a large number of NLP tasks can be solved with pre-trained Transformers with no substantial task-specific changes of the architecture. However, our experiments show that the encoding strategy can have a great impact on the quality of the fine-tuning. The combination between cross-encoding and multi-input models worked better than one cross-encoder and allowed us to achieve comparable results with the state-of-the-art without the use of any external data.
机译:在这项工作中,我们展示了我们的实证尝试来确定使用变压器语言模型的正确策略,以识别与致命统一一致的句子。我们解决了来自Comve的前两个任务(Wang等,2020A)竞争。我们工作的起点是根据哪个大量NLP任务的BERT假设可以通过预先训练的变压器来解决,没有实质的架构的特定任务的变化。然而,我们的实验表明,编码策略可能对微调的质量产生很大影响。交叉编码和多输入模型之间的组合优于一个跨编码器,使我们能够在不使用任何外部数据的情况下实现与现有技术的可比结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号