首页> 外文会议>International Conference on Computational Linguistics >A Closer Look at Linguistic Knowledge in Masked Language Models: The Case of Relative Clauses in American English
【24h】

A Closer Look at Linguistic Knowledge in Masked Language Models: The Case of Relative Clauses in American English

机译:仔细看看蒙面语言模型的语言知识:美国英语中相对条款的情况

获取原文

摘要

Transformer-based language models achieve high performance on various tasks, but we still lack understanding of the kind of linguistic knowledge they learn and rely on. We evaluate three models (BERT, RoBERTa, and ALBERT), testing their grammatical and semantic knowledge by sentence-level probing, diagnostic cases, and masked prediction tasks. We focus on relative clauses (in American English) as a complex phenomenon needing contextual information and antecedent identification to be resolved. Based on a naturalistic dataset, probing shows that all three models indeed capture linguistic knowledge about grammaticality, achieving high performance. Evaluation on diagnostic cases and masked prediction tasks considering fine-grained linguistic knowledge, however, shows pronounced model-specific weaknesses especially on semantic knowledge, strongly impacting models' performance. Our results highlight the importance of (a) model comparison in evaluation task and (b) building up claims of model performance and the linguistic knowledge they capture beyond purely probing-based evaluations.
机译:基于变压器的语言模型在各种任务上实现了高性能,但我们仍然缺乏对他们学习和依赖的语言知识的那种。我们评估三种模型(BERT,Roberta和Albert),通过句子级探测,诊断情况和屏蔽预测任务测试他们的语法和语义知识。我们专注于相对条款(美国英语)作为需要解决上下文信息和先行识别的复杂现象。基于自然主义的数据集,探测表明,所有三种模型确实捕获了语言性的语言知识,实现了高性能。然而,考虑细粒度语言知识的诊断情况和屏蔽预测任务的评估显示明显的模型特异性弱点,特别是在语义知识,强烈影响模型的性能。我们的结果突出了(a)模型比较在评估任务中的重要性和(b)建立了模型性能的索赔以及它们捕获超出纯粹探测的评估的语言知识。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号