首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Cross-lingual projection for class-based language models
【24h】

Cross-lingual projection for class-based language models

机译:基于类的语言模型的跨语言投影

获取原文

摘要

This paper presents a cross-lingual projection technique for training class-based language models. We borrow from previous success in projecting POS tags and NER mentions to that of a trained class-based language model. We use a CRF to train a model to predict when a sequence of words is a member of a given class and use this to label our language model training data. We show that we can successfully project the contextual cues for these classes across pairs of languages and retain a high quality class model in languages with no supervised class data. We present empirical results that show the quality of the projected models as well as their effect on the down-stream speech recognition objective. We are able to achieve over 70% of the WER reduction when using the projected class models as compared to models trained on human annotations.
机译:本文提出了一种用于训练基于班级的语言模型的跨语言投影技术。我们借鉴了以前在投射POS标签方面的成功经验,而NER提到了经过培训的基于班级的语言模型。我们使用CRF来训练模型,以预测单词序列何时是给定类的成员,并使用它来标记我们的语言模型训练数据。我们证明了我们可以跨语言对成功地为这些类投影上下文提示,并在没有监督类数据的情况下在语言中保留高质量的类模型。我们提供的经验结果表明了投影模型的质量以及它们对下游语音识别目标的影响。与使用人工注释训练的模型相比,使用投影的类模型,我们能够实现WER减少70%以上。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号