首页> 外文会议>INTERSPEECH 2012 >Paraphrastic Language Models
【24h】

Paraphrastic Language Models

机译:释义语言模型

获取原文

摘要

In natural languages multiple word sequences can represent the same underlying meaning. Only modelling the observed surface word sequence can result in poor context coverage, for example, when using n-gram language models (LM). To handle this issue, this paper presents a novel form of language model, the paraphrastic LM. A phrase level transduction model that is statistically learned from standard text data is used to generate paraphrase variants. LM probabilities are then estimated by maximizing their marginal probability. Significant error rate reductions of 0.5%-0.6% absolute were obtained on a state-of-the-art conversational telephone speech recognition task using a paraphrastic multi-level LM modelling both word and phrase sequences.
机译:在自然语言中,多个单词序列可以表示相同的潜在含义。只有建模观察到的表面字序列可以导致较差的上下文覆盖,例如,使用n克语言模型(LM)。为了处理这个问题,本文提出了一种新颖的语言模型,释义LM。从标准文本数据统计学学习的短语级转换模型用于生成释义变体。然后通过最大化其边际概率来估计LM概率。在最先进的会话电话语音识别任务中,使用了一个单词和短语序列,在最先进的会话电话语音识别任务中获得了0.5%-0.6%的显着差错率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号