首页> 外文会议>IEEE Workshop on Spoken Language Technology >Semantic language models for Automatic Speech Recognition
【24h】

Semantic language models for Automatic Speech Recognition

机译:用于自动语音识别的语义语言模型

获取原文

摘要

We are interested in the problem of semantics-aware training of language models (LMs) for Automatic Speech Recognition (ASR). Traditional language modeling research have ignored semantic constraints and focused on limited size histories of words. Semantic structures may provide information to capture lexically realized long-range dependencies as well as the linguistic scene of a speech utterance. In this paper, we present a novel semantic LM(SELM) that is based on the theory of frame semantics. Frame semantics analyzes meaning of words by considering their role in the semantic frames they occur and by considering their syntactic properties. We show that by integrating semantic frames and target words into recurrent neural network LMs we can gain significant improvements in perplexity and word error rates. We have evaluated the semantic LM on the publicly available ASR baselines on the Wall Street Journal (WSJ) corpus. SELMs achieve 50% and 64% relative reduction in perplexity compared to n-gram models by using frames and target words respectively. In addition, 12% and 7% relative improvements in word error rates are achieved by SELMs on the Nov'92 and Nov'93 test sets with respect to the baseline tri-gram LM.
机译:我们对自动语音识别(ASR)的语言模型(LM)的语义感知训练问题感兴趣。传统语言建模研究已经忽略了语义约束,而将注意力集中在有限的单词历史上。语义结构可以提供信息以捕获词汇实现的远程依存关系以及语音表达的语言场景。在本文中,我们提出了一种基于框架语义理论的新型语义LM(SELM)。框架语义通过考虑单词在其出现的语义框架中的作用并考虑其句法属性来分析单词的含义。我们表明,通过将语义框架和目标词集成到递归神经网络LM中,我们可以在困惑和词错误率方面获得显着改善。我们已经在《华尔街日报》(WSJ)语料库上公开使用的ASR基准上评估了语义LM。与n-gram模型相比,分别使用框架和目标词,SE​​LM的困惑度相对降低了50%和64%。此外,相对于基线三元语法LM,通过SELM在Nov'92和Nov'93测试集上实现的单词错误率相对提高了12%和7%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号