首页> 外文学位 >Online Learning of Large Margin Hidden Markov Models for Automatic Speech Recognition.
【24h】

Online Learning of Large Margin Hidden Markov Models for Automatic Speech Recognition.

机译:在线学习大余量隐马尔可夫模型以进行自动语音识别。

获取原文
获取原文并翻译 | 示例

摘要

Over the last two decades, large margin methods have yielded excellent performance on many tasks. The theoretical properties of large margin methods have been intensively studied and are especially well-established for support vector machines (SVMs). However, the scalability of large margin methods remains an issue due to the amount of computation they require. This is especially true for applications involving sequential data.;In this thesis we are motivated by the problem of automatic speech recognition (ASR) whose large-scale applications involve training and testing on extremely large data sets. The acoustic models used in ASR are based on continuous-density hidden Markov models (CD-HMMs). Researchers in ASR have focused on discriminative training of HMMs, which leads to models with significantly lower error rates. More recently, building on the successes of SVMs and various extensions thereof in the machine learning community, a number of researchers in ASR have also explored large margin methods for discriminative training of HMMs.;This dissertation aims to apply various large margin methods developed in the machine learning community to the challenging large-scale problems that arise in ASR. Specifically, we explore the use of sequential, mistake-driven updates for online learning and acoustic feature adaptation in large margin HMMs. The updates are applied to the parameters of acoustic models after the decoding of individual training utterances. For large margin training, the updates attempt to separate the log-likelihoods of correct and incorrect transcriptions by an amount proportional to their Hamming distance. For acoustic feature adaptation, the updates attempt to improve recognition by linearly transforming the features computed by the front end. We evaluate acoustic models trained in this way on the TIMIT speech database. We find that online updates for large margin training not only converge faster than analogous batch optimizations, but also yield lower phone error rates than approaches that do not attempt to enforce a large margin.;We conclude this thesis with a discussion of future research directions, highlighting in particular the challenges of scaling our approach to the most difficult problems in large-vocabulary continuous speech recognition.
机译:在过去的二十年中,大余量方法在许多任务上表现出出色的性能。大幅度裕度方法的理论性质已经得到了深入的研究,对于支持向量机(SVM)尤其是行之有效的。但是,由于大余量方法需要大量计算,因此它们的可伸缩性仍然是一个问题。在涉及顺序数据的应用程序中尤其如此。;本文是由自动语音识别(ASR)问题引起的,该问题的大规模应用涉及对超大型数据集的训练和测试。 ASR中使用的声学模型基于连续密度隐马尔可夫模型(CD-HMM)。 ASR的研究人员专注于HMM的判别训练,这导致模型的错误率大大降低。最近,基于SVM的成功及其在机器学习社区中的各种扩展,ASR的许多研究人员还探索了用于训练HMM的判别式训练的大幅度方法。机器学习社区应对ASR中出现的具有挑战性的大规模问题。具体来说,我们探索在大幅度HMM中使用顺序错误驱动的更新来进行在线学习和声学特征调整。在对单个训练发音进行解码之后,会将更新应用于声学模型的参数。对于大幅度的训练,更新尝试将正确和错误转录的对数似然比与汉明距离成比例。对于声学特征自适应,更新尝试通过线性变换前端计算的特征来提高识别度。我们评估在TIMIT语音数据库中以这种方式训练的声学模型。我们发现,用于大幅度培训的在线更新不仅比类似的批次优化收敛得更快,而且比不试图大幅度提高预算的方法产生的电话错误率更低。我们在总结本文的基础上讨论了未来的研究方向,特别强调了将我们的方法扩展到大词汇量连续语音识别中最困难的问题所面临的挑战。

著录项

  • 作者

    Cheng, Chih-Chieh.;

  • 作者单位

    University of California, San Diego.;

  • 授予单位 University of California, San Diego.;
  • 学科 Artificial Intelligence.;Computer Science.
  • 学位 Ph.D.
  • 年度 2011
  • 页码 136 p.
  • 总页数 136
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号