首页> 外文期刊>IEICE Transactions on Information and Systems >One-Pass Semi-Dynamic Network Decoding Using a Subnetwork Caching Model for Large Vocabulary Continuous Speech Recongnition
【24h】

One-Pass Semi-Dynamic Network Decoding Using a Subnetwork Caching Model for Large Vocabulary Continuous Speech Recongnition

机译:大词汇量连续语音识别使用子网络缓存模型的一次通过半动态网络解码

获取原文
获取原文并翻译 | 示例
           

摘要

This paper presents a new decoding framework for large vocabulary continuous speech recognition that can handle a static search network dynamically. Generally, a static network decoder can use a search space that is globally optimized in advance, and therefore it can run at high speed during decoding. However, its large memory requirement due to the large network size or the spatial complexity of the optimization algorithm often makes it impractical. Our new one-pass semi-dynamic network decoding scheme aims at incorporating such an optimized search network with memory efficiency, but without losing speed. In this framework, a complete search network is organized on the basis of self-structuring subnetworks and is nearly minimized using a modified tail-sharing algorithm. While the decoder runs, it caches subnetworks needed for decoding in memory, whereas static network decoders keep the complete network in memory. The subnetwork caching model is controlled by two levels of caches: local cache obtained by subnetwork caching operations and global cache obtained by subnetwork preloading operations. The model can also be controlled adaptively by using subnetwork profiling operations. Furthermore, it is made simple and fast with compactly designed self-structuring subnetworks. Experimental results on a 25 k-word Korean broadcast news transcription task show that the semi-dynamic decoder can run almost as fast as an equivalent static network decoder under various memory configurations by using the subnetwork caching model.
机译:本文提出了一种新的用于大词汇量连续语音识别的解码框架,该框架可以动态处理静态搜索网络。通常,静态网络解码器可以使用预先全局优化的搜索空间,因此,它可以在解码期间高速运行。然而,由于网络规模大或优化算法的空间复杂性,其大的存储需求常常使其不切实际。我们新的单程半动态网络解码方案旨在将这种优化的搜索网络并入存储效率,但又不会降低速度。在这个框架中,一个完整的搜索网络是在自构造子网络的基础上组织的,并且使用改进的尾部共享算法几乎将其最小化。解码器运行时,它将缓存所需的子网缓存在内存中,而静态网络解码器则将完整的网络保留在内存中。子网缓存模型由两个级别的缓存控制:通过子网缓存操作获取的本地缓存和通过子网预加载操作获取的全局缓存。还可以通过使用子网概要分析操作来自适应地控制模型。此外,它通过紧凑设计的自构造子网变得简单快捷。 25 k字韩语广播新闻转录任务的实验结果表明,通过使用子网缓存模型,在各种内存配置下,半动态解码器的运行速度几乎可以与等效静态网络解码器一样快。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号