【24h】

MODELING SPEECH PERCEPTION WITH RESTRICTED BOLTZMANN MACHINES

机译:用受限的Boltzmann机器建模语音感知

获取原文
获取原文并翻译 | 示例

摘要

Restricted Boltzmann Machines (RBMs) appear to be a good candidate to model information processing in the cerebral cortex, since they employ a simple unsupervised learning rule that can be applied to many domains and allows the training of multiple layers of representation. In this paper, we apply the RBM learning algorithm to speech perception. We show that RBMs can be used to achieve good performance in the recognition of isolated spoken digits using a multi-layer deep belief network (consisting of a number of stacked RBMs). This performance, however, appears to depend on the fine-tuning of weights with the supervised back-propagation algorithm. To investigate how central the role of back-propagation, we compare the performance of a number of deep-belief networks using fine-tuning with the performance of the same network architectures without fine-tuning. Furthermore, since one of the main strengths of RBMs is to build up multiple layers of representation, we combine the question of fine-tuning with the question of how beneficial additional layers are for the performance of the networks. To see whether the representations that emerge on higher levels make classification easier, we also apply a simple perception classification to the different levels of the deep-belief networks when it is trained without fine-tuning.
机译:受限的玻尔兹曼机器(RBM)似乎是在大脑皮层中建模信息处理的一个很好的候选者,因为它们采用了一种简单的无监督学习规则,该规则可以应用于多个领域,并且可以训练多层表示。在本文中,我们将RBM学习算法应用于语音感知。我们表明,在使用多层深度信任网络(由多个堆叠的RBM组成)的孤立的语音数字识别中,RBM可以用于实现良好的性能。但是,这种性能似乎取决于使用监督反向传播算法进行权重的微调。为了研究反向传播的作用有多重要,我们将许多使用微调的深信度网络的性能与不进行微调的相同网络体系结构的性能进行了比较。此外,由于RBM的主要优势之一是建立了多层表示,因此我们将微调问题与附加层对网络性能的有益程度结合起来。为了查看出现在较高级别的表示是否使分类更容易,我们还对未经训练的深信度网络的不同级别应用了简单的感知分类。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号