首页> 美国卫生研究院文献>other >Learning in the Machine: Random Backpropagation and the Deep Learning Channel
【2h】

Learning in the Machine: Random Backpropagation and the Deep Learning Channel

机译:机器学习:随机反向传播和深度学习渠道

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Random backpropagation (RBP) is a variant of the backpropagation algorithm for training neural networks, where the transpose of the forward matrices are replaced by fixed random matrices in the calculation of the weight updates. It is remarkable both because of its effectiveness, in spite of using random matrices to communicate error information, and because it completely removes the taxing requirement of maintaining symmetric weights in a physical neural system. To better understand random backpropagation, we first connect it to the notions of local learning and learning channels. Through this connection, we derive several alternatives to RBP, including skipped RBP (SRPB), adaptive RBP (ARBP), sparse RBP, and their combinations (e.g. ASRBP) and analyze their computational complexity. We then study their behavior through simulations using the MNIST and CIFAR-10 bechnmark datasets. These simulations show that most of these variants work robustly, almost as well as backpropagation, and that multiplication by the derivatives of the activation functions is important. As a follow-up, we study also the low-end of the number of bits required to communicate error information over the learning channel. We then provide partial intuitive explanations for some of the remarkable properties of RBP and its variations. Finally, we prove several mathematical results, including the convergence to fixed points of linear chains of arbitrary length, the convergence to fixed points of linear autoencoders with decorrelated data, the long-term existence of solutions for linear systems with a single hidden layer and convergence in special cases, and the convergence to fixed points of non-linear chains, when the derivative of the activation functions is included.
机译:随机反向传播(RBP)是用于训练神经网络的反向传播算法的一种变体,其中在权重更新的计算中,前向矩阵的转置被固定的随机矩阵代替。尽管它的有效性(尽管使用了随机矩阵来传递错误信息)还是因为它完全消除了在物理神经系统中保持对称权重的繁重工作,但它的显着性令人印象深刻。为了更好地理解随机反向传播,我们首先将其与本地学习和学习渠道的概念联系起来。通过这种联系,我们得出了RBP的几种替代方案,包括跳过RBP(SRPB),自适应RBP(ARBP),稀疏RBP及其组合(例如ASRBP),并分析了它们的计算复杂性。然后,我们使用MNIST和CIFAR-10 bechnmark数据集通过模拟研究它们的行为。这些模拟表明,这些变体中的大多数都可以很好地工作,几乎与反向传播一样好,并且与激活函数的导数相乘很重要。作为后续,我们还研究了通过学习通道传达错误信息所需的低位位数。然后,我们对RBP及其变化的一些非凡特性提供了部分直观的解释。最后,我们证明了一些数学结果,包括收敛到任意长度的线性链的固定点,具有解相关数据的线性自动编码器的固定点的收敛性,具有单个隐藏层的线性系统解的长期存在性以及收敛性在特殊情况下,并且包括激活函数的导数时,收敛到非线性链的固定点。

著录项

  • 期刊名称 other
  • 作者单位
  • 年(卷),期 -1(260),-1
  • 年度 -1
  • 页码 1–35
  • 总页数 63
  • 原文格式 PDF
  • 正文语种
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号