首页> 外文期刊>Neurocomputing >Understanding dropout as an optimization trick
【24h】

Understanding dropout as an optimization trick

机译:了解丢失作为优化技巧

获取原文
获取原文并翻译 | 示例
           

摘要

As one of standard approaches to train deep neural networks, dropout has been applied to regularize large models to avoid overfitting, and the improvement in performance by dropout has been explained as avoiding co-adaptation between nodes. However, when correlations between nodes are compared after training the networks with or without dropout, one question arises if co-adaptation avoidance explains the dropout effect completely. In this paper, we propose an additional explanation of why dropout works and propose a new technique to design better activation functions. First, we show that dropout can be explained as an optimization technique to push the input towards the saturation area of nonlinear activation function by accelerating gradient information flowing even in the saturation area in backpropagation. Based on this explanation, we propose a new technique for activation functions, gradient acceleration in activation function (GAAF), that accelerates gradients to flow even in the saturation area. Then, input to the activation function can climb onto the saturation area which makes the network more robust because the model converges on a flat region. Experiment results support our explanation of dropout and confirm that the proposed GAAF technique improves image classification performance with expected properties. (C) 2020 Elsevier B.V. All rights reserved.
机译:作为培训深度神经网络的标准方法之一,已应用于规范大型模型以避免过度装备,并且已经解释了通过辍学的性能的提高为避免节点之间的共同适应。然而,当节点之间的相关性在培训或没有辍学的情况下比较网络之间进行比较时,如果共适应避免完全解释辍学效果,则会出现一个问题。在本文中,我们提出了一个额外的解释,为什么辍学工作和提出一种新技术来设计更好的激活功能。首先,我们示出了通过加速梯度信息,即使在BackProjagation中的饱和区域中,可以将丢失作为优化技术推出朝向非线性激活功能的饱和区域的输入。在此解释的基础上,我们提出了一种用于激活功能的新技术,激活函数(GaAF)中的梯度加速度,即使在饱和区域中也加速梯度流动。然后,输入到激活函数可以升高到饱和区域,这使得网络更稳健,因为模型会聚在平坦区域上。实验结果支持我们对辍学的解释,并确认所提出的GAAF技术通过预期的性能提高了图像分类性能。 (c)2020 Elsevier B.v.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2020年第jul20期|64-70|共7页
  • 作者

    Hahn Sangchul; Choi Heeyoul;

  • 作者单位

    Handong Global Univ Dept Informat & Commun Engn Pohang 37554 South Korea;

    Handong Global Univ Dept Informat & Commun Engn Pohang 37554 South Korea;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Deep learning; Dropout; Activation function;

    机译:深度学习;辍学;激活功能;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号