首页> 外文会议>International Workshop of Physical Agents >Survival Loss: A Neuron Death Regularizer
【24h】

Survival Loss: A Neuron Death Regularizer

机译:生存损失:神经元死亡规范器

获取原文

摘要

We found that combining the L2 regularizer with Adam kills up to 60% of filters in ResNet-110 trained on CIFAR-100 as opposed to combining L2 with Momentum. It does not have a significant impact in terms of accuracy though, where both reach similar values. However, we found that this can be a serious issue if the impaired model is used as a pre-trained model for another more complex dataset (e.g. larger number of categories). This situation actually happens in continual learning. In this paper we conduct a study on the impact of dead filters in continual learning when the dataset increases its difficulty over time and more power from the network is required. Furthermore, we propose a new regularization term referred to as survival loss, that complements L2 to avoid filters to die when combined with Adam. We show that the survival loss improves accuracy in a simulated continual learning set-up, with the prospect of higher improvements as more iterations arrive.
机译:我们发现将L2规范器与ADAM杀死最多60%的Reset-110中的滤波器,而不是将L2与动量组合在一起。 它在准确性方面没有显着影响,其中都达到了类似的值。 但是,如果使用受损的模型用作另一个复杂数据集的预先训练的模型,我们发现这可能是一个严重的问题(例如,更大数量的类别)。 这种情况实际上发生在持续学习中。 在本文中,当数据集增加其难以随着时间的推移和来自网络的更多电力时,我们对持续学习的持续学习的影响进行了研究。 此外,我们提出了一种新的正则化术语作为生存损失,这补充了L2,以避免在与ADAM相结合时死亡。 我们表明生存损失提高了模拟的持续学习设置中的准确性,随着更多迭代到达的推动力提高了更高的改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号