首页> 外文会议>2018 55th ACM/ESDA/IEEE Design Automation Conference >Long Live TIME: Improving Lifetime for Training-In-Memory Engines by Structured Gradient Sparsification
【24h】

Long Live TIME: Improving Lifetime for Training-In-Memory Engines by Structured Gradient Sparsification

机译:使用寿命长:通过结构化的梯度稀疏化来提高内存训练引擎的使用寿命

获取原文
获取原文并翻译 | 示例

摘要

Deeper and larger Neural Networks (NNs) have made breakthroughs in many fields. While conventional CMOS-based computing platforms are hard to achieve higher energy efficiency. RRAM-based systems provide a promising solution to build efficient Training-In-Memory Engines (TIME). While the endurance of RRAM cells is limited, it's a severe issue as the weights of NN always need to be updated for thousands to millions of times during training. Gradient sparsification can address this problem by dropping off most of the smaller gradients but introduce unacceptable computation cost. We proposed an effective framework, SGS-ARS, including Structured Gradient Sparsification (SGS) and Aging-aware Row Swapping (ARS) scheme, to guarantee write balance across whole RRAM crossbars and prolong the lifetime of TIME. Our experiments demonstrate that 356× lifetime extension is achieved when TIME is programmed to train ResNet-50 on Imagenet dataset with our SGS-ARS framework.
机译:越来越大的神经网络(NN)在许多领域都取得了突破。传统的基于CMOS的计算平台很难实现更高的能源效率。基于RRAM的系统为构建有效的内存中训练引擎(TIME)提供了有希望的解决方案。尽管RRAM单元的耐用性受到限制,但是这是一个严重的问题,因为在训练过程中,NN的权重始终需要更新数千到数百万次。梯度稀疏化可以通过降低大多数较小的梯度来解决此问题,但会带来不可接受的计算成本。我们提出了一个有效的框架SGS-ARS,包括结构化梯度稀疏化(SGS)和可感知老化的行交换(ARS)方案,以确保整个RRAM交叉开关之间的写平衡并延长TIME的寿命。我们的实验表明,使用我们的SGS-ARS框架对TIME进行编程以在Imagenet数据集上训练ResNet-50时,可以实现356倍的寿命扩展。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号