首页> 外文会议>IEEE International Parallel and Distributed Processing Symposium Workshops >RadiX-Net: Structured Sparse Matrices for Deep Neural Networks
【24h】

RadiX-Net: Structured Sparse Matrices for Deep Neural Networks

机译:基拉网:深神经网络的结构稀疏矩阵

获取原文

摘要

The sizes of deep neural networks (DNNs) are rapidly outgrowing the capacity of hardware to store and train them. Research over the past few decades has explored the prospect of sparsifying DNNs before, during, and after training by pruning edges from the underlying topology. The resulting neural network is known as a sparse neural network. More recent work has demonstrated the remarkable result that certain sparse DNNs can train to the same precision as dense DNNs at lower runtime and storage cost. An intriguing class of these sparse DNNs is the X-Nets, which are initialized and trained upon a sparse topology with neither reference to a parent dense DNN nor subsequent pruning. We present an algorithm that deterministically generates RadiX-Nets: sparse DNN topologies that, as a whole, are much more diverse than X-Net topologies, while preserving X-Nets' desired characteristics. We further present a functional-analytic conjecture based on the longstanding observation that sparse neural network topologies can attain the same expressive power as dense counterparts.
机译:深度神经网络(DNN)的大小迅速超越了存储和训练的硬件容量。过去几十年的研究已经探讨了在潜在拓扑中修剪边缘之前,期间和之后幸福DNN的前景。所得到的神经网络被称为稀疏神经网络。最近的工作已经证明了这一结果的显着结果,即某些稀疏的DNN可以以较低的运行时和储存成本为密集的DNN培训到与密集DNN相同的精度。这些稀疏DNN的有趣类是X-网,它在稀疏拓扑上初始化和培训,既不引用父致密DNN也不是随后的修剪。我们提出了一种确定性地生成基拉网的算法:稀疏的DNN拓扑,整体而言比X-Net拓扑更多样化,同时保留X-Net的所需特征。我们进一步提出了一种功能 - 分析猜测,基于长期观察,即稀疏的神经网络拓扑可以以致密的对应物达到相同的表现力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号