首页> 外文会议>International Conference on Computer and Knowledge engineering >Efficient implementation of a generalized convolutional neural networks based on weighted euclidean distance
【24h】

Efficient implementation of a generalized convolutional neural networks based on weighted euclidean distance

机译:基于加权欧氏距离的广义卷积神经网络的有效实现

获取原文

摘要

Convolutional Neural Networks (CNNs) are multi-layer deep structures that have been very successful in visual recognition tasks. These networks basically consist of the convolution, pooling, and the nonlinearity layers, each of which operates on the representation produced by the preceding layer and generates a new representation. Convolution layers naturally compute some inner product between a plane represented by the weight parameters and input patches. Recently, Generalized Convolutional Neural Networks (GCNN) have been introduced which justify the use of some kernels or distance functions in place of the inner product operator inside the convolution layers. Although GCNNs gained interesting results on the MNIST dataset, their application to more challenging datasets is hindered by lack of an efficient implementation. In this paper, we focus on a specific generalized convolution operator which is based on the weighted L2 norm distance (WL2Dist). By replacing the nonlinear part with three convolution operators and using effective matrix-matrix multiplications, we were able to efficiently compute the WL2Dist convolution layer both on CPU and GPU. Our experiments show that, on CPU (GPU), the proposed implementation of the WL2Dist layer achieves a 5.5x (21x) speed-up over the initial BLAS-based (CUDA-based) implementations.
机译:卷积神经网络(CNN)是多层深层结构,在视觉识别任务中非常成功。这些网络基本上由卷积,池化和非线性层组成,每个层都对前一层产生的表示进行操作并生成新的表示。卷积层自然计算出由权重参数表示的平面和输入面片之间的一些内积。最近,已经引入了广义卷积神经网络(GCNN),它证明使用某些内核或距离函数代替卷积层内部的内积运算符是合理的。尽管GCNN在MNIST数据集上获得了有趣的结果,但由于缺乏有效的实现方式而使它们无法应用于更具挑战性的数据集。在本文中,我们专注于基于加权L2范数距离(WL2Dist)的特定广义卷积算子。通过使用三个卷积运算符替换非线性部分并使用有效的矩阵矩阵乘法,我们能够在CPU和GPU上高效地计算WL2Dist卷积层。我们的实验表明,在CPU(GPU)上,建议的WL2Dist层实现比初始的基于BLAS(基于CUDA)的实现快5.5倍(21倍)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号