首页> 外文会议>International Computer Symposium >The Brain Memory Architecture HW/SW Co-Design Platform with Adaptive CNN Algorithm
【24h】

The Brain Memory Architecture HW/SW Co-Design Platform with Adaptive CNN Algorithm

机译:大脑内存架构HW / SW共设计平台,具有自适应CNN算法

获取原文

摘要

As the demand for machine learning, edge computing, and the Internet of Things technology increases, computing efficiency and energy consumption has become an important basis for computing choices. Although the graphics processing unit(GPU) has a high degree of parallel computing capability, its energy consumption is large, and the data transmission is limited by the system bus bandwidth. Therefore, our laboratory previously proposed the Brain Memory Architecture prototype architecture, which integrates FPGA and memory as a computing architecture, which has the advantages of high-efficiency, and low-power computing and does not require data exchange through the system bus. Based on this prototype architecture, this paper constructs the Brain Memory Architecture HW/SW Co-Design Platform (BMCD platform) to provide a good user interface so that users can easily build a hardware and software collaborative design computing environment. Through the library provided by the platform to establish the data transmission and calculation between acceleration hardware and memory to solve the bandwidth limitation of the traditional system bus. In this platform, the AXI4-stream interconnect core is provided as a standard interface for data handshaking with acceleration hardware, which reduces user design complexity and maintains the scalability of connection with other computing IP cores. In platform evaluation, design and adaptive CNN algorithm for hardware and software design platform, provide data quantization methods to reduce data bits to reduce the required data bandwidth and storage space and propose a dynamic adjustment algorithm for integer and decimal ratios to correct the accuracy and design problems that may be caused by data quantization. With this adaptive CNN algorithm architecture and BMCD platform to construct a rapid data transmission. This paper finally analyzes the comparison of the weight transmission time of different CNN models with the CPU and the GPU. The method proposed in this paper can reach about 20 times faster than the CPU and about 10 times faster than the GPU.
机译:至于机器学习,边计算和物联网技术增加互联网,计算效率和能源消耗的需求已成为计算选择的重要依据。尽管图形处理单元(GPU)具有高度的并行计算能力,它的能量消耗大,并且将数据传输是由系统总线带宽的限制。因此,我们的实验室先前提出的大脑内存架构的原型架构,集成了FPGA和内存的计算架构,它具有高效率的优势,低功耗计算,并不需要通过系统总线进行数据交换。本文在此基础上的原型架构,构建了脑内存架构HW / SW协同设计平台(BMCD平台),以提供良好的用户界面,使用户可以轻松地构建一个软硬件协同设计的计算环境。通过由平台提供的库建立加速硬件和存储器之间的数据传输和计算,解决了传统系统总线的带宽限制。在该平台上,AXI4-流互连芯被提供作为用于数据与加速硬件,从而降低了用户设计的复杂性和维护与其他计算IP核连接的可扩展性握手的标准接口。在平台评价,设计和自适应CNN算法的硬件和软件设计平台,提供数据的量化方法,以减少数据的比特以减少所需的数据带宽和存储空间,提出了一种动态调整算法用于整数和十进制比率来校正的准确性和设计可以由数据量化引起的问题。与此自适应算法CNN架构和BMCD平台构建快速数据传输。本文最后分析的不同型号CNN与CPU和GPU的重量传输时间的比较。在本文所提出的方法可达到约比CPU快20倍和大约10倍,比GPU更快。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号