首页> 外国专利> Neural network processor using compression and decompression of activation data to reduce memory bandwidth utilization

Neural network processor using compression and decompression of activation data to reduce memory bandwidth utilization

机译:神经网络处理器使用激活数据的压缩和解压缩来减少内存带宽利用率

摘要

A deep neural network ("DNN") module can compress and decompress neuron-generated activation data to reduce the utilization of memory bus bandwidth. The compression unit can receive an uncompressed chunk of data generated by a neuron in the DNN module. The compression unit generates a mask portion and a data portion of a compressed output chunk. The mask portion encodes the presence and location of the zero and non-zero bytes in the uncompressed chunk of data. The data portion stores truncated non-zero bytes from the uncompressed chunk of data. A decompression unit can receive a compressed chunk of data from memory in the DNN processor or memory of an application host. The decompression unit decompresses the compressed chunk of data using the mask portion and the data portion. This can reduce memory bus utilization, allow a DNN module to complete processing operations more quickly, and reduce power consumption.
机译:深度神经网络(“ DNN”)模块可以压缩和解压缩神经元生成的激活数据,以减少内存总线带宽的利用率。压缩单元可以接收由DNN模块中的神经元生成的未压缩数据块。压缩单元生成压缩的输出块的掩码部分和数据部分。掩码部分对未压缩数据块中零字节和非零字节的存在和位置进行编码。数据部分存储来自未压缩数据块的截断的非零字节。解压缩单元可以从DNN处理器中的内存或应用程序主机的内存中接收压缩的数据块。解压缩单元使用掩码部分和数据部分对压缩的数据块进行解压缩。这样可以降低内存总线利用率,允许DNN模块更快地完成处理操作,并降低功耗。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号