首页> 外国专利> ACCELERATING SPARSE MATRIX MULTIPLICATION IN STORAGE CLASS MEMORY-BASED CONVOLUTIONAL NEURAL NETWORK INFERENCE

ACCELERATING SPARSE MATRIX MULTIPLICATION IN STORAGE CLASS MEMORY-BASED CONVOLUTIONAL NEURAL NETWORK INFERENCE

机译:基于存储类内存的卷积神经网络推论的加速稀疏矩阵乘法

摘要

Techniques are presented for accelerating in-memory matrix multiplication operations for a convolution neural network (CNN) inference in which the weights of a filter are stored in the memory of a storage class memory device, such as a ReRAM or phase change memory based device. To improve performance for inference operations when filters exhibit sparsity, a zero column index and a zero row index are introduced to account for columns and rows having all zero weight values. These indices can be saved in a register on the memory device and when performing a column/row oriented matrix multiplication, if the zero row/column index indicates that the column/row contains all zero weights, the access of the corresponding bit/word line is skipped as the result will be zero regardless of the input.
机译:呈现用于加速用于卷积神经网络(CNN)推断的存储器矩阵乘法操作的技术,其中滤波器的权重存储在存储类存储器设备的存储器中,例如RERAM或相位改变存储器的设备。为了提高推理操作的性能,当过滤器表现出稀疏性时,引入零列索引和零行索引以解释具有所有零权重值的列和行。这些索引可以保存在存储器设备上的寄存器中,并且执行列/行取向矩阵乘法时,如果零行/列索引指示列/行包含所有零权重,则相应位/字线的访问由于输入,因此结果将是零的。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号