首页> 外文会议>Simulation Multi-Conference >SYSTOLIC SPARSE MATRIX VECTOR MULTIPLY IN THE AGE OF TPUS AND ACCELERATORS
【24h】

SYSTOLIC SPARSE MATRIX VECTOR MULTIPLY IN THE AGE OF TPUS AND ACCELERATORS

机译:TPU和加速器时代的收缩系稀疏矩阵矢量乘以

获取原文

摘要

Tensor Processing Units has brought back systolic arrays as a computational alternative to high performance computing. Recently Google presented a Tensor Processing Unit for handling matrix multiplication using systolic arrays. This unit is designed for dense matrices only. As they stated, sparse architectural support was omitted momentarily but they will focus on sparsity in future designs. We propose a systolic array to compute the Sparse Matrix Vector product in T2(n) ≈「(nnz)/2」+2n+2 using 2n+2 processing elements. The systolic array we propose also use accumulators to collect the partial results of the resulting vector and supports adapting tiling.
机译:张量处理单元使收缩阵列作为高性能计算的计算替代品。最近谷歌提出了一种使用收缩阵列处理矩阵乘法的张量处理单元。该装置仅设计用于密集的矩阵。正如他们所说,稀疏的建筑支持暂时被省略,但他们将专注于未来设计中的稀疏性。我们建议使用2N + 2处理元件计算T2(n)≈(nnz)/ 2「+ 2n + 2中的稀疏矩阵载体乘积。收缩系统阵列我们提出也使用累加器来收集所得到的矢量的部分结果并支持平铺。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号