首页> 外文会议>Symposium on VLSI Circuits >SNAP: A 1.67 — 21.55TOPS/W Sparse Neural Acceleration Processor for Unstructured Sparse Deep Neural Network Inference in 16nm CMOS
【24h】

SNAP: A 1.67 — 21.55TOPS/W Sparse Neural Acceleration Processor for Unstructured Sparse Deep Neural Network Inference in 16nm CMOS

机译:SNAP:用于1.6nm CMOS中非结构化稀疏深度神经网络推理的1.67 — 21.55TOPS / W稀疏神经加速处理器

获取原文

摘要

A Sparse Neural Acceleration Processor (SNAP) is designed to exploit unstructured sparsity in deep neural networks (DNNs). SNAP uses parallel associative search to discover input pairs to maintain an average 75% hardware utilization. SNAP's two-level partial sum reduce eliminates access contention and cuts the writeback traffic by 22×. Through diagonal and row configurations of PE arrays, SNAP supports any CONV and FC layers. A 2.4mm2 16nm SNAP test chip is measured to achieve a peak effectual efficiency of 21.55TOPS/W (16b) at 0.55V and 260MHz for CONV layers with 10% weight and activation density. Operating on pruned ResNet-50, SNAP achieves 90.98fps at 0.80V and 480MHz, dissipating 348mW.
机译:稀疏神经加速处理器(SNAP)旨在利用深度神经网络(DNN)中的非结构稀疏性。 SNAP使用并行关联搜索来发现输入对,以维持平均75%的硬件利用率。 SNAP的两级部分和减少减少了访问争用,并将回写流量减少了22倍。通过PE阵列的对角线和行配置,SNAP支持任何CONV和FC层。一个2.4毫米 2 对于具有10%重量和激活密度的CONV层,对16nm SNAP测试芯片进行了测量,以在0.55V和260MHz时达到21.55TOPS / W(16b)的峰值有效效率。在修剪的ResNet-50上运行时,SNAP在0.80V和480MHz时可达到90.98fps,耗散348mW。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号