首页> 外文会议>IEEE International Solid- State Circuits Conference >14.3 A 65nm Computing-in-Memory-Based CNN Processor with 2.9-to-35.8TOPS/W System Energy Efficiency Using Dynamic-Sparsity Performance-Scaling Architecture and Energy-Efficient Inter/Intra-Macro Data Reuse
【24h】

14.3 A 65nm Computing-in-Memory-Based CNN Processor with 2.9-to-35.8TOPS/W System Energy Efficiency Using Dynamic-Sparsity Performance-Scaling Architecture and Energy-Efficient Inter/Intra-Macro Data Reuse

机译:14.3基于65nm的基于内存计算的CNN处理器,具有动态稀疏性能扩展架构和节能的内部/宏内部数据复用功能,具有2.9至35.8TOPS / W的系统能效

获取原文

摘要

Computing-in-Memory (CIM) is a promising solution for energy-efficient neural network (NN) processors. Previous CIM chips [1], [4] mainly focus on the memory macro itself, lacking insight on the overall system integration. Recently, a CIM-based system processor [5] for speech recognition demonstrated promising energy efficiency. No prior work systematically explores sparsity optimization for a CIM processor. Directly mapping sparse NN models onto regular CIM macros is ineffective, since sparse data is usually randomly distributed and CIM macros cannot be power gated even when many zeros exist. For a high compression rate and high efficiency, the granularity of sparsity [6] needs to be explored based on CIM characteristics. Moreover, system-level weight mapping to a CIM macro and data-reuse strategies are not well explored - these directions are important for CIM macro utilization and energy efficiency.
机译:内存中计算(CIM)是节能神经网络(NN)处理器的有前途的解决方案。以前的CIM芯片[1],[4]主要集中在内存宏本身,而对整体系统集成缺乏了解。最近,用于语音识别的基于CIM的系统处理器[5]展示了有希望的能源效率。没有任何先前的工作系统地探索CIM处理器的稀疏性优化。将稀疏NN模型直接映射到常规CIM宏是无效的,因为稀疏数据通常是随机分布的,即使存在许多零,CIM宏也无法进行功率门控。为了获得高压缩率和高效率,需要基于CIM特性探索稀疏性的粒度[6]。此外,还没有很好地探索到CIM宏的系统级权重映射和数据重用策略-这些方向对于CIM宏的利用和能效非常重要。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号