首页> 外文期刊>IEEE Solid-State Circuits Letters >A 8.93-TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity With All Parameters Stored On-Chip
【24h】

A 8.93-TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity With All Parameters Stored On-Chip

机译:8.93 - 顶/ W LSTM经常性神经网络加速器,具有分层粗粒稀疏性,具有存储片上的所有参数

获取原文
获取原文并翻译 | 示例
           

摘要

Long short-term memory (LSTM) networks are widely used for speech applications but pose difficulties for efficient implementation on hardware due to large weight storage requirements. We present an energy-efficient LSTM recurrent neural network (RNN) accelerator, featuring an algorithm-hardware co-optimized memory compression technique called hierarchical coarse-grain sparsity (HCGS). Aided by HCGS-based block-wise recursive weight compression, we demonstrate LSTM networks with up to 16× fewer weights while achieving minimal accuracy loss. The prototype chip fabricated in 65-nm LP CMOS achieves 8.93/7.22 TOPS/W for 2-/3-layer LSTM RNNs trained with HCGS for TIMIT/TED-LIUM corpora.
机译:长期内存(LSTM)网络广泛用于语音应用,但由于大量存储要求,在硬件上有效实现困难。我们提出了一个节能的LSTM经常性神经网络(RNN)加速器,具有算法 - 硬件共同优化内存压缩技术,称为分层粗粒稀稀物(HCG)。通过基于HCGS的块递归重量压缩,我们向LSTM网络展示了高达16倍的重量,同时实现了最小的精度损耗。在65-NM LP CMOS中制造的原型芯片达到8.93 / 7.22顶/倍,用于2- / 3层LSTM RNNS,用HCG为Timit / Ted-lim Coresta培训。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号