...
首页> 外文期刊>IEEE Journal of Solid-State Circuits >An 8.93 TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity for On-Device Speech Recognition
【24h】

An 8.93 TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity for On-Device Speech Recognition

机译:8.93个顶部/ W LSTM经常性神经网络加速器,具有用于设备的分层粗粒稀疏性,用于设备上的语音识别

获取原文
获取原文并翻译 | 示例

摘要

Long short-term memory (LSTM) is a type of recurrent neural networks (RNNs), which is widely used for time-series data and speech applications, due to its high accuracy on such tasks. However, LSTMs pose difficulties for efficient hardware implementation because they require a large amount of weight storage and exhibit computation complexity. Prior works have proposed compression techniques to alleviate the storage/computation requirements of LSTMs but elementwise sparsity schemes incur sizable index memory overhead and structured compression techniques report limited compression ratios. In this article, we present an energy-efficient LSTM RNN accelerator, featuring an algorithm-hardware co-optimized memory compression technique called hierarchical coarse-grain sparsity (HCGS). Aided by the HCGS-based blockwise recursive weight compression, we demonstrate LSTM networks with up to 16x fewer weights while achieving minimal error rate degradation. The prototype chip fabricated in 65-nm LP CMOS achieves up to 8.93 TOPS/W for real-time speech recognition using compressed LSTMs based on HCGS. HCGS-based LSTMs have demonstrated energy-efficient speech recognition with low error rates for TIMIT, TED-LIUM, and LibriSpeech data sets.
机译:长短期内存(LSTM)是一种经常性的神经网络(RNN),其广泛用于时间序列数据和语音应用,因为它对这些任务的高精度。然而,LSTMS对于有效的硬件实现困难,因为它们需要大量的重量存储并表现出计算复杂性。先前作品已经提出了压缩技术,以减轻LSTMS的存储/计算要求,但是元素稀疏性方案产生的可征收索引存储器开销和结构化压缩技术报告有限的压缩比。在本文中,我们提出了一个节能LSTM RNN加速器,其特征在于一种算法 - 硬件共同优化内存压缩技术,称为分层粗粒稀疏(HCG)。通过基于HCG的块块递归重量压缩,我们展示了最多16倍的权重的LSTM网络,同时实现最小的错误率劣化。使用基于HCG的压缩LSTMS,65-NM LP CMOS中制造的原型芯片可实现高达8.93个顶部/倍的实时语音识别。基于HCGS的LSTMS已经证明了节能语音识别,速度,TED Lium和LibrisPeech数据集的低误差率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号