首页> 外文会议>Asia and South Pacific Design Automation Conference >EFFORT: Enhancing Energy Efficiency and Error Resilience of a Near-Threshold Tensor Processing Unit
【24h】

EFFORT: Enhancing Energy Efficiency and Error Resilience of a Near-Threshold Tensor Processing Unit

机译:EFFORT:增强近阈值张量处理单元的能效和抗错能力

获取原文

摘要

Modern deep neural network (DNN) applications demand a remarkable processing throughput usually unmet by traditional Von Neumann architectures. Consequently, hardware accelerators, comprising a sea of multiplier and accumulate (MAC) units, have recently gained prominence in accelerating DNN inference engine. For example, Tensor Processing Units (TPU) account for a lion's share of Google's datacenter inference operations. The proliferation of real-time DNN predictions is accompanied with a tremendous energy budget. In quest of trimming the energy footprint of DNN accelerators, we propose EFFORT-an energy optimized, yet high performance TPU architecture, operating at the Near-Threshold Computing (NTC) region. EFFORT promotes a better-than-worst-case design by operating the NTC TPU at a substantially high frequency while keeping the voltage at the NTC nominal value. In order to tackle the timing errors due to such aggressive operation, we employ an opportunistic error mitigation strategy. Additionally, we implement an in-situ clock gating architecture, drastically reducing the MACs' dynamic power consumption. Compared to a cutting-edge error mitigation technique for TPUs, EFFORT enables up to 2.5× better performance at NTC with only 2% average accuracy drop across 3 out of 4 DNN datasets.
机译:现代深度神经网络(DNN)应用程序要求卓越的处理吞吐量,这通常是传统的冯·诺依曼架构无法满足的。因此,包含加速器和累加(MAC)单元的硬件加速器最近在加速DNN推理引擎方面获得了突出的地位。例如,张量处理单元(TPU)占Google数据中心推理操作的最大份额。实时DNN预测的激增伴随着巨大的能源预算。为了缩小DNN加速器的能耗,我们提出了EFFORT,这是一种经过能源优化的高性能TPU架构,在近阈值计算(NTC)区域运行。 EFFORT通过在相当高的频率下运行NTC TPU,同时将电压保持在NTC标称值,从而促进了一种比最坏情况更好的设计。为了解决由于此类激进操作而导致的时序错误,我们采用了机会主义的错误缓解策略。此外,我们实现了原地时钟门控架构,从而大大降低了MAC的动态功耗。与用于TPU的尖端错误缓解技术相比,EFFORT在NTC上的性能提高了2.5倍,在4个DNN数据集中的3个中,平均准确率仅下降了2%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号