首页> 外文会议>Design, Automation Test in Europe Conference Exhibition >Exploiting Approximate Computing for Deep Learning Acceleration
【24h】

Exploiting Approximate Computing for Deep Learning Acceleration

机译:利用深度学习加速的近似计算

获取原文

摘要

Deep Neural Networks (DNNs) have emerged as a powerful and versatile set of techniques to address challenging artificial intelligence (AI) problems. Applications in domains such as image/video processing, natural language processing, speech synthesis and recognition, genomics and many others have embraced deep learning as the foundational technique. DNNs achieve superior accuracy for these applications using very large models which require 100s of MBs of data storage, ExaOps of computation and high bandwidth for data movement. Despite advances in computing systems, training state-of-the-art DNNs on large datasets takes several days/weeks, directly limiting the pace of innovation and adoption. In this paper, we discuss how these challenges can be addressed via approximate computing. Based on our earlier studies demonstrating that DNNs are resilient to numerical errors from approximate computing, we present techniques to reduce communication overhead of distributed deep learning training via adaptive residual gradient compression (AdaComp), and computation cost for deep learning inference via Prameterized clipping ACTivation (PACT) based network quantization. Experimental evaluation demonstrates order of magnitude savings in communication overhead for training and computational cost for inference while not compromising application accuracy.
机译:深度神经网络(DNN)已成为一种强大而多功能的技术,以解决挑战人工智能(AI)问题。在图像/视频处理,自然语言处理,语音合成和识别,基因组学和许多其他域等域中的应用程序都将深入学习作为基础技术。使用非常大型型号来实现这些应用的卓越精度,这些应用需要1000 MB的数据存储,exaops的计算和高带宽进行数据移动。尽管计算系统进步,但大型数据集上的培训最先进的DNN需要几天/周,直接限制了创新和采用的步伐。在本文中,我们讨论了如何通过近似计算解决这些挑战。基于我们之前的研究,证明DNN对近似计算的数值误差是有弹性的,我们通过自适应残留梯度压缩(ADACOMP)来降低分布式深度学习训练的通信开销的技术,以及通过跳闸剪切激活的深度学习推断的计算成本(基于PATT)的网络量化。实验评估表明了用于训练和计算成本的通信开销中的幅度节省的秩序,而推理的推理同时没有损害应用准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号