首页> 外文会议>2018 Design, Automation amp; Test in Europe Conference amp; Exhibition >Exploiting approximate computing for deep learning acceleration
【24h】

Exploiting approximate computing for deep learning acceleration

机译:利用近似计算促进深度学习

获取原文
获取原文并翻译 | 示例

摘要

Deep Neural Networks (DNNs) have emerged as a powerful and versatile set of techniques to address challenging artificial intelligence (AI) problems. Applications in domains such as image/video processing, natural language processing, speech synthesis and recognition, genomics and many others have embraced deep learning as the foundational technique. DNNs achieve superior accuracy for these applications using very large models which require 100s of MBs of data storage, ExaOps of computation and high bandwidth for data movement. Despite advances in computing systems, training state-of-the-art DNNs on large datasets takes several days/weeks, directly limiting the pace of innovation and adoption. In this paper, we discuss how these challenges can be addressed via approximate computing. Based on our earlier studies demonstrating that DNNs are resilient to numerical errors from approximate computing, we present techniques to reduce communication overhead of distributed deep learning training via adaptive residual gradient compression (AdaComp), and computation cost for deep learning inference via Prameterized clipping ACTivation (PACT) based network quantization. Experimental evaluation demonstrates order of magnitude savings in communication overhead for training and computational cost for inference while not compromising application accuracy.
机译:深度神经网络(DNN)已成为解决挑战性人工智能(AI)问题的功能强大且用途广泛的技术组合。图像/视频处理,自然语言处理,语音合成和识别,基因组学等领域的应用已将深度学习作为基础技术。 DNN使用非常大的模型为这些应用程序提供了卓越的精度,这些模型需要100 MB的数据存储,ExaOps计算和高带宽的数据移动能力。尽管计算系统取得了进步,但在大型数据集上训练最先进的DNN仍需要数天/周,这直接限制了创新和采用的速度。在本文中,我们讨论了如何通过近似计算解决这些挑战。基于我们先前的研究表明DNN可以抵抗近似计算带来的数值误差,我们提出了通过自适应残差梯度压缩(AdaComp)减少分布式深度学习训练的通信开销的技术,以及通过参数化削波ACTivation减少深度学习推理的计算成本的技术(基于PACT)的网络量化。实验评估表明,在不牺牲应用精度的前提下,可以节省大量通信训练开销和推理计算成本。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号