首页> 外文会议>IEEE/ACM International Symposium on Low Power Electronics and Design >TEA-DNN: the Quest for Time-Energy-Accuracy Co-optimized Deep Neural Networks
【24h】

TEA-DNN: the Quest for Time-Energy-Accuracy Co-optimized Deep Neural Networks

机译:TEA-DNN:对时间-能量-精度共同优化的深度神经网络的追求

获取原文

摘要

Embedded deep learning platforms have witnessed two simultaneous improvements. First, the accuracy of convolutional neural networks (CNNs) has been significantly improved through the use of automated neural-architecture search (NAS) algorithms to determine CNN structure. Second, there has been increasing interest in developing hardware accelerators for CNNs that provide improved inference performance and energy consumption compared to GPUs. Such embedded deep learning platforms differ in the amount of compute resources and memory-access bandwidth, which would affect performance and energy consumption of CNNs. It is therefore critical to consider the available hardware resources in the network architecture search. To this end, we introduce TEA-DNN, a NAS algorithm targeting multi-objective optimization of execution time, energy consumption, and classification accuracy of CNN workloads on embedded architectures. TEA-DNN leverages energy and execution time measurements on embedded hardware when exploring the Pareto-optimal curves across accuracy, execution time, and energy consumption and does not require additional effort to model the underlying hardware. We apply TEA-DNN for image classification on actual embedded platforms (NVIDIA Jetson TX2 and Intel Movidius Neural Compute Stick). We highlight the Pareto-optimal operating points that emphasize the necessity to explicitly consider hardware characteristics in the search process. To the best of our knowledge, this is the most comprehensive study of Pareto-optimal models across a range of hardware platforms using actual measurements on hardware to obtain objective values.
机译:嵌入式深度学习平台见证了两个同时的改进。首先,通过使用自动神经体系结构搜索(NAS)算法确定CNN结构,卷积神经网络(CNN)的准确性得到了显着提高。其次,人们对开发CNN硬件加速器的兴趣日益浓厚,与GPU相比,该硬件加速器可提供更好的推理性能和能耗。此类嵌入式深度学习平台的计算资源数量和内存访问带宽不同,这将影响CNN的性能和能耗。因此,至关重要的是要在网络体系结构搜索中考虑可用的硬件资源。为此,我们介绍了TEA-DNN,这是一种NAS算法,旨在对嵌入式体系结构上CNN工作负载的执行时间,能耗和分类精度进行多目标优化。当探索准确性,执行时间和能耗方面的帕累托最优曲线时,TEA-DNN会利用嵌入式硬件上的能量和执行时间测量,而无需花费额外的精力来对底层硬件进行建模。我们将TEA-DNN应用到实际嵌入式平台(NVIDIA Jetson TX2和Intel Movidius神经计算棒)上的图像分类。我们重点介绍了帕累托最优操作​​点,强调了在搜索过程中明确考虑硬件特征的必要性。据我们所知,这是对各种硬件平台上的帕累托最优模型的最全面的研究,使用对硬件的实际测量来获得目标值。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号