首页> 外文会议>IEEE/ACM International Symposium on Low Power Electronics and Design >TEA-DNN: the Quest for Time-Energy-Accuracy Co-optimized Deep Neural Networks
【24h】

TEA-DNN: the Quest for Time-Energy-Accuracy Co-optimized Deep Neural Networks

机译:TEA-DNN:追求时能 - 精度协调深神经网络

获取原文
获取外文期刊封面目录资料

摘要

Embedded deep learning platforms have witnessed two simultaneous improvements. First, the accuracy of convolutional neural networks (CNNs) has been significantly improved through the use of automated neural-architecture search (NAS) algorithms to determine CNN structure. Second, there has been increasing interest in developing hardware accelerators for CNNs that provide improved inference performance and energy consumption compared to GPUs. Such embedded deep learning platforms differ in the amount of compute resources and memory-access bandwidth, which would affect performance and energy consumption of CNNs. It is therefore critical to consider the available hardware resources in the network architecture search. To this end, we introduce TEA-DNN, a NAS algorithm targeting multi-objective optimization of execution time, energy consumption, and classification accuracy of CNN workloads on embedded architectures. TEA-DNN leverages energy and execution time measurements on embedded hardware when exploring the Pareto-optimal curves across accuracy, execution time, and energy consumption and does not require additional effort to model the underlying hardware. We apply TEA-DNN for image classification on actual embedded platforms (NVIDIA Jetson TX2 and Intel Movidius Neural Compute Stick). We highlight the Pareto-optimal operating points that emphasize the necessity to explicitly consider hardware characteristics in the search process. To the best of our knowledge, this is the most comprehensive study of Pareto-optimal models across a range of hardware platforms using actual measurements on hardware to obtain objective values.
机译:嵌入式深度学习平台目睹了两个同时改进。首先,通过使用自动神经结构搜索(NAS)算法来确定CNN结构,已经显着改善了卷积神经网络(CNNS)的准确性。其次,对CNN的开发硬件加速器的兴趣越来越兴趣,与GPU相比提供了改进的推理性能和能量消耗。这种嵌入的深度学习平台在计算资源和内存访问带宽的数量中不同,这会影响CNN的性能和能量消耗。因此,考虑网络架构搜索中的可用硬件资源是至关重要的。为此,我们介绍了嵌入式架构上CNN工作负载的多目标优化的NAS算法,瞄准了多目标优化的NAS算法。 TEA-DNN在精度,执行时间和能量消耗探索Pareto-Optimal曲线时利用能量和执行时间测量在嵌入式硬件上,并且不需要额外的努力来模拟底层硬件。我们在实际嵌入式平台上应用TEA-DNN进行图像分类(NVIDIA Jetson TX2和Intel Movidius神经计算棒)。我们突出了帕累托 - 最佳操作点,强调必须明确地考虑搜索过程中的硬件特征。据我们所知,这是使用硬件上的实际测量来获得客观值的一系列硬件平台的帕累托最佳模型最全面的研究。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号