首页> 外文期刊>IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems >Standing on the Shoulders of Giants: Hardware and Neural Architecture Co-Search With Hot Start
【24h】

Standing on the Shoulders of Giants: Hardware and Neural Architecture Co-Search With Hot Start

机译:站在巨人的肩膀上:硬件和神经结构与热门开始共同搜索

获取原文
获取原文并翻译 | 示例
       

摘要

Hardware and neural architecture co-search that automatically generates artificial intelligence (AI) solutions from a given dataset are promising to promote AI democratization; however, the amount of time that is required by current co-search frameworks is in the order of hundreds of GPU hours for one target hardware. This inhibits the use of such frameworks on commodity hardware. The root cause of the low efficiency in existing co-search frameworks is the fact that they start from a "cold" state (i.e., search from scratch). In this article, we propose a novel framework, namely, HotNAS, that starts from a "hot" state based on a set of existing pretrained models (also known as model zoo) to avoid lengthy training time. As such, the search time can be reduced from 200 GPU hours to less than 3 GPU hours. In HotNAS, in addition to hardware design space and neural architecture search space, we further integrate a compression space to conduct model compressing during the co-search, which creates new opportunities to reduce latency, but also brings challenges. One of the key challenges is that all of the above search spaces are coupled with each other, e.g., compression may not work without hardware design support. To tackle this issue, HotNAS builds a chain of tools to design hardware to support compression, based on which a global optimizer is developed to automatically co-search all the involved search spaces. Experiments on ImageNet dataset and Xilinx FPGA show that, within the timing constraint of 5 ms, neural architectures generated by HotNAS can achieve up to 5.79% Top-1 and 3.97% Top-5 accuracy gain, compared with the existing ones.
机译:从给定数据集自动生成人工智能(AI)解决方案的硬件和神经结构的共同搜索很有希望促进AI民主化;但是,当前共同搜索框架所需的时间量为一个目标硬件的数百个GPU小时数。这禁止使用这种框架对商品硬件。现有共同搜索框架中低效率的根本原因是它们从“冷”状态开始(即,从头开始搜索)。在本文中,我们提出了一种新颖的框架,即热带,从“热”状态开始,这是基于一套现有的预磨模(也称为模型动物园)来避免冗长的训练时间。因此,搜索时间可以从200GPu小时减少到小于3GPu小时。在Hotnas中,除了硬件设计空间和神经结构搜索空间之外,我们还将压缩空间进一步集成,以在共同搜索期间进行模型压缩,从而创造了降低延迟的新机会,也可以带来挑战。关键挑战之一是所有上述搜索空间都彼此耦合,例如,在没有硬件设计支持的情况下压缩可能无法工作。为了解决这个问题,HotNAS构建了一系列工具来设计硬件以支持压缩,基于哪个全局优化器以自动共同搜索所有涉及的搜索空间。 Imagenet DataSet和Xilinx FPGA的实验表明,在5毫秒的时序约束中,HotNA产生的神经架构可以实现高达5.79%的顶级 - 1和3.97%的前5个精度增益,比如现有的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号