首页> 外文会议>IEEE International Symposium on Performance Analysis of Systems and Software >CLAN: Continuous Learning using Asynchronous Neuroevolution on Commodity Edge Devices
【24h】

CLAN: Continuous Learning using Asynchronous Neuroevolution on Commodity Edge Devices

机译:CLAN:在商品边缘设备上使用异步神经进化进行持续学习

获取原文

摘要

Recent advancements in machine learning algorithms, especially the development of Deep Neural Networks (DNNs) have transformed the landscape of Artificial Intelligence (AI). With every passing day, deep learning based methods are applied to solve new problems with exceptional results. The portal to the real world is the edge. The true impact of AI can only be fully realized if we can have AI agents continuously interacting with the real world and solving everyday problems. Unfortunately, high compute and memory requirements of DNNs acts a huge barrier towards this vision. Today we circumvent this problem by deploying special purpose inference hardware on the edge while procuring trained models from the cloud. This approach, however, relies on constant interaction with the cloud for transmitting all the data, training on massive GPU clusters, and downloading updated models. This is challenging for bandwidth, privacy, and constant connectivity concerns that autonomous agents may exhibit. In this paper we evaluate techniques for enabling adaptive intelligence on edge devices with zero interaction with any high-end cloud/server. We build a prototype distributed system of Raspberry Pis communicating via WiFi running NeuroEvolutionary (NE) learning and inference. We evaluate the performance of such a collaborative system and detail the compute/communication characteristics of different arrangements of the system that trade-off parallelism versus communication. Using insights from our analysis, we also propose algorithmic modifications to reduce communication by up to 3.6x during the learning phase to enhance scalability even further and match performance of higher end computing devices at scale. We believe that these insights will enable algorithm-hardware co-design efforts for enabling continuous learning on the edge.
机译:机器学习算法的最新进展,尤其是深度神经网络(DNN)的发展,改变了人工智能(AI)的格局。每天,基于深度学习的方法都被用来解决新问题,并取得了非凡的效果。通往现实世界的门户是边缘。只有让AI代理不断与现实世界互动并解决日常问题,才能完全实现AI的真正影响。不幸的是,DNN的高计算和内存要求对实现这一愿景构成了巨大障碍。今天,我们通过在边缘部署专用推理硬件同时从云中获取经过训练的模型来规避此问题。但是,这种方法依靠与云的持续交互来传输所有数据,在大型GPU集群上进行训练以及下载更新的模型。这对于自治代理可能表现出的带宽,隐私和持续的连接性问题具有挑战性。在本文中,我们评估了在与任何高端云/服务器零交互的边缘设备上启用自适应智能的技术。我们构建了Raspberry Pis分布式原型系统,该系统通过运行NeuroEvolutionary(NE)学习和推理的WiFi进行通信。我们评估了这种协作系统的性能,并详细介绍了权衡并行性与通信性的系统不同安排的计算/通信特性。利用我们的分析见解,我们还提出了算法修改方案,以在学习阶段将通信减少多达3.6倍,以进一步增强可扩展性并大规模匹配高端计算设备的性能。我们相信,这些见识将有助于算法-硬件协同设计工作,以实现在边缘的持续学习。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号