首页> 外文会议>IEEE International Solid- State Circuits Conference >F1: Striking the Balance Between Energy Efficiency Flexibility: General-Purpose vs Special-Purpose ML Processors
【24h】

F1: Striking the Balance Between Energy Efficiency Flexibility: General-Purpose vs Special-Purpose ML Processors

机译:F1:敲击能源效率与灵活性之间的平衡:通用VS专用ML处理器

获取原文

摘要

The forum provides a comprehensive full-stack (hardware and software) view of ML acceleration from cloud to edge. The first talk focuses on the main design and benchmarking challenges facing large general-purpose accelerators, including multi-die scaling, and describes strategies for conducting relevant research as the complexity gap between research prototype and product continues to widen. The second talk looks at how to leverage and specialize the open-source RISC-V ISA for edge ML, exploring the trade-offs between different forms of acceleration such as lightweight ISA extensions and tightly-coupled memory accelerators. The third talk details an approach based on a practical unified architecture for ML that can be easily “tailored” to fit in different scenarios ranging from smart watches, smartphones, autonomous cars to intelligent cloud. The fourth talk explores the co-design of hardware and DNN models to achieve state-of-the-art performance for real-time, extremely energy/throughput-constrained inference applications. The fifth talk deals with ML on reconfigurable logic, discussing many examples of forms of specializations implemented on FPGAs and their impact on potential applications, flexibility, performance and efficiency. The sixth talk describes the software complexities for enabling ML APIs for various different types of specialized hardware accelerators (GPU, TPUs, including EdgeTPU). The seventh talk look into how to optimize the training process for sparse and low-precision network models for general platforms as well as next-generation memristor-based ML engines.
机译:该论坛提供了一个全面的全堆叠(硬件和软件)视图的ML加速度从云到Edge。第一款谈话侧重于大型通用加速器面临的主要设计和基准挑战,包括多模缩放,并描述了在研究原型和产品之间的复杂性差距继续扩大的复杂性差距。第二次谈话介绍如何利用和专用开源RISC-V ISA for Edge ML,探索不同形式的加速度之间的权衡,例如轻量级ISA扩展和紧密耦合的内存加速器。第三次谈论详细说明了一种基于ML的实用统一架构的方法,可以很容易地“定制”,以适应不同的场景,从智能手表,智能手机,自动车辆到智能云。第四个谈话探讨了硬件和DNN模型的共同设计,以实现实时,极其能量/吞吐量约束推理应用的最先进的性能。第五次谈论可重构逻辑上的ML,讨论了在FPGA上实施的专业形式的许多例子及其对潜在应用,灵活性,性能和效率的影响。第六次谈话描述了用于为各种不同类型的专用硬件加速器启用ML API的软件复杂性(GPU,TPU,包括EDGEDPU)。第七次谈论如何优化用于普通平台的稀疏和低精密网络模型的培训过程以及基于下一代忆阻器的ML引擎。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号