首页> 外文会议>International Conference on Application-specific Systems, Architectures and Processors >An Adaptive Memory Management Strategy Towards Energy Efficient Machine Inference in Event-Driven Neuromorphic Accelerators
【24h】

An Adaptive Memory Management Strategy Towards Energy Efficient Machine Inference in Event-Driven Neuromorphic Accelerators

机译:事件驱动的神经形态加速器中面向节能机器推理的自适应内存管理策略

获取原文

摘要

Spiking neural networks are viable alternatives to classical neural networks for edge processing in low-power embedded and IoT devices. To reap their benefits, neuromorphic network accelerators that tend to support deep networks still have to expend great effort in fetching synaptic states from a large remote memory. Since local computation in these networks is event-driven, memory becomes the major part of the system's energy consumption. In this paper, we explore various opportunities of data reuse that can help mitigate the redundant traffic for retrieval of neuron meta-data and post-synaptic weights. We describe CyNAPSE, a baseline neural processing unit and its accompanying software simulation as a general template for exploration on various levels. We then investigate the memory access patterns of three spiking neural network benchmarks that have significantly different topology and activity. With a detailed study of locality in memory traffic, we establish the factors that hinder conventional cache management philosophies from working efficiently for these applications. To that end, we propose and evaluate a domain-specific management policy that takes advantage of the forward visibility of events in a queue-based event-driven simulation framework. Subsequently, we propose network-adaptive enhancements to make it robust to network variations. As a result, we achieve 13-44% reduction in system power consumption and 8-23% improvement over conventional replacement policies.
机译:尖刺神经网络是低功耗嵌入式和IoT设备中边缘处理的经典神经网络的可行替代方案。为了获得其好处,趋向于支持深度网络的神经形态网络加速器仍然必须花费大量的精力来从大型远程存储器中获取突触状态。由于这些网络中的本地计算是事件驱动的,因此内存成为系统能耗的主要部分。在本文中,我们探索了数据重用的各种机会,这些机会可以帮助减轻用于检索神经元元数据和突触后权重的冗余流量。我们将CyNAPSE,基线神经处理单元及其随附的软件仿真描述为在各个级别进行探索的通用模板。然后,我们研究具有明显不同的拓扑和活动的三个尖峰神经网络基准测试的内存访问模式。通过详细研究内存流量的局部性,我们建立了阻碍常规缓存管理理念有效地为这些应用程序工作的因素。为此,我们提出并评估了特定于域的管理策略,该策略利用了基于队列的事件驱动的仿真框架中事件的前瞻性可见性。随后,我们提出了网络自适应增强功能,使其对网络变化具有鲁棒性。结果,与传统的更换策略相比,我们的系统功耗降低了13-44%,改进了8-23%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号