首页> 外文期刊>Very Large Scale Integration (VLSI) Systems, IEEE Transactions on >Memory-Augmented Neural Networks on FPGA for Real-Time and Energy-Efficient Question Answering
【24h】

Memory-Augmented Neural Networks on FPGA for Real-Time and Energy-Efficient Question Answering

机译:关于FPGA的内存增强神经网络,用于实时和节能问题应答

获取原文
获取原文并翻译 | 示例
       

摘要

Memory-augmented neural networks (MANNs) were introduced to handle long-term dependent data efficiently. MANNs have shown promising results in question answering (QA) tasks that require holding contexts for answering a given question. As demands for QA on edge devices have increased, the utilization of MANNs in resource-constrained environments has become important. To achieve fast and energy-efficient inference of MANNs, we can exploit application-specific hardware accelerators on field-programmable gate arrays (FPGAs). Although several accelerators for conventional deep neural networks have been designed, it is difficult to efficiently utilize the accelerators with MANNs due to different requirements. In addition, characteristics of QA tasks should be considered for further improving the efficiency of inference on the accelerators. To address the aforementioned issues, we propose an inference accelerator of MANNs on FPGA. To fully utilize the proposed accelerator, we introduce fast inference methods considering the features of QA tasks. To evaluate our proposed approach, we implemented the proposed architecture on an FPGA and measured the execution time and energy consumption for the bAbI data set. According to our thorough experiments, the proposed methods improved speed and energy efficiency of the inference of MANNs up to about 25.6 and 28.4 times, respectively, compared with those of CPU.
机译:引入内存增强的神经网络(MANNS)以有效地处理长期相关数据。 MANNS已经显示出有前途的结果,要求持有持有上下文以回答给定问题的问题。随着对边缘设备对QA的需求增加,人体在资源受限环境中的利用变得重要。为了实现MANNS的快速和节能推断,我们可以在现场可编程门阵列(FPGA)上利用特定于应用的硬件加速器。虽然设计了几种用于传统的深神经网络的加速器,但由于不同的要求,难以有效地利用与人工人物的加速器。此外,应考虑QA任务的特性,以进一步提高加速器的推理效率。为解决上述问题,我们提出了在FPGA上的曼诺的推论加速器。为了充分利用所提出的加速器,我们介绍考虑QA任务的特征的快速推断方法。为了评估我们所提出的方法,我们在FPGA上实施了所提出的架构,并测量了BABI数据集的执行时间和能耗。根据我们彻底的实验,与CPU的实验相比,拟议的方法分别提高了人数推理的速度和能量效率高达约25.6%和28.4倍。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号