首页> 外文期刊>Journal of instrumentation: an IOP and SISSA journal >Implementation of high-performance, sub-microsecond deep neural networks on FPGAs for trigger applications
【24h】

Implementation of high-performance, sub-microsecond deep neural networks on FPGAs for trigger applications

机译:用于触发器应用的FPGA上高性能,次微秒的深神经网络的实现

获取原文
获取原文并翻译 | 示例
           

摘要

Deep neural networks are already widely used for physics analysis, but there are only few applications within low-level hardware triggers, and typically only with small networks. Modern high-end FPGAs offer Tera-scalearith meticper formance, and there by provide a significant amount of operations per data set even for MHz-range data rates. We present a bottom-up approach of implementing typical neural network layers, in which we took into account both the special constraints that come from high-performance trigger systems, such as the ATLAS hardware trigger at the LHC, as well as an efficient implementation. By specifically designing each layer type to match our requirements, we could develop a framework that reaches 90 to 100% processing efficiency for large layers, requires only few extra resources for data flow and controlling, and offers latencies in the range of only tens to hundreds of nanoseconds for entire (deep) networks. Additionally, a toolkit was built around these optimized layer implementations, which facilitates the creation of the FPGA implementation of a trained NN model.
机译:深度神经网络已经广泛用于物理分析,但在低级硬件触发器中只有很少的应用,通常只有小型网络。现代高端FPGA提供TERA-Scalearith梅普尔格特,即使对于MHz范围数据速率,每个数据都提供了大量的操作。我们提出了实现典型神经网络层的自下而上的方法,其中我们考虑了来自高性能触发系统的特殊约束,例如LHC处的ATLAS硬件触发器,以及有效的实现。通过专门设计每层类型以匹配我们的要求,我们可以开发一个框架,达到大型层的90%到100%的处理效率,只需要额外的资源进行数据流和控制,并在仅限数十个范围内提供延迟整个(深)网络的纳秒。此外,围绕这些优化的图层实现构建了一个工具包,这有助于创建训练的NN模型的FPGA实现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号