首页> 外文会议>International Symposium on Microarchitecture >ASPEN: A Scalable In-SRAM Architecture for Pushdown Automata
【24h】

ASPEN: A Scalable In-SRAM Architecture for Pushdown Automata

机译:Aspen:用于下推自动机的可扩展内的SRAM架构

获取原文

摘要

Many applications process some form of tree-structured or recursively-nested data, such as parsing XML or JSON web content as well as various data mining tasks. Typical CPU processing solutions are hindered by branch misprediction penalties while attempting to reconstruct nested structures and also by irregular memory access patterns. Recent work has demonstrated improved performance for many data processing applications through memory-centric automata processing engines. Unfortunately, these architectures do not support a computational model rich enough for tasks such as XML parsing. In this paper, we present ASPEN, a general-purpose, scalable, and reconfigurable memory-centric architecture for processing of tree-like data. We take inspiration from previous automata processing architectures, but support the richer deterministic pushdown automata computational model. We propose a custom datapath capable of performing the state matching, stack manipulation, and transition routing operations of pushdown automata, all efficiently stored and computed in memory arrays. Further, we present compilation algorithms for transforming large classes of existing grammars to pushdown automata executable on ASPEN, and demonstrate their effectiveness on four different languages: Cool (object oriented programming), DOT (graph visualization), JSON, and XML. Finally, we present an empirical evaluation of two application scenarios for ASPEN: XML parsing, and frequent subtree mining. The proposed architecture achieves an average 704.5 ns per KB parsing XML compared to 9983 ns per KB in a state-of-the-art XML parser across 23 benchmarks. We also demonstrate a 37.2x and 6x better end-to-end speedup over CPU and GPU implementations of subtree mining.
机译:许多应用程序处理某种形式的树结构或递归嵌套数据,例如解析XML或JSON Web内容以及各种数据挖掘任务。典型的CPU处理解决方案因分支错误规定的惩罚而受到阻碍,同时尝试重建嵌套结构以及不规则的存储器访问模式。最近的工作已经通过以内存为中心的自动处理引擎对许多数据处理应用进行了改进的性能。不幸的是,这些架构不支持足够丰富的计算模型,以获得XML解析等任务。在本文中,我们提出了Aspen,一种通用,可扩展和可重构的内存中心架构,用于处理类似树的数据。我们从以前的自动机加工架构中获取灵感,但支持更丰富的确定性推动自动数据计算模型。我们提出了一种自定义数据路径,能够执行推动自动机的状态匹配,堆栈操作和转换路由操作,全部有效地存储和计算在存储器阵列中。此外,我们提供用于将大类现有语法转换为在Aspen上可执行的汇编算法,并在四种不同语言上展示其有效性:COOL(面向对象编程),点(图形可视化),JSON和XML。最后,我们对ASPEN的两个应用场景提供了一个实证评估:XML解析和频繁的子树挖掘。拟议的体系结构平均每kB分析XML平均达到704.5ns,而在23个基准中的最先进的XML解析器中为每kB为9983 ns。我们还展示了37.2倍和6倍的最终到最终加速,通过CPU和子树挖掘的GPU实现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号