【24h】

CAPE: A Content-Addressable Processing Engine

机译:CAPE:内容可寻址的处理引擎

获取原文

摘要

Processing-in-memory (PIM) architectures attempt to overcome the von Neumann bottleneck by combining computation and storage logic into a single component. The content-addressable parallel processing paradigm (CAPP) from the seventies is an in-situ PIM architecture that leverages content-addressable memories to realize bit-serial arithmetic and logic operations, via sequences of search and update operations over multiple memory rows in parallel. In this paper, we set out to investigate whether the concepts behind classic CAPP can be used successfully to build an entirely CMOS-based, general-purpose microarchitecture that can deliver manyfold speedups while remaining highly programmable. We conduct a full-stack design of a Content-Addressable Processing Engine (CAPE), built out of dense push-rule 6T SRAM arrays. CAPE is programmable using the RISC-V ISA with standard vector extensions. Our experiments show that CAPE achieves an average speedup of 14 (up to 254) over an area-equivalent (slightly under 9 mm2 at 7 nm) out-of-order processor core with three levels of caches.
机译:加工内存(PIM)架构尝试通过将计算和存储逻辑组合成单个组件来克服von neumann瓶颈。来自七十年代的内容可寻址并行处理范例(CAPP)是一种原位PIM架构,其利用内容可寻址的存储器来实现比特串行算术和逻辑操作,通过搜索和更新在多个存储行上并行上的多个存储行。在本文中,我们开始调查经典CAPP后面的概念是否可以成功地用于构建一个完全基于CMOS的通用微体系结构,可以提供多重加速度,同时保持高度可编程。我们开展内容可寻址处理引擎(CAPE)的全堆叠设计,采用密集推挽规则6T SRAM阵列。 CAPE是使用标准矢量扩展的RISC-V ISA进行编程。我们的实验表明,Cape通过面积等效物(略低于9毫米)的平均加速14(最多254) 2 在7个nm)以上的超级处理器核心,具有三个级别的高速缓存。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号