首页> 外文期刊>Computer architecture news >Reducing Memory Access Latency with Asymmetric DRAM Bank Organizations
【24h】

Reducing Memory Access Latency with Asymmetric DRAM Bank Organizations

机译:通过非对称DRAM Bank组织减少内存访问延迟

获取原文
获取原文并翻译 | 示例

摘要

DRAM has been a de facto standard for main memory, and advances in process technology have led to a rapid increase in its capacity and bandwidth. In contrast, its random access latency has remained relatively stagnant, as it is still around 100 CPU clock cycles. Modern computer systems rely on caches or other latency tolerance techniques to lower the average access latency. However, not all applications have ample parallelism or locality that would help hide or reduce the latency. Moreover, applications' demands for memory space continue to grow, while the capacity gap between last-level caches and main memory is unlikely to shrink. Consequently, reducing the main-memory latency is important for application performance. Unfortunately, previous proposals have not adequately addressed this problem, as they have focused only on improving the bandwidth and capacity or reduced the latency at the cost of significant area overhead. We propose asymmetric DRAM bank organizations to reduce the average main-memory access latency. We first analyze the access and cycle times of a modern DRAM device to identify key delay components for latency reduction. Then we reorganize a subset of DRAM banks to reduce their access and cycle times by half with low area overhead. By synergistically combining these reorganized DRAM banks with support for non-uniform bank accesses, we introduce a novel DRAM bank organization with center high-aspect-ratio mats called CHARM. Experiments on a simulated chip-multiprocessor system show that CHARM improves both the instructions per cycle and system-wide energy-delay product up to 21% and 32%, respectively, with only a 3% increase in die area.
机译:DRAM已经成为主存储器的事实上的标准,并且处理技术的进步导致其容量和带宽迅速增加。相比之下,它的随机访问延迟仍然相对停滞,因为它仍在100个CPU时钟周期左右。现代计算机系统依靠缓存或其他等待时间容忍技术来降低平均访问等待时间。但是,并非所有应用程序都具有足够的并行度或局部性,有助于隐藏或减少延迟。此外,应用程序对内存空间的需求持续增长,而最后一级缓存和主内存之间的容量差距不太可能缩小。因此,减少主内存延迟对于应用程序性能很重要。不幸的是,以前的提议没有充分解决这个问题,因为它们只集中在提高带宽和容量或减少等待时间上,而代价是大量的区域开销。我们建议使用非对称的DRAM银行组织来减少平均主存储器访问延迟。我们首先分析现代DRAM设备的访问时间和循环时间,以确定关键延迟组件以减少等待时间。然后,我们重组了一部分DRAM存储库,以减少它们的访问和循环时间,同时降低了区域开销。通过将重组后的DRAM存储区与对非均匀存储区访问的支持进行协同组合,我们引入了一种新颖的DRAM存储区组织,其中心高纵横比垫称为CHARM。在模拟的芯片多处理器系统上进行的实验表明,CHARM可以将每个周期的指令和整个系统的能量延迟积分别提高21%和32%,而芯片面积仅增加3%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号