【24h】

4th International Workshop on On-Chip Memory Hierarchies and Interconnects (OMHI)

机译:第四届片上存储器层次结构和互连国际研讨会(OMHI)

获取原文

摘要

Current chip multiprocessors (CMPs) include several levels of on-chip caches to avoid the huge latencies of accessing the off-chip DRAM main memory modules. These caches must be efficiently interconnected to avoid performance penalties. On-chip networks are used to interconnect the memory hierarchy inside the processor chip. Latencies can be significantly affected by the devised on-chip memory hierarchy and the interconnect design, whose impact on the overall latency strongly depends on the core count. Consequently, this problem aggravates with the increasing core counts, which is the current commercial trend. By contrast, the main concern in GPUs is on memory bandwidth instead of latencies. Current GPUs are designed to hide memory latencies through fine-grained multithreading. The main goal of on-chip memories in current GPUs is to reduce off-chip memory traffic. In this context, the programmer plays a key role in improving cache access locality. Hence we can conclude that CPUs and GPUs require memory organizations with different characteristics. Thus, as current heterogeneous CPU-GPU systems are proliferating in the market, the memory system must be designed to efficiently support both types of memory organizations: latency-oriented and bandwidth-oriented.
机译:当前的芯片多处理器(CMP)包括几级片上高速缓存,以避免访问片外DRAM主存储模块的巨大延迟。这些缓存必须有效地互连以避免性能损失。片上网络用于互连处理器芯片内部的存储器层次结构。设计的片上存储器层次结构和互连设计会严重影响延迟,而延迟对整体延迟的影响则很大程度上取决于内核数。因此,该问题随着核心数量的增加而加剧,这是当前的商业趋势。相比之下,GPU中的主要问题在于内存带宽而不是延迟。当前的GPU旨在通过细粒度的多线程来隐藏内存延迟。当前GPU中的片上存储器的主要目标是减少片外存储器流量。在这种情况下,程序员在提高缓存访问位置方面起着关键作用。因此,我们可以得出结论,CPU和GPU需要具有不同特征的内存组织。因此,随着当前异构CPU-GPU系统在市场上的激增,必须将存储器系统设计为有效地支持两种类型的存储器组织:面向延迟和面向带宽。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号