首页> 外文期刊>Computer architecture news >There and Back Again: Optimizing the Interconnect in Networks of Memory Cubes
【24h】

There and Back Again: Optimizing the Interconnect in Networks of Memory Cubes

机译:一遍又一遍:优化存储立方体网络中的互连

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

High-performance computing, enterprise, and datacenter servers are driving demands for higher total memory capacity as well as memory performance. Memory "cubes" with high per-package capacity (from 3D integration) along with high-speed point-to-point interconnects provide a scalable memory system architecture with the potential to deliver both capacity and performance. Multiple such cubes connected together can form a "Memory Network" (MN), but the design space for such MNs is quite vast, including multiple topology types and multiple memory technologies per memory cube. In this work, we first analyze several MN topologies with different mixes of memory package technologies to understand the key tradeoffs and bottlenecks for such systems. We find that most of a MN's performance challenges arise from the interconnection network that binds the memory cubes together. In particular, arbitration schemes used to route through MNs, ratio of NVM to DRAM, and specific topologies used have dramatic impact on performance and energy results. Our initial analysis indicates that introducing non-volatile memory to the MN presents a unique tradeoff between memory array latency and network latency. We observe that placing NVM cubes in a specific order in the MN improves performance by reducing the network size/diameter up to a certain NVM to DRAM ratio. Novel MN topologies and arbitration schemes also provide performance and energy deltas by reducing the hop count of requests and response in the MN. Based on our analyses, we introduce three techniques to address MN latency issues: (1) Distance-based arbitration scheme to improve queuing latencies throughout the network, (2) skip-list topology, derived from the classic data structure, to improve network latency and link usage, and (3) the MetaCube, a denser memory cube that leverages advanced packaging technologies to improve latency by reducing MN size.
机译:高性能计算,企业和数据中心服务器推动了对更高的总内存容量以及内存性能的需求。具有高封装能力(来自3D集成)的存储器“多维数据集”以及高速点对点互连,提供了可扩展的存储器系统体系结构,具有同时交付容量和性能的潜力。连接在一起的多个此类多维数据集可以形成一个“内存网络”(MN),但是此类MN的设计空间非常大,每个内存多维数据集包括多种拓扑类型和多种内存技术。在这项工作中,我们首先分析具有不同混合内存封装技术的几种MN拓扑,以了解此类系统的关键权衡和瓶颈。我们发现,MN的大多数性能挑战来自将内存多维数据集绑定在一起的互连网络。特别是,用于通过MN进行路由的仲裁方案,NVM与DRAM的比率以及所使用的特定拓扑结构对性能和能耗结果产生了巨大影响。我们的初步分析表明,将非易失性内存引入MN会在内存阵列延迟和网络延迟之间做出独特的权衡。我们观察到,将NVM多维数据集以特定顺序放置在MN中,可以通过将网络大小/直径减小到特定的NVM与DRAM比率来提高性能。新的MN拓扑和仲裁方案还通过减少MN中的请求和响应的跳数来提供性能和能量增量。根据我们的分析,我们引入三种技术来解决MN延迟问题:(1)基于距离的仲裁方案,以改善整个网络的排队等待时间;(2)从经典数据结构派生的跳过列表拓扑,以改善网络延迟(3)MetaCube,这是一种密度更高的存储多维数据集,它利用高级封装技术通过减小MN的大小来提高延迟。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号