首页> 外文期刊>Computer architecture news >Efficient Synonym Filtering and Scalable Delayed Translation for Hybrid Virtual Caching
【24h】

Efficient Synonym Filtering and Scalable Delayed Translation for Hybrid Virtual Caching

机译:混合虚拟缓存的有效同义词过滤和可伸缩延迟翻译

获取原文
获取原文并翻译 | 示例

摘要

Conventional translation look-aside buffers (TLBs) are required to complete address translation with short latencies, as the address translation is on the critical path of all memory accesses even for L1 cache hits. Such strict TLB latency restrictions limit the TLB capacity, as the latency increase with large TLBs may lower the overall performance even with potential TLB miss reductions. Furthermore, TLBs consume a significant amount of energy as they are accessed for every instruction fetch and data access. To avoid the latency restriction and reduce the energy consumption, virtual caching techniques have been proposed to defer translation to after L1 cache misses. However, an efficient solution for the synonym problem has been a critical issue hindering the wide adoption of virtual caching. Based on the virtual caching concept, this study proposes a hybrid virtual memory architecture extending virtual caching to the entire cache hierarchy, aiming to improve both performance and energy consumption. The hybrid virtual caching uses virtual addresses augmented with address space identifiers (ASID) in the cache hierarchy for common non-synonym addresses. For such non-synonyms, the address translation occurs only after last-level cache (LLC) misses. For uncommon synonym addresses, the addresses are translated to physical addresses with conventional TLBs before L1 cache accesses. To support such hybrid translation, we propose an efficient synonym detection mechanism based on Bloom filters which can identify synonym candidates with few false positives. For large memory applications, delayed translation alone cannot solve the address translation problem, as fixed-granularity delayed TLBs may not scale with the increasing memory requirements. To mitigate the translation scalability problem, this study proposes a delayed many segment translation designed for the hybrid virtual caching. The experimental results show that our approach effectively lowers accesses to the TLBs, leading to significant power savings. In addition, the approach provides performance improvement with scalable delayed translation with variable length segments.
机译:需要常规的转换后备缓冲器(TLB)以较短的延迟完成地址转换,因为地址转换位于所有内存访问的关键路径上,即使对于L1缓存命中也是如此。这种严格的TLB延迟限制限制了TLB的容量,因为即使使用大型TLB,延迟的增加也可能会降低整体性能,甚至可能会降低TLB的丢失率。此外,在每次取指令和访问数据时访问TLB时,都会消耗大量能量。为了避免等待时间限制并减少能耗,已提出虚拟缓存技术以将转换推迟到L1缓存未命中之后。但是,对于同义词问题的有效解决方案一直是阻碍虚拟缓存广泛采用的关键问题。基于虚拟缓存的概念,本研究提出了一种混合虚拟内存体系结构,将虚拟缓存扩展到整个缓存层次结构,旨在提高性能和能耗。混合虚拟高速缓存使用在高速缓存层次结构中添加了地址空间标识符(ASID)的虚拟地址作为普通的非同义词地址。对于此类非同义词,仅在最后一级缓存(LLC)未命中后才进行地址转换。对于不常见的同义词地址,在访问L1缓存之前,这些地址将使用常规TLB转换为物理地址。为了支持这种混合翻译,我们提出了一种基于Bloom过滤器的有效同义词检测机制,该机制可以识别几乎没有误报的同义词候选者。对于大内存应用程序,仅延迟转换无法解决地址转换问题,因为固定粒度的延迟TLB可能无法随着存储需求的增长而扩展。为了减轻翻译的可伸缩性问题,本研究提出了一种为混合虚拟缓存设计的延迟的多段翻译。实验结果表明,我们的方法有效地减少了对TLB的访问,从而显着节省了功率。另外,该方法通过可变长度段的可扩展延迟翻译来提供性能改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号