首页> 外文会议>IEEE/ACM International Symposium on Microarchitecture >Cache bursts: A new approach for eliminating dead blocks and increasing cache efficiency
【24h】

Cache bursts: A new approach for eliminating dead blocks and increasing cache efficiency

机译:缓存突发:消除死区段的新方法以及增加高速缓存效率

获取原文

摘要

Data caches in general-purpose microprocessors often contain mostly dead blocks and are thus used inefficiently. To improve cache efficiency, dead blocks should be identified and evicted early. Prior schemes predict the death of a block immediately after it is accessed; however, these schemes yield lower prediction accuracy and coverage. Instead, we find that predicting the death of a block when it just moves out of the MRU position gives the best tradeoff between timeliness and prediction accuracy/coverage. Furthermore, the individual reference history of a block in the L1 cache can be irregular because of data/control dependence. This paper proposes a new class of dead-block predictors that predict dead blocks based on bursts of accesses to a cache block. A cache burst begins when a block becomes MRU and ends when it becomes non-MRU. Cache bursts are more predictable than individual references because they hide the irregularity of individual references. When used at the L1 cache, the best burst-based predictor can identify 96% of the dead blocks with a 96% accuracy. With the improved dead-block predictors, we evaluate three ways to increase cache efficiency by eliminating dead blocks early: replacement optimization, bypassing, and prefetching. The most effective approach, prefetching into dead blocks, increases the average L1 efficiency from 8% to 17% and the L2 efficiency from 17% to 27%. This increased cache efficiency translates into higher overall performance: prefetching into dead blocks outperforms the same prefetch scheme without dead-block prediction by 12% at the L1 and by 13% at the L2.
机译:通用微处理器中的数据高速缓存通常含有大多数死区块,因此使用效率低下。为了提高缓存效率,应提前识别和驱逐死块。现有计划在访问后立即预测块的死亡;然而,这些方案产生了更低的预测准确性和覆盖范围。相反,我们发现预测块的死亡,当它刚刚移出MRU位置时,可以在及时性和预测准确度/覆盖之间提供最佳权衡。此外,由于数据/控制依赖性,L1高速缓存中的块的个体参考历史可以是不规则的。本文提出了一种新的DEAT块预测因子,其基于对缓存块的访问突发预测死块。当块变为MRU时,高速缓存突发开始,并在变为非MRU时结束。缓存突发比单个参考更可预测,因为它们隐藏了个别参考的不规则性。在L1缓存中使用时,最佳的基于突发的预测器可以识别96%的死块,精度为96%。通过改进的死区阻滞预测因子,我们通过消除早期消除死区块来评估三种方式来提高缓存效率:替换优化,绕过和预取。最有效的方法预先达到死区块,将平均L1效率从8%增加到17%,L2效率从17%到27%。这种增加的高速缓存效率转化为更高的整体性能:预取块在L1处的L1和L2处的13%达到相同的预取方案,在L1和13%的情况下达到相同的预取方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号