首页> 外文会议>Annual IEEE/ACM International Symposium on Microarchitecture >In-Network Coherence Filtering: Snoopy Coherence without Broadcasts
【24h】

In-Network Coherence Filtering: Snoopy Coherence without Broadcasts

机译:网络内连贯过滤:没有广播的史努比连贯

获取原文

摘要

With transistor miniaturization leading to an abundance of on-chip resources and uniprocessor designs providing diminishing returns, the industry has moved beyond single-core microprocessors and embraced the many-core wave. Scalable cache coherence protocol implementations are necessary to allow fast sharing of data among various cores and drive the many-core revolution forward. Snoopy coherence protocols, if realizable, have the desirable property of having low storage overhead and not adding indirection delay to cache-to-cache accesses. There are various proposals, like Token Coherence (TokenB), Uncorq, Intel QPI, INSO and Timestamp Snooping, that tackle the ordering of requests in snoopy protocols and make them realizable on unordered networks. However, snoopy protocols still have the broadcast overhead because each coherence request goes to all cores in the system. This has substantial network bandwidth and power implications. In this work, we propose embedding small in-network coherence niters inside on-chip routers that dynamically track sharing patterns among various cores. This sharing information is used to filter away redundant snoop requests that are traveling towards unshared cores. Filtering these useless messages saves network bandwidth and power and makes snoopy protocols on many-core systems truly scalable. Our in-network coherence filters are able to reduce the total number of snoops in the system on an average by 41.9%, thereby reducing total network traffic by 25.4% on 16-processor chip multiprocessor (CMP) systems running parallel applications. For 64-processor CMP systems, our filtering technique on an average achieves 46.5% reduction in total number of snoops that ends up reducing the total network traffic by 27.3%, on an average.
机译:利用晶体管小型化导致芯片上的片上资源和单处理器设计提供减少回报,业界已超越单核微处理器,并拥有多核波。可扩展的高速缓存相干协议协议实现是必要的,以允许在各种核心之间快速分享数据并向前进的多核革命驱动。 Snoopy Cherence协议,如果可实现,则具有具有低存储开销的理想性能,并且不会向缓存到缓存访问增加间接延迟。有各种建议,如令牌相干性(tokenb),Uncorq,英特尔QPI,INSO和时间戳窥探,可以解决史努比协议中的请求的排序,并使它们可实现无序网络。然而,史努比协议仍然具有广播开销,因为每个相干请求都进入系统中的所有核心。这具有实质性的网络带宽和功率影响。在这项工作中,我们建议在片上路由器内部嵌入小型网络中的连贯介质,以动态跟踪各种核心之间的共享模式。此共享信息用于过滤向未共享核心行驶的冗余窥探请求。过滤这些无用的消息可节省网络带宽和电源,并在许多核心系统上进行窥探协议,真正可扩展。我们网络的相干滤波器能够平均减少系统中的循环总数,平均为41.9%,从而在16处理器芯片多处理器(CMP)系统上减少了25.4%的网络流量。对于64处理器CMP系统,我们平均的过滤技术实现了46.5%的总数减少,最终将网络流量降低27.3%,平均值。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号