首页> 外文期刊>Journal of supercomputing >A novel dynamic network data replication scheme based on historical access record and proactive deletion
【24h】

A novel dynamic network data replication scheme based on historical access record and proactive deletion

机译:一种基于历史访问记录和主动删除的动态网络数据复制新方案

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Data replication is becoming a popular technology in many fields such as cloud storage, Data grids and P2P systems. By replicating files to other serversodes, we can reduce network traffic and file access time and increase data availability to react natural and man-made disasters. However, it does not mean that more replicas can always have a better system performance. Replicas indeed decrease read access time and provide better fault-tolerance, but if we consider write access, maintaining a large number of replications will result in a huge update overhead. Hence, a tradeoff between read access time and write updating cost is needed. File popularity is an important factor in making decisions about data replication. To avoid data access fluctuations, historical file popularity can be used for selecting really popular files. In this research, a dynamic data replication strategy is proposed based on two ideas. The first one employs historical access records which are useful for picking up a file to replicate. The second one is a proactive deletion method, which is applied to control the replica number to reach an optimal balance between the read access time and the write update overhead. A unified cost model is used as a means to measure and compare the performance of our data replication algorithm and other existing algorithms. The results indicate that our new algorithm performs much better than those algorithms.
机译:数据复制正在成为许多领域的流行技术,例如云存储,数据网格和P2P系统。通过将文件复制到其他服务器/节点,我们可以减少网络流量和文件访问时间,并提高数据可用性以应对自然和人为灾难。但是,这并不意味着更多的副本始终可以具有更好的系统性能。副本确实可以减少读取访问时间并提供更好的容错能力,但是如果考虑写入访问,则维护大量复制将导致巨大的更新开销。因此,需要在读取访问时间和写入更新成本之间进行权衡。文件流行度是决定数据复制的重要因素。为了避免数据访问波动,可以使用历史文件流行度来选择真正流行的文件。在这项研究中,基于两个思想提出了一种动态数据复制策略。第一个使用历史访问记录,这些记录对于选择要复制的文件很有用。第二种是主动删除方法,用于控制副本数以在读取访问时间和写入更新开销之间达到最佳平衡。统一成本模型用作衡量和比较我们的数据复制算法和其他现有算法性能的一种方式。结果表明,我们的新算法比那些算法具有更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号