首页> 外文会议>IEEE Symposium on Reliable Distributed Systems >Improving Reliability of Deduplication-Based Storage Systems with Per-File Parity
【24h】

Improving Reliability of Deduplication-Based Storage Systems with Per-File Parity

机译:通过每个文件奇偶校验提高基于重复数据删除的存储系统的可靠性

获取原文
获取外文期刊封面目录资料

摘要

The reliability issue in deduplication-based storage systems has not received adequate attention. Existing approaches introduce data redundancy after files have been deduplicated, either by replication on critical data chunks, i.e., chunks with high reference count, or RAID schemes on unique data chunks, which means that these schemes are based on individual unique data chunks rather than individual files. This can leave individual files vulnerable to losses, particularly in the presence of transient and unrecoverable data chunk errors such as latent sector errors. To address this file reliability issue, this paper proposes a Per-File Parity (short for PFP) scheme to improve the reliability of deduplication-based storage systems. PFP computes the XOR parity within parity groups of data chunks of each file after the chunking process but before the data chunks are deduplicated. Therefore, PFP can provide parity redundancy protection for all files by intra-file recovery and a higher-level protection for data chunks with high reference counts by inter-file recovery. Our reliability analysis and extensive data-driven, failure-injection based experiments conducted on a prototype implementation of PFP show that PFP significantly outperforms the existing redundancy solutions, DTR and RCR, in system reliability, tolerating multiple data chunk failures and guaranteeing file availability upon multiple data chunk failures. Moreover, a performance evaluation shows that PFP only incurs an average of 5.7% performance degradation to the deduplication-based storage system.
机译:基于重复数据删除的存储系统中的可靠性问题尚未引起足够的重视。现有的方法在对文件进行了重复数据删除之后引入了数据冗余,方法是在关键数据块(即具有高引用计数的块)上进行复制,或者在唯一数据块上进行RAID方案,这意味着这些方案基于单独的唯一数据块而不是单独的数据块文件。这可能会使单个文件容易丢失,特别是在存在暂时性和不可恢复的数据块错误(例如潜在扇区错误)的情况下。为了解决此文件可靠性问题,本文提出了一种基于文件奇偶校验(PFP)的方案,以提高基于重复数据删除的存储系统的可靠性。在分块处理之后但在对数据块进行重复数据删除之前,PFP会计算每个文件的数据块的奇偶校验组内的XOR奇偶校验。因此,PFP可以通过文件内恢复为所有文件提供奇偶校验冗余保护,并通过文件间恢复为具有高引用计数的数据块提供更高级别的保护。我们对PFP的原型实施进行的可靠性分析和大量基于数据驱动的基于故障注入的实验表明,PFP在系统可靠性方面明显优于现有的冗余解决方案DTR和RCR,可以承受多个数据块故障,并保证多个文件块的可用性数据块故障。此外,性能评估表明,PFP仅会对基于重复数据删除的存储系统造成平均5.7%的性能下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号