期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Scalable high performance de-duplication backup via hash join
1
作者 Tian-ming YANG Dan FENG Zhong-ying NIU Ya-ping WAN 《Journal of Zhejiang University-Science C(Computers and Electronics)》 SCIE EI 2010年第5期315-327,共13页
Apart from high space efficiency,other demanding requirements for enterprise de-duplication backup are high performance,high scalability,and availability for large-scale distributed environments.The main challenge is ... Apart from high space efficiency,other demanding requirements for enterprise de-duplication backup are high performance,high scalability,and availability for large-scale distributed environments.The main challenge is reducing the significant disk input/output(I/O) overhead as a result of constantly accessing the disk to identify duplicate chunks.Existing inline de-duplication approaches mainly rely on duplicate locality to avoid disk bottleneck,thus suffering from degradation under poor duplicate locality workload.This paper presents Chunkfarm,a post-processing de-duplication backup system designed to improve capacity,throughput,and scalability for de-duplication.Chunkfarm performs de-duplication backup using the hash join algorithm,which turns the notoriously random and small disk I/Os of fingerprint lookups and updates into large sequential disk I/Os,hence achieving high write throughput not influenced by workload locality.More importantly,by decentralizing fingerprint lookup and update,Chunkfarm supports a cluster of servers to perform de-duplication backup in parallel;it hence is conducive to distributed implementation and thus applicable to large-scale and distributed storage systems. 展开更多
关键词 Backup system De-duplication POST-PROCESSING fingerprint lookup Scalability
原文传递
基于MD5和布鲁姆过滤器的重复数据删除算法
2
作者 车玉婧 胡波 费向东 《计算机光盘软件与应用》 2011年第16期56-57,共2页
针对目前重复数据处理技术的低效性和不可靠性,本文提出了一种基于MD5算法和布鲁姆过滤器的重复数据删除算法。新算法采用两级布鲁姆过滤器并有效结合MDS算法的方式,在发挥布鲁姆过滤器空间效率的同时汲取了MD5算法的可靠性,使得文... 针对目前重复数据处理技术的低效性和不可靠性,本文提出了一种基于MD5算法和布鲁姆过滤器的重复数据删除算法。新算法采用两级布鲁姆过滤器并有效结合MDS算法的方式,在发挥布鲁姆过滤器空间效率的同时汲取了MD5算法的可靠性,使得文件级别和数据块级别的重复数据删除策略交替工作。测试分析表明,新算法性能稳定并且实现了高效且可靠的重复数据删除功能。 展开更多
关键词 重复数据删除技术 MD5算法 布鲁姆过滤器 指纹查询
下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部