A logistic based mathematical model to optimize duplicate elimination ratio in content defined chunking based big data storage system

8Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

Deduplication is an efficient data reduction technique, and it is used to mitigate the problem of huge data volume in big data storage systems. Content defined chunking (CDC) is the most widely used algorithm in deduplication systems. The expected chunk size is an important parameter of CDC, and it influences the duplicate elimination ratio (DER) significantly. We collected two realistic datasets to perform an experiment. The experimental results showed that the current approach of setting the expected chunk size to 4 KB or 8 KB empirically cannot optimize DER. Therefore, we present a logistic based mathematical model to reveal the hidden relationship between the expected chunk size and the DER. This model provides a theoretical basis for optimizing DER by setting the expected chunk size reasonably. We used the collected datasets to verify this model. The experimental results showed that the R2 values, which describe the goodness of fit, are above 0.9, validating the correctness of this mathematic model. Based on the DER model, we discussed how to make DER close to the optimum by setting the expected chunk size reasonably.

Cite

CITATION STYLE

APA

Wang, L., Dong, X., Zhang, X., Guo, F., Wang, Y., & Gong, W. (2016). A logistic based mathematical model to optimize duplicate elimination ratio in content defined chunking based big data storage system. Symmetry, 8(7). https://doi.org/10.3390/sym8070069

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free