frequency and similarity aware partitioning for cloud storage based on space time utility...

13
TSINGHUA SCIENCE AND TECHNOLOGY ISSNll 1007-0214 ll 02/10 ll pp233-245 Volume 20, Number 3, June 2015 Frequency and Similarity-Aware Partitioning for Cloud Storage Based on Space-Time Utility Maximization Model Jianjiang Li, Jie Wu, and Zhanning Ma Abstract: With the rise of various cloud services, the problem of redundant data is more prominent in the cloud storage systems. How to assign a set of documents to a distributed file system, which can not only reduce storage space, but also ensure the access efficiency as much as possible, is an urgent problem which needs to be solved. Space-efficiency mainly uses data de-duplication technologies, while access-efficiency requires gathering the files with high similarity on a server. Based on the study of other data de-duplication technologies, especially the Similarity-Aware Partitioning (SAP) algorithm, this paper proposes the Frequency and Similarity-Aware Partitioning (FSAP) algorithm for cloud storage. The FSAP algorithm is a more reasonable data partitioning algorithm than the SAP algorithm. Meanwhile, this paper proposes the Space-Time Utility Maximization Model (STUMM), which is useful in balancing the relationship between space-efficiency and access-efficiency. Finally, this paper uses 100 web files downloaded from CNN for testing, and the results show that, relative to using the algorithms associated with the SAP algorithm (including the SAP-Space-Delta algorithm and the SAP-Space-Dedup algorithm), the FSAP algorithm based on STUMM reaches higher compression ratio and a more balanced distribution of data blocks. Key words: de-duplication; cloud storage; redundancy; frequency 1 Introduction Cloud computing has become an important part of the information technology industry in recent times. The number of cloud storage services is increasing with most vendors providing a variable amount of free storage, such as Facebook, Netflix, and Dropbox; therefore, data volumes are increasing at an unprecedented rate. These cloud-based services must solve two key issues: (1) They should be access- Jianjiang Li and Zhanning Ma are with the Department of Computer Science and Technology, University of Science and Technology Beijing, Beijing 100083, China. E-mail: [email protected]; [email protected]. Jie Wu is with the Department of Computer and Information Sciences, Temple University, Philadelphia, PA 19122, USA. E-mail: [email protected]. To whom correspondence should be addressed. Manuscript received: 2015-03-16; accepted: 2015-05-04 efficient to read or write the files in the system in terms of minimizing network accesses. This is a crucial factor with increasing network congestion in the underlying data center networks [1, 2] . (2) They should be space- efficient to manage the high volumes of data [3] . The continuous integration of global information leads to more and more high-value data. The exponential growth of high-value data leads to many challenges to the enterprise’s IT departments. The explosive data growth tends to consume a large amount of storage space of the data center, then forces the enterprises to buy more storage facilities to upgrade. At the same time, mass data is also a considerable challenge for network transmission. Faced with this phenomenon, we must consider other ways to resolve the anguish brought by mass data. From the economic perspective, people always want to spend the least money to store the most data. Even if the amount of data has not been large in www.redpel.com +917620593389 www.redpel.com +917620593389

Upload: redpel-dot-com

Post on 12-Apr-2017

151 views

Category:

Education


0 download

TRANSCRIPT

Page 1: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

TSINGHUA SCIENCE AND TECHNOLOGYISSNll1007-0214ll02/10llpp233-245Volume 20, Number 3, June 2015

Frequency and Similarity-Aware Partitioning for Cloud Storage Basedon Space-Time Utility Maximization Model

Jianjiang Li, Jie Wu, and Zhanning Ma�

Abstract: With the rise of various cloud services, the problem of redundant data is more prominent in the cloud

storage systems. How to assign a set of documents to a distributed file system, which can not only reduce storage

space, but also ensure the access efficiency as much as possible, is an urgent problem which needs to be solved.

Space-efficiency mainly uses data de-duplication technologies, while access-efficiency requires gathering the files

with high similarity on a server. Based on the study of other data de-duplication technologies, especially the

Similarity-Aware Partitioning (SAP) algorithm, this paper proposes the Frequency and Similarity-Aware Partitioning

(FSAP) algorithm for cloud storage. The FSAP algorithm is a more reasonable data partitioning algorithm than the

SAP algorithm. Meanwhile, this paper proposes the Space-Time Utility Maximization Model (STUMM), which is

useful in balancing the relationship between space-efficiency and access-efficiency. Finally, this paper uses 100

web files downloaded from CNN for testing, and the results show that, relative to using the algorithms associated

with the SAP algorithm (including the SAP-Space-Delta algorithm and the SAP-Space-Dedup algorithm), the

FSAP algorithm based on STUMM reaches higher compression ratio and a more balanced distribution of data

blocks.

Key words: de-duplication; cloud storage; redundancy; frequency

1 Introduction

Cloud computing has become an important part ofthe information technology industry in recent times.The number of cloud storage services is increasingwith most vendors providing a variable amountof free storage, such as Facebook, Netflix, andDropbox; therefore, data volumes are increasing at anunprecedented rate. These cloud-based services mustsolve two key issues: (1) They should be access-

� Jianjiang Li and Zhanning Ma are with the Department ofComputer Science and Technology, University of Scienceand Technology Beijing, Beijing 100083, China. E-mail:[email protected]; [email protected].� Jie Wu is with the Department of Computer and Information

Sciences, Temple University, Philadelphia, PA 19122, USA.E-mail: [email protected].�To whom correspondence should be addressed.

Manuscript received: 2015-03-16; accepted: 2015-05-04

efficient to read or write the files in the system in termsof minimizing network accesses. This is a crucial factorwith increasing network congestion in the underlyingdata center networks[1, 2]. (2) They should be space-efficient to manage the high volumes of data[3].

The continuous integration of global informationleads to more and more high-value data. Theexponential growth of high-value data leads to manychallenges to the enterprise’s IT departments. Theexplosive data growth tends to consume a large amountof storage space of the data center, then forces theenterprises to buy more storage facilities to upgrade.At the same time, mass data is also a considerablechallenge for network transmission. Faced with thisphenomenon, we must consider other ways to resolvethe anguish brought by mass data.

From the economic perspective, people always wantto spend the least money to store the most data.Even if the amount of data has not been large in

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 2: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

234 Tsinghua Science and Technology, June 2015, 20(3): 233-245

the past, researchers have already been committed todiscovering efficient compression storage. Therefore,how to reduce the amount of data transfer to completethe file synchronization is an issue which can bringtremendous value.

Space efficiency is often achieved by the processof de-duplication[2–6], which splits all the files inthe system into blocks, and maintains only a uniquecopy of each block. De-duplication can improvethe efficiency of storage, and reduce the networktraffic. Typically, most de-duplication solutions havefocused on reducing the space-overhead within a singleserver[7–9]. Under the environment in which the globaldata’s quantity grows sharply, at the requests of businessapplications and the requirements of legal compliance,de-duplication technology has become the companies’first choice. De-duplication technology changes thedata protection method by reducing the amount ofstored data, which also improves the economics of diskbackup. This has also gradually been recognized as thenext generation backup technology in industry, which isthe essential technology in data centers.

At present, de-duplication technology has beenconcerned by the researchers and industry, and has hadvery fruitful research results. At the same time, manyalgorithms have been designed to detect and deleteduplicate data in storage space or network transmission.Using data de-duplication technology to develop a safe,stable, and efficient cloud storage system has veryimportant practical significance, both in terms of savingstorage space and saving network bandwidth, evensaving energy. The development of the de-duplicationtechnique also enlarges the applying scope. Therefore,it not only can be used in backup and archiving, but alsocan cover the other network and desktop applications.

Reference [10] shows a Similarity-Aware Partitioning(SAP) approach that is both access-efficient and space-efficient. But there are many inaccuracies in delta-encoding. For example, there are three non-empty filesA1, A2, and A3. A1 D B1B2B3B4B5, A2 D B6B7,and A3 D B1B2B3B6B7 (where Bi is a block).Based on the delta-encoding of the SAP algorithm,ıfA1; A2g D fB6; B7g , ıfA1; A3g D fB6; B7g , sinceıfA1; A2g D ıfA1; A3g D fB6; B7g , then judging thesimilarity between A1 and A3 is equal to the similaritybetween A1 and A2 . This method is lopsided becauseit only considers the differences between two files,without considering the size of the file itself. Moreover,the complexity of the SAP algorithm is very high;

the amount of data on each server might have a bigdifference.

In this paper, we present an efficient Frequencyand Similarity-Aware Partitioning (FSAP) algorithmfor cloud storage based on a Space-Time UtilityMaximization Model (STUMM), which is morereasonable than the SAP algorithm. When calculatingthe similarity between two files, the FSAP algorithm notonly considers the size of the redundant data betweentwo files, but also considers the size of the two files.The FSAP algorithm also presents how to remove theredundancy among the different servers. In differentdegrees of further removal of redundancy (by settingdifferent thresholds), the space used to store files isdifferent, and the time to access files is also different.How to balance the relationship between space andaccess time to get maximum benefit is a big question. Inorder to solve this problem, we proposed a space-timeutility maximization model.

In Sections 2 and 3, we will describe the FSAPalgorithm in detail, and show the reasons why theFSAP algorithm is more reasonable than the SAPalgorithm. In Section 4, we will present the Space-Time Utility Maximization Model. The experimentalresults are presented in Section 5. We conducted twoexperiments: one for comparing the compression ratioand accessing efficiency between the FSAP algorithmand the SAP algorithm, and one for the reasonabilityof the Space-Time Utility Maximization Model. Finally,we will describe the related work and draw conclusions.

2 Data Partitioning Based on the SimilarityCoefficient of Files

In this section, we define the similarity coefficientbetween two files, compare it with the delta-encodingprinciple in the SAP algorithm[10], and present theprocess of data partitioning.

2.1 The definition of similarity coefficient betweentwo files

Definition 1 The similarity coefficient between anytwo non-empty files Ai and Aj (where, i ¤ j ) is

Sim.Ai ; Aj / D .num.Ai \ Aj /=num.Ai //�

.num.Ai \ Aj /=num.Aj // D

.num.Ai \ Aj //2=.num.Ai / � num.Aj //:

Ai \ Aj refers to the blocks contained by both fileAi and file Aj . num.Ai \Aj / represents the number ofblocks contained by both file Ai and file Aj . Therefore,

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 3: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 235

num.Ai \Aj / > 0; num.Ai / and num.Aj / indicate thenumber of blocks respectively contained by two non-empty files Ai and Aj , therefore, num.Ai / > 1 andnum.Ai / > 1.

Property 1 Sim.Ai ; Aj / 2 Œ0; 1�:

Generally, 0 6 num.Ai \ Aj / 6 num.Ai / and 0 6num.Ai \ Aj / 6 num.Aj /. So, 0 6 Sim.Ai ; Aj / 6 1.

If file Ai and file Aj are completely different, whichmeans that the blocks contained by file Ai and file Aj

are completely different, then num.Ai \ Aj / D 0. So,Sim.Ai ; Aj / D 0.

If file Ai and file Aj are identical, which means thatthe blocks contained by file Ai and file Aj are identical,without regard to the order of blocks, then num.Ai \

Aj / D num.Ai / D num.Aj /. So, Sim.Ai ; Aj / D 1.When num.Ai \ Aj / is fixed, Sim.Ai ; Aj / will

decline as the increase of num.Ai / or num.Aj /. Inparticular, when num.Ai /� num.Ai \ Aj / ornum.Aj /� num.Ai \ Aj /, Sim.Ai ; Aj / is close to 0.

Definition 2 Especially, if one of Ai and Aj

(where, i ¤ j ) is null (i.e., it does not contain anyblock), then the similarity coefficient between Ai andAj is defined as 0. If Ai and Aj (where, i ¤ j ) areboth empty (i.e., both Ai and Aj do not contain anyblock), then the similarity coefficient between Ai andAj is defined as 1.

Property 2 Sim.Ai ; Aj / D Sim.Aj ; Ai /:

If file Ai and file Aj are two non-empty files,according to Definition 1, then

Sim.Ai ; Aj / D .num.Ai \ Aj //2=.num.Ai / �

num.Aj //;

Sim.Aj ; Ai / D .num.Aj \ Ai //2=.num.Aj / �

num.Ai // D .num.Ai \ Aj //2=.num.Ai / � num.Aj //:

Therefore, Sim.Ai ; Aj / D Sim.Aj ; Ai /.If one of Ai and Aj is null, according to Definition 2,

then Sim.Ai ; Aj / D Sim.Aj ; Ai / D 0.If Ai and Aj are both empty, according to Definition

2, then Sim.Ai ; Aj / D Sim.Aj ; Ai / D 1.In conclusion, Sim.Ai ; Aj / D Sim.Aj ; Ai /.

According to Property 2, after Sim.Ai ; Aj / iscalculated, we can know the value of Sim.Aj ; Ai /, andhence cut the computational effort by half.

The similarity coefficient of two non-empty filesproposed in this paper is different from the similarityof two non-empty files based on delta-encoding inSAP algorithm[10]. For example, according to therepresentation method of blocks, three non-empty filesA1, A2, and A3 are expressed as A1 D B1B2B3B4B5,A2 D B6B7, and A3 D B1B2B3B6B7. According to

the delta-encoding in the SAP algorithm,ıfA1; A2g D fB6; B7g and ıfA1; A3g D fB6; B7g:

That is to say, ıfA1; A2g D ıfA1; A3g D fB6; B7g;determining the similarity of A1 and A3 is equivalent tothe similarity of A1 and A2. However, the definition ofthe similarity coefficient between two non-empty filesproposed by this paper shows

Sim.A1; A2/ D .num.A1 \ A2//2=.num.A1/ �

num.A2// D 0;

Sim.A1; A3/ D .num.A1 \ A3//2=.num.A1/ �

num.A3// D 32=.5 � 5/ D 0:36:

That is to say, A1 and A3 are more similar than A1

andA2. From the actual situation ofA1,A2, andA3,A1

and A2 do not have common blocks. A1 and A3 havethree common blocks: B1, B2, and B3. It also showsthat the similarity degree between A1 and A3 is muchhigher than the similarity degree between A1 and A2.Therefore, the similarity coefficient between two non-empty files proposed by this paper is more reasonablethan the delta-encoding principle in the SAP algorithm.

Table 1 shows ten files A1, A2, � � � , A10 accordingto the representation method of blocks, and Table 2shows the similarity coefficients between every two filesin Table 1, calculated according to the definition ofthe similarity coefficient proposed by this paper. Table2 can be regarded as a symmetry matrix, where allelements on the primary diagonal are 1.

Table 1 Files A1, A2, � � � , A10 according to the representationmethod of blocks (where, Bi is a file block).

A1 D B1B2B7 A6 D B3B5B6

A2 D B3B7 A7 D B3B4B6

A3 D B2B4B1B8 A8 D B4B1B7

A4 D B5B3B5B8 A9 D B5B3B7B8

A5 D B6B8 A10 D B4B7B6

Table 2 The similarity coefficients between any two files ofTable 1.

A1 A2 A3 A4 A5 A6 A7 A8 A9 A10

A1 1 0.17 0.33 0 0 0 0 0.44 0.08 0.11A2 0.17 1 0 0.13 0 0.17 0.17 0.17 0.5 0.17A3 0.33 0 1 0.06 0.13 0 0.08 0.33 0.06 0.08A4 0 0.13 0.06 1 0.13 0.67 0.08 0 0.75 0A5 0 0 0.13 0.13 1 0.17 0.17 0 0.13 0.17A6 0 0.17 0 0.67 0.17 1 0.44 0 0.33 0.11A7 0 0.17 0.08 0.08 0.17 0.44 1 0.11 0.08 0.44A8 0.44 0.17 0.33 0 0 0 0.11 1 0.08 0.44A9 0.08 0.5 0.06 0.75 0.13 0.33 0.08 0.08 1 0.08A10 0.11 0.17 0.08 0 0.17 0.11 0.44 0.44 0.88 1

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 4: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

236 Tsinghua Science and Technology, June 2015, 20(3): 233-245

2.2 The process of data partitioning based onsimilarity coefficients

When the total size of all files and the capacity ofeach server are known, we demonstrate how to evenlypartition these files among servers according to thesimilarity coefficients of files. The files with highsimilarity coefficients are likely to be partitioned tothe same server, which minimizes storage space whenkeeping a high access efficiency. Next, we will goover an example to demonstrate the process of datapartitioning based on similarity coefficients.

(1) According to the total size of all files and thecapacity of each server, we can calculate the numberof servers needed, which is the maximum number ofthe required servers (in general, after the redundancyis removed, the space occupied by these files is farless than the amount of space occupied by these filesstored separately). These files can be placed accordingto the calculated number of servers. After completionof placing these files, if the amount of data placed onthe server is less than the capacity of the server, we canfine-tune the data. For example, we can fine-tune thedata on the server which has the smallest total similaritycoefficient (or compression) to the other server. Here,we might as well suppose that the number of serversis 2 (when the number of server is 1, all of the data isplaced on the same server).

(2) According to the attraction (bigger similaritycoefficient) and repulsion (smaller similaritycoefficient) principle, we can choose pairs of fileswhich will be placed on different servers. At first,we find two pairs of files with the largest similaritycoefficient from Table 2, and respectively place them onserver1 and server2. We choose A4 and A9 placed onserver1, because the similarity coefficient of A4 and A9

is 0.75, which is the biggest similarity coefficient. Thenwe choose the pair, which has the biggest similaritycoefficient, from the rest of the pairs of files. Thesimilarity coefficients of A10-A7, A10-A8, A6-A7, andA1-A8 are all 0.44. As can be seen, among these fourpairs of files, A7, A8, and A10 all appear twice, but A1

and A6 both appear only once. Therefore, we shouldselect two files from A7, A8, A10 stored on server2.The similarity coefficients between A7, A8, A10, andA9 on the other server (i.e., server1) are identical, sowe might as well choose A8 and A10 placed on server2.This process is shown in Fig. 1.A4A9 D fB3B5B7B8g (Now the size of data stored

Fig. 1 The first step of data partitioning.

on server1 is 4 blocks.) A8A10 D fB1B4B6B7g (Nowthe size of data stored on server2 is 4 blocks.)

(3) From Table 3, we find file A2 which has thebiggest similarity coefficient with the current blocksstored on server1 (the similarity coefficient is 0.5),so it will be placed on server1. From Table 4, wefind out files A1 and A7 have the biggest similaritycoefficient with the blocks currently stored on server2(the similarity coefficients are all 0.33). The similaritycoefficients between files A1, A7 and the blockcurrently stored on server1 are all 0.08. According to theattraction and repulsion principle, here we can chooseA1 (of course, we can also choose A7). This process isshown in Fig. 2.A4A9A2 D fB3B5B7B8g (Now the size of data

stored on server1 is 4 blocks.) A8A10A1 D

fB1B2B4B6B7g (Now the size of data stored on server2is 5 blocks.)

(4) From Table 5, we find file A3 which has thebiggest similarity coefficient with the current blocksstored on server2 (the similarity coefficient is 0.45), soit will be placed on server2. At the same time, fromTable 6, we find file A6 which has the biggest similaritycoefficient with the block currently stored on server1(the similarity coefficient is 0.33), so it will be placedon server1. This process is shown in Fig. 3.A4A9A2A6 D fB3B5B6B7B8g (Now the size of

data stored on server1 is 5 blocks.) A8A10A1A3 D

Table 3 Similarity coefficients between A4A9 on server1 andfiles not placed on any server.

A1 A2 A3 A5 A6 A7

A4A9 0.08 0.5 0.06 0.13 0.33 0.08

Table 4 Similarity coefficients between A8A10 on server2and files not placed on any server.

A1 A2 A3 A5 A6 A7

A8A10 0.33 0.13 0.25 0.13 0.08 0.33

Fig. 2 The second step of data partitioning.

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 5: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 237

Table 5 Similarity coefficients between A8A10A1 on server2and files not placed on any server.

A3 A5 A6 A7

A8A10A1 0.45 0.1 0.07 0.27

Table 6 Similarity coefficients between A4A9A2 on server1and files not placed on any server (after file A2 is stored onserver1, blocks on server1 does not change, so the similaritycoefficients do not change).

A3 A5 A6 A7

A4A9A2 0.06 0.13 0.33 0.08

Fig. 3 The third step of data partitioning.

fB1B2B4B6B7B8g (Now the size of data stored onserver2 is 6 blocks.)

(5) From Table 7, we find file A5, which has thebiggest similarity coefficient with the blocks currentlystored on server1 (the similarity coefficient is 0.4), so itwill be placed on server1.A4A9A2A6A5 D fB3B5B6B7B8g (At present, the

size of the data stored on server1 is still 5 blocks).As can be seen from Tables 7 and 8, now the

similarity coefficient between A7 and the blockscurrently stored on server1 is larger than the similaritycoefficient between A7 and the blocks currently storedon server2, and the current number of blocks storedon server1 is less than the number of blocks storedon server2, so in order to balance the distributionof blocks on servers and follow the attraction andrepulsion principle, A7 will be also placed on server1.This process is shown in Fig. 4.A4A9A2A6A5A7 D fB3B4B5B6B7B8g (Now

Table 7 Similarity coefficients between A4A9A2A6 onserver1 and files not placed on any server.

A5 A7

A4A9A2A6 0.4 0.27

Table 8 Similarity coefficients between A8A10A1A3 onserver2 and files not placed on any server.

A5 A7

A8A10A1A3 0.33 0.22

Fig. 4 The fourth step of data partitioning.

the size of data stored on server1 is 6 blocks.)A8A10A1A3 D fB1B2B4B6B7B8g (Now the size ofdata stored on server1 is 6 blocks.)

Now, all of files have been placed on the appropriateservers based on the similarity of files.

2.3 Redundancy removal based on the occurrencefrequency of blocks

From the process of data partitioning described above,the redundant blocks on server1 and server2 are B4,B6, B7, and B8. If the capacity of server1 and server2is greater than the size of blocks needed to be placedon them, then the redundant blocks do not need to beremoved. Otherwise, we need to decide the degree ofredundancy removal based on the capacity of server1and server2. The most extreme case is: blocks on allservers cannot be redundant. In the above example,among files placed on server1, the occurrence frequencyof B4, B6, B7, and B8 is 1, 3, 2, 3, respectively; amongfiles placed on server2, the occurrence frequency of B4,B6, B7, and B8 is 3, 1, 3, 1, respectively. To minimizethe amount of access to the servers which store blocks,we can remove B4 and B7 stored on the server1, andremove B6 and B8 stored on server2. Eventually, theblocks placed on server1 areB3B5B6B8, and the blocksplaced on server2 are B1B2B4B7. To obtain A2, A7 orA9 stored on server1 is required to access server2 once(the total number of accesses is 3). Similarly, to obtainA3 or A10 stored on server2, it is required to access toserver1 once (the total number of times is 2). Therefore,redundancy removal based on the frequency of blocksis able to realize the data storage with smaller storagespace and higher access efficiency.

3 The FSAP Algorithm

In Section 2, we have explained the process of theFSAP algorithm through an example, including how

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 6: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

238 Tsinghua Science and Technology, June 2015, 20(3): 233-245

to calculate the similarity coefficients between eachset of two files, how to balance the distribution ofblocks on servers based on the similarity of files, andhow to remove redundancy based on the occurrencefrequency of blocks. The FSAP algorithm can realizethe data storage with smaller storage space and higheraccess efficiency. The key step of the FSAP algorithm(as shown in Algorithm 1) will be introduced in detailbelow.

Algorithm 1 The FSAP AlgorithmInput:

n: the total number of files which will be distributed onserversset of files A D fA1; A2; � � � ; Ang

C : server capacityS.A/: the total size of AK: the maximum number of the required serversd : remaining server capacityN.A/: the number of files in Anum.Ai /: the number of blocks in file Ai

q: block sizeOutput:

the result of data partitioning

1: Initialize: the files in each server: D ∅; K = S.A/=C ; d= C

2: find K pairs files with a similarity coefficient that is thelargest in the set Sim.Ai ; Aj /, 1 6 i < j 6 n

3: for all 1 6 m 6 K do4: add Aim

and Ajminto Dm: Dm Dm C Aim

C Ajm

5: remove Aimand Ajm

from A: A A � Aim� Ajm

;N.A/ D N.A/ �K

6: L1:7: construct a K �N.A/ matrix M8: for all 1 6 x 6 K do9: for all 1 6 y 6 N.A/ do

10: Mxy D Sim.Dx ; Ay/

11: while M is not empty do12: find the largest element Mx1y1

in the M13: if C > num.Dx1

[ Ay1/ � q then

14: remove x1-th row y1-th column from M15: Dx1

Dx1C Ay1

16: A A � Ay1

17: else18: remove x1-th row from M19: if A is not empty then20: goto L121: according to the occurrence frequency, remove redundant

blocks from Dm, 1 6 m 6 K

22: remove the redundant blocks among different servers basedon STUMM

23: D D fD1;D2; � � � ;DKg

24: return D

� Line 1: The result of data partitioning is initialized,that is D ∅. At the same time, the remainingserver capacity is initialized by d D C , andthe maximum number of the required servers iscalculated by K D S.A/=C .� Line 2: The similarity coefficients between

each two files (Sim.Ai ; Aj /, 1 6 i < j 6 n) arecalculated according to the definition proposed inthis paper, then K pairs of files, whose similaritycoefficients are among the top K, are found. TheK pairs of files are denoted by <Aim , Ajm

>,where 1 6 m 6 n.� Lines 3-5: The K pairs of files are placed on K

servers separately, that is, each server stores a pairof files, denoted by Dm. Files Aim and Ajm

areadded intoDm, then they are deleted fromA. AfterK pairs of files are processed, the number of filesin A will be reduced correspondingly.� Line 6: L1 is a label of the goto statement.� Lines 7-10: A K �N.A/ matrix M is constructed,

denoted by M D Sim.Dx; Ay/ (1 6 x 6 k, 1 6y 6 N.A/). Each row of M corresponds to thesimilarity coefficients between the files placed oneach server and the files not placed on any server.Each element of M is calculated by Sim.Dx; Ay/,where 1 6 x 6 K and 1 6 y 6 N.A/.� Lines 11-18: This is a circulate process when M is

not empty. Only the largest elementMx1y1in M is

found in each loop; in other words, file Ay1is the

most suitable file which should be placed on thex1-th server. Before these operations are executed,this algorithm needs to judge whether there areenough remaining spaces. On the other hand, inorder to balance the amount of files on each server,only one file is placed on each server in each loop.If C > num.Dx1

[ Ay1/ � q, then x1-th row y1-

th column will be removed from M. At the sametime, file Ay1

and Ajmare added into Dx1

, then itis deleted from A. If C 6 num.Dx1

[ Ay1/ � q,

then x1-th row will be removed from M.� Lines 19-20: If the set ofA is not empty, lines 7-18

will be executed repeatedly.� Line 21: If the set of A has become empty,

redundant blocks will be removed from Dm

(where, 1 6 m 6 K) based on the occurrencefrequency of blocks. The redundant blocks withlower occurrence frequencies will be removedfrom the severs which store them.� Line 22: According to the capacity of each server

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 7: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 239

and the access time tolerated by users, the degreeof redundancy removal will be determined. InSection 4, we will introduce the STUMM modelin detail.� Lines 23-24: the result of data partitioning is

outputted.The complexity of the FSAP algorithm proposed in

this paper is lower than the SAP algorithm: In the firststep, the similarity coefficients between any two non-empty files have a large amount of calculations, andthe memory usage is also larger. However, since theadjacency matrix in the FSAP algorithm is a symmetricmatrix, the amount of calculation and the occupancy ofthe memory reduces by half.

The amount of files on each server is nearly sameafter data partitioning: In order to balance the amount offiles stored on each server, only one file, which has thehighest similarity, is placed on each server in each loop.Therefore, the amount of files on each server is nearlythe same. This data partitioning is more reasonable thanthat of the SAP algorithm.

4 Space-Time Utility Maximization Model

Definition 3 Space-time utility P is a weightedaverage of data space usage and data access time.Namely,

P D ˛=S C .1 � ˛/=T; 0 6 ˛ 6 1;

where S is data space usage, which means the totalamount of data space usage on servers. T is data accesstime, which indicates the average of data access timeto obtain all data files stored on servers. The SAPalgorithm is just a special case (without considerationof removing the redundant data blocks among differentstorage servers) of the FSAP algorithm proposed in thispaper. In addition, the SAP algorithm only considersdata space usage S . However, the FSAP algorithm notonly considers data space usage S , but also considersdata access time T .

When ˛ D 1, space-time utility P is completelycontrolled by data space usage S . On the other hand,when ˛ D 0, space-time utility P is completelycontrolled by data access time T . This paper takes intoaccount the effect of data space usage and data accesstime on space-time utility. Therefore, the range of ˛should be 0 6 ˛ 6 1. If 0 6 ˛ < 0:5, then data accesstime can do more to influence space-time utility thandata space usage. If 0:5 < ˛ 6 1, then data space usagecan do more to influence space-time utility than data

access time. If ˛ D 0:5, then data space usage does thesame to influence space-time utility as data access time.This paper will discuss how to solve the problem, underwhich circumstances the value of time utility P is themaximum.

Generally, If all data can be obtained locally, dataaccess time T is a constant T0 which does not changewith data space usage S . Otherwise, when data spaceusage S becomes greater, data access time T willdecrease due to the lower probability of data accessacross servers; when data space usage S becomessmaller, data access time T will increase because of thehigher probability of data access across servers. Dataaccess time T is a function of data space usage S ; here,suppose that the function is T D ˇ=S , where, ˇ is aconstant greater than 0.

(1) If all data can be obtained locally, then Pmax D

maxf˛=S C .1 � ˛/=T0g. At this time, as data spaceusage S increases, data access time T is still constantT0 and will not decrease. Therefore, if all data canbe obtained locally, Pmax D ˛=minS C .1 � ˛/=T0.That is space-time utility P can reach the maximumvalue when redundant blocks within each server arecompletely eliminated. As we know, this is certain afterdata partitioning using the FSAP algorithm is proposedin this paper. Therefore, Pmax D ˛=S C .1 � ˛/=T0.

(2) If some data may be obtained remotely, thenPmax D maxf˛=SC.1�˛/=.ˇ=S/g D maxf˛=SC.1�˛/=ˇ �Sg and @P=@S D �˛=S2C.1�˛/=ˇ. Therefore,when @P=@S D 0, namely�˛=S2C.1�˛/=ˇ D 0, thatis, when S D

p.˛ � ˇ/=.1 � ˛/, the space-time utility

P can reach the extreme value ˛=p.˛ � ˇ/=.1 � ˛/ C

.1 � ˛/=ˇ �p.˛ � ˇ/=.1 � ˛/. S can be determined

according to different values of ˛ and ˇ.In order to obtain the maximum value of P , both S

and T need to be converted into a unified dimension.The greater the S , the more storage cost paid by the datacenter; the larger the T , the higher the latency which canbe tolerated by the data center.

When the size of files assigned to each server ismuch less than the capacity of each server Sfile, theredundancy may be allowed, i.e., different servers canhave the same blocks. When the size of files on eachserver is close to or exceeds the capacity of eachserver Sfile, the redundant blocks among servers mustbe removed according to the occurrence frequency ofblocks.

While redundancy removal technology improves theutilization of space, it reduces the reliability of data.

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 8: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

240 Tsinghua Science and Technology, June 2015, 20(3): 233-245

Researchers have done experiments to analyze howmuch redundancy removal contributes to reducing thereliability of data[11].

The results of these experiments show that theavailability of files is far less than the availability of datablocks. For example, when the equipment failure ratereaches 6%, 99.5% of data blocks are available, but only96% of files are accessible. The reason for the resultsis that multiple files share common blocks, so there is adependency between different files.

The influence of redundancy removal on thereliability of data has long attracted the attention ofscholars. The simplest way to improve the reliabilityof data is to create a copy using the Mirror technology,but it will greatly reduce the utilization rate of storagespace, like Ceph[12] and GPFS[13]. All of these systemsrarely take into account the utilization rate of storagespace and the reliability of data.

For cloud storage systems, in order to increasethe availability of data, we should also keep someredundant data. How much redundant data should bekept is a difficult problem. Using the Space-Timeutility maximization model, according to the capacityof each server and the access time tolerated by users,we can determine the degree of redundancy removaland make a more reasonable operation to obtain themaximum value of P . In Section 5, we will verify itby experiments.

5 Performance Evaluation and Analysis

In this section, firstly, compare the compression ratioand the distribution of data blocks between SAP andFSAP by experiment; secondly, verify the reasonabilityof the Space-Time utility maximization model.

5.1 Comparison of the utilization rate of storagespace of the FSAP algorithm with that of theSAP algorithm

In this part, when the number of servers is maxf2,the number of files/20g (Each server is configured asthe following: dual-core CPU and 4 GB memory),we compare the utilization rate of storage spaceof the FSAP1 algorithm, the FSAP2 algorithm,the SAP-Space-Delta algorithm, and the SAP-Space-Dedup algorithm. The FSAP1 algorithm is normalcompression, and the FSAP2 algorithm is deepcompression whose threshold value is 2; namely, if thefrequency of the occurrence of redundant data blocks onone server is smaller than 2, these redundant blocks on

this server will be removed. For example, the frequencyof the occurrence of data block B1 on server1 is 1, andthe frequency of the occurrence of B1 on server2 is 2.In the case of the FSAP1 algorithm, the redundant datablocks of B1 on both server1 and server2 will not beremoved, but in the case of the FSAP2 algorithm, B1 onserver1 will be removed, while B1 on server2 will beretained.

We perform basic experiments on a set of files: 100random web files downloaded from CNN. In the courseof these experiments, we keep the block-size constant at48, change the number of files in increments of 5, andmeasure the compression ratio achieved by the FSAP1

algorithm, the FSAP2 algorithm, the SAP-Space-Delta algorithm, and the SAP-Space-Dedup algorithmrelative to the total size of the uncompressed files. Thetest results are shown in Table 9 and Fig. 5. On the otherhand, 100 files are stored on 5 servers and Table 10shows the distribution of data blocks using the FSAP1

algorithm, the FSAP2 algorithm, the SAP-Space-Deltaalgorithm, and the SAP-Space-Dedup algorithm.

For files on CNN, the FSAP1 algorithm achievesa compression ratio of 23%, the FSAP2 algorithmachieves a compression ratio of 29.35%, the SAP-Space-Delta algorithm achieves a compression ratio of

Table 9 The test results of the FSAP algorithm and the SAPalgorithm.

The numberof files

Compression ratio (%)SAP-

Space-DeltaSAP-

Space-DedupFSAP1 FSAP2

5 11.14 11.74 12.60 19.9710 11.42 14.79 14.12 20.6015 11.80 18.04 18.15 24.0520 12.49 19.77 19.59 24.9525 12.00 20.77 20.30 25.5230 11.26 21.04 20.71 25.7335 12.81 23.36 22.56 27.3240 13.71 24.96 25.25 29.6745 11.60 24.51 25.46 29.7450 11.80 25.53 26.38 30.6255 12.72 26.61 26.73 30.8960 12.52 23.45 23.37 28.7165 12.38 24.18 23.94 29.2770 12.16 24.51 24.38 29.5675 11.93 24.76 24.72 29.7280 11.83 22.71 23.32 29.1885 11.63 23.02 23.50 29.3090 11.52 23.38 23.84 29.6495 11.86 24.2 24.26 29.98100 11.65 22.44 23.00 29.35

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 9: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 241

Fig. 5 The compression ratio of files from CNN usingdifferent algorithms.

Table 10 The distribution of data blocks using differentalgorithms.

server1 server2 server3 server4 server5SAP-Space-Delta 7536 9988 11 848 14 362 20 721

SAP-Space-Dedup 6843 9055 10 521 12 560 17 606FSAP1 10 582 11 267 12 635 11 122 10 567FSAP2 9196 10 012 11 138 9565 9180

11.65%, and the SAP-Space-Dedup algorithm achievesa compression ratio of 22.44%. The compression ratioof CNN filed using the FSAP1 algorithm is obviouslyhigher than the compression ratio of CNN filed usingthe SAP-Space-Delta algorithm, and it is slightly higherthan the SAP-Space-Dedup algorithm. Because eachof the test web files downloaded from CNN areapproximately the same size, the curve of the FSAP1

algorithm and the SAP-Space-Dedup algorithm are verysimilar. Even in this case, the compression ratio of theFSAP1 algorithm is still higher than the compressionratio of the SAP-Space-Dedup algorithm.

From the test results, we can find that (1) in thecase of a fixed number of servers, as the numberof files increases, the compression ratio graduallyincreases; (2) if the number of servers increases, thenthe compression ratio will decrease. As an analogy tothe actual situation, if the capacity of each server issufficient, then more files will be stored, and a highercompression ratio will be achieved. Conversely, if thecapacity of a server is insufficient, then we need toincrease the number of servers, and the compressionratio will decline.

From Table 10, we can find that the distributionof data blocks using the SAP-Space-Delta algorithmand the SAP-Space-Dedup algorithm are seriouslyimbalanced. The number of data blocks on different

servers are obviously different. But the distributionof data blocks using the FSAP1 algorithm and theFSAP2 algorithm is basically the same. The test resultsshow that the FSAP algorithm achieves an even greaterdistribution of data blocks than the SAP algorithm.

In general, the test results show that the FSAPalgorithm proposed in this paper is better than theSAP algorithm. Further, the compression ratio usingthe FSAP2 algorithm is far greater than that usingthe FSAP1 algorithm, which also shows that theFSAP algorithm has a significant effect on redundancyremoval of data blocks among different servers.

5.2 The test of Space-Time utility maximizationmodel

In order to test Space-Time utility maximization model,we perform experiments on a set of 100 random webfiles downloaded from CNN using the FSAP algorithm.The number of servers is fixed at 5, and each server isconfigured as dual-core CPU and 4 GB memory. Wedefine a variable — threshold. For the redundant datablocks, if the occurrence frequency on one server issmaller than the threshold value, these redundant blockson this server will be removed (under the premise ofretain at least one). If the frequency on each server islarger than the threshold, these redundant blocks willnot be removed, while the frequency is equal to thethreshold, we only retain the redundant blocks on asingle server. As the maximum occurrence frequencyis 7 or 8[10], we take the maximum threshold value is10. We observe how S and T change over the thresholdvalue of redundancy removal. The changes of S and Twill directly effect P .

The test procedure is as follows:(1) To obtain the change curves of S over the

threshold value of redundancy removal. S is the totalnumber of data blocks retained on 5 storage servers.

(2) To obtain the change curves of T over thethreshold value of redundancy removal. T is expressedby the results of numerical simulation. Its calculationstandard is as follows: if one file is completely storedin one server, its access time is the number of datablocks in this file. After deep redundancy removal ofdata blocks, when one data block is removed, T will beincreased by 1 because of the need to obtain the datablock from other servers.

(3) The dimensions of S and T are different; we needto unify them. It is an important step in the test of theSpace-Time utility maximization model. The method

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 10: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

242 Tsinghua Science and Technology, June 2015, 20(3): 233-245

adopted in this paper is as follows:S 0 D .1=S �min.1=S//=.max.1=S/ �min.1=S//;T 0 D .1=T �min.1=T //=.max.1=T /�min.1=T //:(4) To obtain the change curves of P over the

threshold value of redundancy removal. Now,P D ˛ � S 0 C .1 � ˛/ � T 0; 0 6 ˛ 6 1:

The test results are shown in Table 11 and Figs. 6and 7. From the test results, we can find that as thethreshold value of redundancy removal increases, dataspace usage S decreases; meanwhile, data access timeT increases. When the threshold value of redundancyremoval is 2, 3, and 4, data space usage S significantlyreduces and data access time T obviously increases. Atlast, data space usage and data access time tend to be aconstant values, which means that at this time there areno redundant data blocks among servers. At the sametime, Table 12 and Fig. 8 show that P can always obtainthe extremum value in the range of threshold. When˛ D 0, P gets the maximum value at the point of theminimum threshold; when ˛ D 1, P gets the maximumvalue at the point of the maximum threshold. When˛ changes, P gets the maximum value at different

thresholds. The curves of P show that space-time utilitymaximization model is reasonable. Additionally, usingthe Space-Time maximization model proposed in thispaper, by observing the change of P with the thresholdvalue of redundancy removal, we can achieve the bestbalance between data space usage and data access time.

6 Related Work

Data de-duplication, as a compression method, hasbeen widely used in most backup systems to improvebandwidth and space efficiency[14]. This technology canremove redundant data in the data stream. The finer thegrain size, the more the redundant data removed, whilemore computing resources are used at the same time.De-duplication technology has already received a lot ofattention in the industry.

Single Instance Storage (SIS) on the MicrosoftWindows Server2000 achieves the same file stored onlyonce, while through forging links to provide a copy ofthe file system[15]. This method is simple, easy to use,and is more suitable for the collection of small files, butit cannot detect the same data block within a different

Table 11 The test results of S and T (N is the number of servers.).

ThresholdN D 2 N D 3 N D 4 N D 5

S T S T S T S T1 49 996 49 996 52 910 52 910 54 832 54 832 56 059 56 0592 45 783 54 209 46 177 59 643 46 353 63 311 46 419 65 6993 45 023 55 729 45 118 61 761 45 108 65 801 45 101 68 3354 44 710 56 668 44 751 62 862 44 751 66 872 44 727 69 4575 44 577 57 200 44 599 63 470 44573 67 584 44 548 70 1736 44 504 57 565 44 517 63 880 44499 67 954 44 481 70 5087 44 461 57 823 44 469 64 168 44448 68 260 44 434 70 7908 44 429 58 047 44 438 64 385 44420 68 456 44 406 70 9869 44 413 58 175 44 414 64 577 44403 68 592 44 384 711 6210 44 399 58 301 44 401 64 694 44 383 68 772 44 366 71 324

Fig. 6 The relationship between S and the threshold valueof redundancy removal. N is the number of servers.

Fig. 7 The relationship between T and the threshold valueof redundancy removal. N is the number of servers.

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 11: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 243

Table 12 The value of P in different ˛̨̨ .Thre- Pshold ˛=0 ˛=0.2 ˛=0.5 ˛=0.6 ˛=0.8 ˛=1

1 1 0.8 0.5 0.4 0.2 02 0.314 42 0.409 13 0.551 19 0.598 55 0.693 25 0.787 963 0.160 63 0.312 88 0.541 25 0.617 37 0.769 62 0.921 874 0.098 71 0.271 23 0.530 01 0.616 27 0.788 79 0.961 315 0.060 24 0.244 27 0.520 32 0.612 34 0.796 38 0.980 416 0.042 50 0.231 52 0.515 05 0.609 56 0.798 58 0.987 617 0.027 70 0.220 70 0.510 18 0.606 68 0.799 67 0.992 668 0.017 49 0.213 13 0.506 58 0.604 40 0.800 04 0.995 689 0.008 36 0.206 30 0.503 20 0.602 18 0.800 12 0.998 0610 0 0.2 0.5 0.6 0.8 1

Fig. 8 The relationship between P and the threshold valueof redundancy removal.

file. In order to delete duplicate data in the datablock level, the fixed-size block method was firstdesigned[16]. In this method, each file is divided intonon-overlapping blocks of fixed size, and each datablock has a hash value. Bell laboratory’s Venti[8] filestorage system saves approximately 30% of the storagespace using this method. IBM applied this technologyto the online TotalStorage SAN File System[16]. Inaddition, the virtual machine migration also usesthe fixed-length block divided method to improvethe migration performance and reduce the networkcost[17]. The technology of using this algorithm is verysimple. The efficiency is also very good, but the changesto a file location are very sensitive, and the file headerinserting a new byte can cause all data blocks to forminto a new block.

Low-Bandwidth File System (LBFS)[18] using theRabin fingerprint[19, 20] instead of the fixed size dividedthe file into a variable block size. This method iscalled block divided method based on content, it makesthe storage system no longer sensitive to positionmodifications. Network buffer system based on value,

DDFS of Data Domain, the P2P file system of Pasta,and backup system of Pastiche also used this technologyto identify and eliminate the redundant data. However,this algorithm has uncertainty on the definition of datablock boundary, and has a higher overhead. Besidesthese algorithms, some techniques, such as delta-encoding[21], Bloomfilter[22], and sparse index[23], werealso used to enhance the performance of de-duplicationtechnology.

7 Conclusions

In this paper, we discuss the problem of redundant datain the cloud storage systems and study the main de-duplication technologies. The main goal of this paper ishow to use less space to store more data files, and at thesame time, not obviously affect access efficiency. Basedon the study of other data de-duplication technologies(especially the SAP algorithm), we first propose a morereasonable data partitioning algorithm — FSAP, thenpresent the Space-Time maximization model; finally,we compare the utilization rate of storage space ofthe FSAP algorithm with that of the SAP algorithm,and test the Space-Time utility maximization modelthrough experiments on a set of 100 random web filesdownloaded from CNN.

The FSAP algorithm overcomes the shortcomings ofthe SAP algorithm that calculates the similarity betweentwo files that are not reasonable, and also reducesthe computational complexity and the space occupied.The experimental results on files from CNN show thatrelative to using the SAP-Space-Delta algorithm andthe SAP-Space-Dedup algorithm, the compression ratiousing the FSAP algorithm respectively increased by17.7% and 6.91%. At the same time, the test resultsshow that the FSAP algorithm achieves a more evendistribution of data blocks than the SAP algorithm.These facts show that the FSAP algorithm proposedin this paper is much more efficient than the SAPalgorithm.

Data de-duplication does not mean that there is nodata redundancy; proper data redundancy can reduceaccess time and increase the reliability of data access.Using the space-time maximization model proposedin this paper, by observing the change of P with thethreshold value of redundancy removal, we can achievethe best balance between data space usage and dataaccess time.

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 12: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

244 Tsinghua Science and Technology, June 2015, 20(3): 233-245

Acknowledgements

This work was supported by the National High-TechResearch and Development (863) Program of China (No.2015AA01A303).

References

[1] T. Benson, A. Akella, and D. A. Maltz, Network trafficcharacteristics of data centers in the wild, in Proceedingsof the 10th ACM SIGCOMM Conference on InternetMeasurement, ACM, 2010, pp. 267–280.

[2] D. T. Meyer and W. J. Bolosky, A study of practicaldeduplication, ACM Transactions on Storage (TOS), vol.7, no. 4, p. 14, 2012.

[3] A. T. Clements, I. Ahmad, M. Vilayannur, and J. Li,Decentralized deduplication in san cluster file systems, inUSENIX Annual Technical Conference, 2009, pp. 101–114.

[4] U. Manber, Finding similar files in a large file system, inUsenix Winter 1994 Technical Conference, 1994, pp. 1–10.

[5] B. S. Baker, On finding duplication and near-duplicationin large software systems, in Reverse Engineering, 1995,Proceedings of 2nd Working Conference on, IEEE, 1995,pp. 86–95.

[6] G. Forman, K. Eshghi, and S. Chiocchetti, Finding similarfiles in large document repositories, in Proceedings ofthe Eleventh ACM SIGKDD International Conference onKnowledge Discovery in Data Mining, ACM, 2005, pp.394–400.

[7] J. R. Douceur, A. Adya, W. J. Bolosky, P. Simon, andM. Theimer, Reclaiming space from duplicate files in aserverless distributed file system, in Distributed ComputingSystems, 2002. Proceedings. 22nd InternationalConference on, IEEE, 2002, pp. 617–624.

[8] S. Quinlan and S. Dorward, Venti: A new approach toarchival storage, FAST, vol. 2, pp. 89–101, 2002.

[9] B. Zhu, K. Li, and R. H. Patterson, Avoiding the diskbottleneck in the data domain deduplication file system,FAST, vol. 8, pp. 1–14, 2008.

[10] B. Balasubramanian, T. Lan, and M. Chiang, Sap:Similarity aware partitioning for efficient cloud storage, inInfocom 2014 Proceedings IEEE, 2014, pp. 592–600.

[11] D. Bhagwat, K. Pollack, D. D. Long, T. Schwarz, E.L. Miller, and J. F. Paris, Providing high reliabilityin a minimum redundancy archival storage system, inModeling, Analysis, and Simulation of Computer andTelecommunication Systems, 2006. MASCOTS 2006. 14thIEEE International Symposium on, IEEE, 2006, pp. 413–421.

[12] T. J. Schwarz, Q. Xin, E. L. Miller, D. D. Long, A.Hospodor, and S. Ng, Disk scrubbing in large archivalstorage systems, in Modeling, Analysis, and Simulationof Computer and Telecommunications Systems, 2004.(MASCOTS 2004). Proceedings. The IEEE ComputerSocietys 12th Annual International Symposium on, IEEE,2004, pp. 409–418.

[13] S. A. Weil, S. A. Brandt, E. L. Miller, D. D. Long, and C.Maltzahn, Ceph: A scalable, high-performance distributedfile system, in Proceedings of the 7th Symposium onOperating Systems Design and Implementation, USENIXAssociation, 2006, pp. 307–320.

[14] R. Zhu, L.-H. Qin, J.-L. Zhou, and H. Zheng, Usingmultithreads to hide deduplication i/o latency with lowsynchronization overhead, Journal of Central SouthUniversity, vol. 20, pp. 1582–1591, 2013.

[15] W. J. Bolosky, S. Corbin, D. Goebel, and J. R. Douceur,Single instance storage in windows 2000, in Proceedingsof the 4th USENIX Windows Systems Symposium, Seattle,USA, 2000, pp. 13–24.

[16] B. Hong, D. Plantenberg, D. D. Long, and M. Sivan-Zimet, Duplicate data elimination in a san file system,in Proceedings of the 21st IEEE / 12th NASA GoddardConference on Mass Storage Systems and Technologies,2004, pp. 301–314.

[17] C. P. Sapuntzakis, R. Chandra, B. Pfaff, J. Chow, M.S. Lam, and M. Rosenblum, Optimizing the migrationof virtual computers, ACM SIGOPS Operating SystemsReview, vol. 36, no. SI, pp. 377–390, 2002.

[18] A. Muthitacharoen, B. Chen, and D. Mazieres, A lowbandwidth network file system, ACM SIGOPS OperatingSystems Review, vol. 35, no. 5, pp. 174–187, 2001.

[19] M. O. Rabin, Fingerprinting by random polynomials,Center for Research in Computing Technology, HarvardUniversity, Report TR-15-81, 1981.

[20] A. Z. Broder, Some applications of rabins fingerprintingmethod, in Sequences II. Springer, 1993, pp. 143–152.

[21] F. Douglis and A. Iyengar, Application-specificdeltaencoding via resemblance detection. in USENIXAnnual Technical Conference, General Track, 2003, pp.113–126.

[22] B. H. Bloom, Space/time trade-offs in hash coding withallowable errors, Communications of the ACM, vol. 13, no.7, pp. 422–426, 1970.

[23] M. Lillibridge, K. Eshghi, D. Bhagwat, V. Deolalikar, G.Trezis, and P. Camble, Sparse indexing: Large scale, inlinededuplication using sampling and locality, in 7th USENIXConference on File and Storage Technologies, 2009, pp.111–123.

www.redpel.com +917620593389

www.redpel.com +917620593389

Page 13: Frequency and similarity aware partitioning for cloud storage based on space time utility maximization model

Jianjiang Li et al.: Frequency and Similarity-Aware Partitioning for Cloud Storage � � � 245

Jie Wu is the chair and a Laura H.Carnell Professor in the Department ofComputer and Information Sciences atTemple University. Prior to joining TempleUniversity, USA, he was a programdirector at the National Science Foundationand a distinguished professor at FloridaAtlantic University. He received his PhD

degree from Florida Atlantic University in 1989. His currentresearch interests include mobile computing and wirelessnetworks, routing protocols, cloud and green computing,network trust and security, and social network applications.Dr. Wu regularly published in scholarly journals, conferenceproceedings, and books. He serves on several editorial boards,including IEEE Transactions on Computers, IEEE Transactionson Service Computing, and Journal of Parallel and DistributedComputing. Dr. Wu was general co-chair/chair for IEEE MASS2006 and IEEE IPDPS 2008 and program co-chair for IEEEINFOCOM 2011. Currently, he is serving as general chairfor IEEE ICDCS 2013 and ACM MobiHoc 2014, and programchair for CCF CNCC 2013. He was an IEEE Computer SocietyDistinguished Visitor, ACM Distinguished Speaker, and chair

for the IEEE Technical Committee on Distributed Processing(TCDP). Dr. Wu is a CCF Distinguished Speaker and a Fellowof the IEEE. He is the recipient of the 2011 China ComputerFederation (CCF) Overseas Outstanding Achievement Award.

Jianjiang Li is currently an associateprofessor at University of Science andTechnology Beijing, China. He receivedhis PhD degree in computer science fromTsinghua University in 2005. He was avisiting scholar at Temple University fromJan. 2014 to Jan. 2015. His current researchinterests include parallel computing, cloud

computing, and parallel compilation.

Zhanning Ma is currently a masterstudent in University of Science andTechnology Beijing for his degree. Hisresearch interests include distributedstorage technology and data duplicateremoval. He received his bachelor’s degreein 2010 from Taishan Medical University.

www.redpel.com +917620593389

www.redpel.com +917620593389