1. background, motivation, and data · cross-geography scientific data transferring trends and...

1
Cross-Geography Scientific Data Transferring Trends and Behavior Zhengchun Liu 1 , Rajkumar Kettimuthu 1 , Ian Foster 1 and Nageswara S.V. Rao 2 ( 1 Argonne National Laboratory, Lemont, IL, USA; 2 Oak Ridge National Laboratory, Oak Ridge, TN, USA) 27 th HPDC Summary Ø Wide area data transfers play an important role in science applications but rely on expensive infrastructure that often delivers disappointing performance in practice. Ø We present a systematic examination of a large set of data transfer log data to characterize transfer characteristics, including the nature of the datasets transferred, throughput achieved, user behavior, and resource usage. Ø Our analysis yields new insights that can help design better data transfer tools, optimize networking and edge resources used for transfers, and improve the performance and experience for end users. Ø Our analysis shows that (i) most of the datasets as well as individual files transferred are very small; (ii) data corruption is not negligible for large data transfers; and (iii) the data transfer nodes utilization is low. 3. Transfer characteristics 1. Background, motivation, and data 2. Dataset characteristics 4. User behaviors 1 0 1 5 1 10 1 15 1 10 1 15 1 30 1 35 1 40 1 45 7ransfer dataset size (byte) 0 20 40 60 80 100 Cumulative % of total transfers 2014 2015 2016 2017 q Most of the datasets moved over the wide area are small. Specifically, the 50th, 75th, and 95th quartiles of dataset size are 6.3 MB, 221.5 MB, and 55.8~GB, respectively. Counterintuitively, the dataset size has decreased year by year from 2014 to 2017. q Majority of individual file size is less than 1MB. The result motivate the need for optimizations aimed at small file transfers. q Image files are the most common file type transferred, followed by raw text files. .dat are likely to be the format that user give casually. Scientific formats such as .h5(hierarchical data format) and .nc(NetCDF) are in the top 10. jpg png no-ext tif txt dat gz h5 nc vtk bin json out Pat tiff xPl log csv test hdf stats fasta fits fa cbf 0 1 2 3 4 5 6 7 8 9 3ercentage to total (%) 0 6 20 40 60 80 100 PercentDge of tiPe Dctive in 2017(%) 0 20 40 60 80 100 PercentDge of D71s (%) Cumulative distribution of idle time percentage; 80% of endpoints were active less than 6% of the time. 2 4 6 8 10 12 14 16 18 1uPber of endpoints accessed 55 60 65 70 75 80 85 90 95 100 Percentage of users 2014 2015 2016 2017 Cumulative distribution of the number of endpoints users have accessed. DTN utilization is surprisingly low. Since the DTN requirement is high for high- throughput DTNs, some good topics for research would be the use of these computing resource: (1)for other purposes; (2)for complex encoding to deal with data corruption and; (3)to compress data to reduce the network bandwidth consumption. q Slightly more than half of the users accessed two or fewer endpoints. q The degree distribution of the number of users per endpoint follows a power-law distribution, similar to other real-world social network graphs. 0 200 400 600 800 1000 1200 1400 1600 0 200 400 600 800 1000 # of endpoints 10 1 10 1 10 3 1umber of users 10 0 10 1 10 1 10 3 # of endpoints Top 100 Post heavily used endpoints 0 20 40 60 80 Percentage of tiPe active (%) Active means that there is at one transfer to/from the endpoint. The log-scaled plot shows that the distribution follows a power law. 1 0 1 5 1 10 1 15 1 10 1 15 1 30 1 35 1 40 )ile size (byte) 0 20 40 60 80 100 Cumulative probability (%) 2014 2015 2016 2017 Year fts_url_copy libglobus_ftp_client globusonline-fxp globus-url-copy gfal2-util Total PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles 2014 N/A N/A 111.23 746.59 39.81 1646.10 13.13 816.67 N/A N/A 176.24 3431.78 2015 48.09 77.29 103.21 841.96 52.89 2424.58 19.27 947.78 0.93 6.70 267.33 4435.13 2016 244.46 295.67 105.75 998.96 88.56 3600.78 14.76 850.76 10.03 74.05 466.91 5922.83 2017 342.12 550.57 40.11 885.65 113.45 3901.27 16.89 898.14 45.93 234.65 585.01 6671.79 Total 634.67 923.53 360.3 3,473.16 294.71 11,572.73 64.05 3,513.35 56.89 315.4 1,495.49 20,461.53 Petabytes and millions of files transferred via GridFTP using different clients. National International Total Year PBytes MFiles PBytes MFiles PBytes MFiles 2014 41.44 1,865 0.78 26.9 42.32 1,892 2015 53.45 2,763 2.55 94.3 56.39 2,873 2016 90.10 3,929 2.84 110.8 93.60 14,042 2017 109.16 4,162 3.23 94.3 113.50 4,264 Data transferred by Globus (i.e., globusonline-fxp) By using Globus GridFTP, about 20 billion files, totaling 1.8 Exabyte between any two of 63,166 unique endpoints were transferred from 2014 to 2017. On average more than 25,000 files are transferred per minute in 2017. q At least one checksum failure occurs per 1.26 TB. The integrity checking is needed though it causes extra load. q The failures are decreasing year by year. Overall, the service is becoming increasingly reliable. q Although some server-to-server transfers achieve high performance (dozens of Gbps), most transfer throughput is low. q There is no clear increasing trend in terms of transfer performance over time. 2014 2015 2016 2017 Year 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 AYerage check errors per 7B 2014 2015 2016 2017 Year 0 50 100 150 200 250 AYerage faults per 7B 1 0 1 5 1 10 1 15 1 10 1 15 1 30 1 35 1 40 7ransfer rate (bit/s) 0 20 40 60 80 100 Cumulative probability (%) 2014 2015 2016 2017 1 0 1 5 1 10 1 15 1 10 1 15 1 30 1 35 1 40 1 45 1 50 1 55 %ytes transferred 0 2 4 6 8 10 12 14 16 % of anual active users 2014 2015 2016 2017 Distribution of bytes transferred per user. q Of all the bytes transferred, 80% are by just 3% of all users; 10% of the users transferred 95% of the data. q The distribution of the number of users per endpoint follows a power-law distribution, similar to other real- world social network graphs. q Most users do not manually tune the transfer parameters. q Thus, transfer tools should be smart enough to choose the optimal parameters. A dataset consists of one or more files and zero or more directories. q Most of the datasets transferred by the Globus transfer service have only one file. And 17.6% of those datasets (or 11% of the total) have a file size > 100 MB, motivating the need for striping the single-file transfer over multiple servers. q The average file size of most datasets transferred is small (on the order of few megabytes). q Repeated transfers are not common, less than 7.7% of the datasets are transferred more than once. When they do occur, the datasets in question are distributed mostly from one (or a few) endpoints to multiple destinations. 5. Endpoint characteristics We believe our findings can help: q Resource providers to optimize the resources used for data transferring; q End users to organize datasets to maximize performance; q Researchers and tool developers to build new (or optimizing the existing) data transfer protocols and tools; Funding agencies to plan investments. Acknowledgements This material was supported in part by the U.S. Department of Energy, Office of Science, Advanced Scientific Computing Research, under Contract DE-AC02-06CH11357 and the DOE RAMSES project fund by Scientific Workflow Analysis program managed by Richard Carlson.

Upload: others

Post on 08-Oct-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 1. Background, motivation, and data · Cross-Geography Scientific Data Transferring Trends and Behavior Zhengchun Liu1, RajkumarKettimuthu1, Ian Foster1and Nageswara S.V. Rao2 (1Argonne

Cross-Geography Scientific Data Transferring Trends and BehaviorZhengchun Liu1, Rajkumar Kettimuthu1, Ian Foster1 and Nageswara S.V. Rao2

(1Argonne National Laboratory, Lemont, IL, USA; 2Oak Ridge National Laboratory, Oak Ridge, TN, USA) 27th HPDC

SummaryØ Wide area data transfers play an important role in science applications but rely on expensive infrastructure that often delivers disappointing performance in practice.

Ø We present a systematic examination of a large set of data transfer log data to characterize transfer characteristics, including the nature of the datasets transferred,throughput achieved, user behavior, and resource usage.

Ø Our analysis yields new insights that can help design better data transfer tools, optimize networking and edge resources used for transfers, and improve the performanceand experience for end users.

Ø Our analysis shows that (i) most of the datasets as well as individual files transferred are very small; (ii) data corruption is not negligible for large data transfers; and (iii)the data transfer nodes utilization is low.

3. Transfer characteristics

1. Background, motivation, and data

2. Dataset characteristics

4. User behaviors

10 15 110 115 110 115 130 135 140 145

7ransfer dataset size (byte)0

20

40

60

80

100

Cum

ulat

ive

% o

f tot

al tr

ansf

ers

20142015

20162017

q Most of the datasets moved over the wide area are small. Specifically, the 50th, 75th, and 95th quartiles of dataset size are 6.3 MB, 221.5 MB, and 55.8~GB, respectively. Counterintuitively, the dataset size has decreased year by year from 2014 to 2017.

q Majority of individual file size is less than 1MB. The result motivate the need for optimizations aimed at small file transfers.q Image files are the most common file type transferred, followed by raw text files. .dat are likely to be the format that user give

casually. Scientific formats such as .h5(hierarchical data format) and .nc(NetCDF) are in the top 10.

jpg

png

no-e

xt tif txt

dat gz h5 nc vtk

bin

json

out

Pat tiff

xPl

log

csv

test hdf

stat

sfa

sta

fits fa cbf0

123456789

3erc

enta

ge to

tota

l (%

)

0 6 20 40 60 80 100PercentDge of tiPe Dctive in 2017(%)

0

20

40

60

80

100

Perc

entD

ge o

f D71

s (%

)

Cumulative distribution of idle time percentage; 80%of endpoints were active less than 6% of the time.

2 4 6 8 10 12 14 16 181uPber of endpoints accessed

55

60

65

70

75

80

85

90

95

100

Perc

enta

ge o

f use

rs

2014 2015 2016 2017

Cumulative distribution of the number ofendpoints users have accessed.

DTN utilization is surprisingly low. Sincethe DTN requirement is high for high-throughput DTNs, some good topics forresearch would be the use of these computingresource:(1)for other purposes;(2)for complex encoding to deal with data

corruption and;(3)to compress data to reduce the network

bandwidth consumption.

q Slightly more than half of the users accessed two or fewer endpoints.q The degree distribution of the number of users per endpoint follows a power-law distribution, similar to

other real-world social network graphs.

0 200 400 600 800 1000 1200 1400 16000

200

400

600

800

1000

# o

f end

poin

ts

101 101 103

1umber of users100

101

101

103

# o

f end

poin

ts

Top 100 Post heavily used endpoints0

20

40

60

80

Perc

enta

ge o

f tiP

e ac

tive

(%)

Active means that there is at one transfer to/from the endpoint.

The log-scaled plot shows that the distribution follows a power law.

10 15 110 115 110 115 130 135 140

)ile size (byte)0

20

40

60

80

100

Cum

ulat

ive

prob

abili

ty (%

)

20142015

20162017

Cross-Geography Scientific Data Transferring Trends and Behavior HPDC ’18, June 11–15, 2018, Tempe, AZ, USA

Table 2: Petabytes and millions of �les transferred via GridFTP using di�erent tools over the past four years.

Year fts_url_copy libglobus_ftp_client globusonline-fxp globus-url-copy gfal2-util TotalPBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles

2014 N/A N/A 111.23 746.59 39.81 1646.10 13.13 816.67 N/A N/A 176.24 3431.782015 48.09 77.29 103.21 841.96 52.89 2424.58 19.27 947.78 0.93 6.70 267.33 4435.132016 244.46 295.67 105.75 998.96 88.56 3600.78 14.76 850.76 10.03 74.05 466.91 5922.832017 342.12 550.57 40.11 885.65 113.45 3901.27 16.89 898.14 45.93 234.65 585.01 6671.79Total 634.67 923.53 360.3 3,473.16 294.71 11,572.73 64.05 3,513.35 56.89 315.4 1,495.49 20,461.53

2.3 Limitations in GridFTP Usage LogsBecause of privacy considerations [28], the GridFTP toolkit reportsthe IP address only of the machine that runs it; in other words, logsfor the STOR command do not have the IP address of the sourceendpoint. Similarly there is no IP address of the destination endpointfor RETR logs. The total number of endpoints (unique IP address)in the past four years is 63,166. There are 20.5 billion STOR logstotaling 1.5 exabytes received and 19.4 billion RETR logs totaling1.8 exabyte transferred. We note that since GridFTP uses unreliableUDP to collect usage and since users can disable the collection,the STOR logs and RETR logs are di�erent. Considering the largenumber of logs even in a short time—on average there are morethan 25,000 STOR and RETR logs per minute in 2017—accuratelymatching a STOR log with a RETR log is almost impossible. On theother hand, Globus transfer (being a hosted service) logs have thisinformation and many other details about the transfers. Arguably,these logs still have some limitations; for example, they do not havethe size of the individual �les in a transfer. Nevertheless, these logsare much more comprehensive than the GridFTP logs.

2.4 Globus Transfer ServiceThe Globus transfer service is a cloud-hosted software-as-a-serviceimplementation of the logic required to orchestrate �le transfersbetween pairs of storage systems [3]. A transfer request speci�es,among other things, a source and destination; the �le(s) and/ordirectory(s) to be transferred; and (optionally) whether to performintegrity checking (enabled by default) and/or to encrypt the data(disabled by default). It provides automatic fault recovery and au-tomatic tuning of optimization parameters to achieve high perfor-mance. Globus can transfer data with either the GridFTP or HTTPprotocols; we focus here on GridFTP transfers, since HTTP supporthas been added only recently.

The Globus transfer service distinguishes between the two typesof GridFTP server installations: Globus Connect Personal (GCP),a lightweight single-user GridFTP server designed to be deployedon personal computers, and Globus Connect Server (GCS), a mul-tiuser GridFTP server designed to be deployed on high-performancestorage systems that may be accessed by many users concurrently.

Globus transfer logs recorded 4,813,091 transfers from 2014/01/01to 2018/01/01, totaling 13.1 billion �les and 305.8 PB. These trans-fers involved 41,900 unique endpoints and 71,800 unique source-to-destination pairs (edges), and 26,100 users. We used the MaxMind IPgeolocation service [25] to obtain approximate endpoint locations.Figure 2 shows the number in each city worldwide. Table 3 showsthe total bytes and �les transferred per year, both within a single

country (nationally) and between countries (internationally). Logsinclude the unique name of the source and destination endpoints,transfer start and end date and time, the user who submitted thetransfer, total bytes, number of �les and number of directories, andnumber of faults and �le integrity failures. The logs also have tun-able parameters. Therefore, the Globus logs are a good supplementto GridFTP logs in order to characterize wide area data transfer.

Table 3: Data transferred by Globus: petabytes and millionsof �les.

National International TotalYear PBytes MFiles PBytes MFiles PBytes MFiles2014 41.44 1,865 0.78 26.9 42.32 1,8922015 53.45 2,763 2.55 94.3 56.39 2,8732016 90.10 3,929 2.84 110.8 93.60 14,0422017 109.16 4,162 3.23 94.3 113.50 4,264

2.5 Analysis FrameworkFour years of raw GridFTP logs were stored in about 100,000 com-pressed �les in json format, for a total of 1.2 TB. We parsed andsaved these logs in MongoDB for our analysis. The raw Globustransfer service logs were saved in millions of tiny �les in jsonformat. Since Globus logs is much smaller than GridFTP logs, weparsed these tiny json �les and saved them as one �le by usingthe Python pickle module (it implements binary protocols for se-rializing and deserializing a Python object structure). In our anal-ysis, we used the Python pandas library [26] to load the Globustransfer logs. We performed all raw data analysis on a Cray Urika-GX platform (a high-performance big data analytics platform opti-mized for multiple work�ows), with the Apache Spark [37] cluster-computing framework. Anonymized sample data �les are availableat https://github.com/ramsesproject/wan-dts-log. The GridFTP logssoon will be publicly available for researchers for further analysisvia the data-sharing service of Globus.

3 DATASET CHARACTERISTICSUsers’ transfers consist of one or more �les. GridFTP clients use oneor more control channel sessions to the GridFTP server(s) (for third-party server-to-server transfers, clients establish control channelsessions with both the source and destination servers). The GridFTPserver handles each control channel session independently and thusdoes not what �les belong to the same transfer. GridFTP logs havestatistics for each individual �le, which could be a separate transferin itself or part of a bigger multi-�le or directory transfer. On the

Petabytes and millions of files transferred via GridFTP using different clients.

Cross-Geography Scientific Data Transferring Trends and Behavior HPDC ’18, June 11–15, 2018, Tempe, AZ, USA

Table 2: Petabytes and millions of �les transferred via GridFTP using di�erent tools over the past four years.

Year fts_url_copy libglobus_ftp_client globusonline-fxp globus-url-copy gfal2-util TotalPBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles PBytes MFiles

2014 N/A N/A 111.23 746.59 39.81 1646.10 13.13 816.67 N/A N/A 176.24 3431.782015 48.09 77.29 103.21 841.96 52.89 2424.58 19.27 947.78 0.93 6.70 267.33 4435.132016 244.46 295.67 105.75 998.96 88.56 3600.78 14.76 850.76 10.03 74.05 466.91 5922.832017 342.12 550.57 40.11 885.65 113.45 3901.27 16.89 898.14 45.93 234.65 585.01 6671.79Total 634.67 923.53 360.3 3,473.16 294.71 11,572.73 64.05 3,513.35 56.89 315.4 1,495.49 20,461.53

2.3 Limitations in GridFTP Usage LogsBecause of privacy considerations [28], the GridFTP toolkit reportsthe IP address only of the machine that runs it; in other words, logsfor the STOR command do not have the IP address of the sourceendpoint. Similarly there is no IP address of the destination endpointfor RETR logs. The total number of endpoints (unique IP address)in the past four years is 63,166. There are 20.5 billion STOR logstotaling 1.5 exabytes received and 19.4 billion RETR logs totaling1.8 exabyte transferred. We note that since GridFTP uses unreliableUDP to collect usage and since users can disable the collection,the STOR logs and RETR logs are di�erent. Considering the largenumber of logs even in a short time—on average there are morethan 25,000 STOR and RETR logs per minute in 2017—accuratelymatching a STOR log with a RETR log is almost impossible. On theother hand, Globus transfer (being a hosted service) logs have thisinformation and many other details about the transfers. Arguably,these logs still have some limitations; for example, they do not havethe size of the individual �les in a transfer. Nevertheless, these logsare much more comprehensive than the GridFTP logs.

2.4 Globus Transfer ServiceThe Globus transfer service is a cloud-hosted software-as-a-serviceimplementation of the logic required to orchestrate �le transfersbetween pairs of storage systems [3]. A transfer request speci�es,among other things, a source and destination; the �le(s) and/ordirectory(s) to be transferred; and (optionally) whether to performintegrity checking (enabled by default) and/or to encrypt the data(disabled by default). It provides automatic fault recovery and au-tomatic tuning of optimization parameters to achieve high perfor-mance. Globus can transfer data with either the GridFTP or HTTPprotocols; we focus here on GridFTP transfers, since HTTP supporthas been added only recently.

The Globus transfer service distinguishes between the two typesof GridFTP server installations: Globus Connect Personal (GCP),a lightweight single-user GridFTP server designed to be deployedon personal computers, and Globus Connect Server (GCS), a mul-tiuser GridFTP server designed to be deployed on high-performancestorage systems that may be accessed by many users concurrently.

Globus transfer logs recorded 4,813,091 transfers from 2014/01/01to 2018/01/01, totaling 13.1 billion �les and 305.8 PB. These trans-fers involved 41,900 unique endpoints and 71,800 unique source-to-destination pairs (edges), and 26,100 users. We used the MaxMind IPgeolocation service [25] to obtain approximate endpoint locations.Figure 2 shows the number in each city worldwide. Table 3 showsthe total bytes and �les transferred per year, both within a single

country (nationally) and between countries (internationally). Logsinclude the unique name of the source and destination endpoints,transfer start and end date and time, the user who submitted thetransfer, total bytes, number of �les and number of directories, andnumber of faults and �le integrity failures. The logs also have tun-able parameters. Therefore, the Globus logs are a good supplementto GridFTP logs in order to characterize wide area data transfer.

Table 3: Data transferred by Globus: petabytes and millionsof �les.

National International TotalYear PBytes MFiles PBytes MFiles PBytes MFiles2014 41.44 1,865 0.78 26.9 42.32 1,8922015 53.45 2,763 2.55 94.3 56.39 2,8732016 90.10 3,929 2.84 110.8 93.60 14,0422017 109.16 4,162 3.23 94.3 113.50 4,264

2.5 Analysis FrameworkFour years of raw GridFTP logs were stored in about 100,000 com-pressed �les in json format, for a total of 1.2 TB. We parsed andsaved these logs in MongoDB for our analysis. The raw Globustransfer service logs were saved in millions of tiny �les in jsonformat. Since Globus logs is much smaller than GridFTP logs, weparsed these tiny json �les and saved them as one �le by usingthe Python pickle module (it implements binary protocols for se-rializing and deserializing a Python object structure). In our anal-ysis, we used the Python pandas library [26] to load the Globustransfer logs. We performed all raw data analysis on a Cray Urika-GX platform (a high-performance big data analytics platform opti-mized for multiple work�ows), with the Apache Spark [37] cluster-computing framework. Anonymized sample data �les are availableat https://github.com/ramsesproject/wan-dts-log. The GridFTP logssoon will be publicly available for researchers for further analysisvia the data-sharing service of Globus.

3 DATASET CHARACTERISTICSUsers’ transfers consist of one or more �les. GridFTP clients use oneor more control channel sessions to the GridFTP server(s) (for third-party server-to-server transfers, clients establish control channelsessions with both the source and destination servers). The GridFTPserver handles each control channel session independently and thusdoes not what �les belong to the same transfer. GridFTP logs havestatistics for each individual �le, which could be a separate transferin itself or part of a bigger multi-�le or directory transfer. On the

Data transferred by Globus (i.e., globusonline-fxp)

By using Globus GridFTP, about 20 billion files, totaling 1.8 Exabyte between any two of 63,166 unique endpoints were transferred from 2014 to 2017. On average more than 25,000 files are transferred per minute in 2017.

q At least one checksum failure occurs per 1.26 TB. The integrity checking is needed though it causes extra load.

q The failures are decreasing year by year. Overall, the service is becoming increasingly reliable.

q Although some server-to-server transfers achieve high performance (dozens of Gbps), most transfer throughput is low.

q There is no clear increasing trend in terms of transfer performance over time.

2014 2015 2016 2017Year

0.0

0.2

0.4

0.6

0.8

1.0

1.2

1.4

1.6

AYer

age

chec

k er

rors

per

7B

2014 2015 2016 2017Year

0

50

100

150

200

250

AYer

age

faul

ts p

er 7

B

10 15 110 115 110 115 130 135 140

7ransfer rate (bit/s)0

20

40

60

80

100

Cum

ulat

ive

prob

abili

ty (%

) 20142015

20162017

10 15 110 115 110 115 130 135 140 145 150 155

%ytes transferred0

2

4

6

8

10

12

14

16

% o

f anu

al a

ctiv

e us

ers

20142015

20162017

Distribution of bytestransferred per user.

q Of all the bytes transferred, 80% are by just 3% of all users; 10% of the users transferred 95% of the data.

q The distribution of the number of users per endpoint follows a power-law distribution, similar to other real-world social network graphs.

q Most users do not manually tune the transfer parameters.

q Thus, transfer tools should be smart enough to choose the optimal parameters.

A dataset consists of one or more files and zero or more directories.

q Most of the datasets transferred by the Globus transfer service have only one file. And 17.6% of those datasets (or 11% of the total) have a file size > 100 MB, motivating the need for striping the single-file transfer over multiple servers.

q The average file size of most datasets transferred is small (on the order of few megabytes).

q Repeated transfers are not common, less than 7.7% of the datasets are transferred more than once. When they do occur, the datasets in question are distributed mostly from one (or a few) endpoints to multiple destinations.

5. Endpoint characteristics

We believe our findings can help: q Resource providers to optimize the resources used for data transferring; q End users to organize datasets to maximize performance; q Researchers and tool developers to build new (or optimizing the existing) data transfer protocols and tools; Funding agencies to plan investments.

AcknowledgementsThis material was supported in part by the U.S. Department of Energy, Office of Science, Advanced Scientific Computing Research,under Contract DE-AC02-06CH11357 and the DOE RAMSES project fund by Scientific Workflow Analysis program managed by RichardCarlson.