june 1, 2015 1 grenchmark : towards a generic framework for analyzing, testing, and comparing grids...
Post on 18-Dec-2015
218 views
TRANSCRIPT
April 18, 20231
GrenchMark: Towards a Generic Framework for Analyzing, Testing, and Comparing Grids
ASCI Conference 2006
A. Iosup, D.H.J. EpemaPDS Group, ST/EWI, TU Delft
April 18, 20232
Outline
• Introduction and Motivation• The GrenchMark Framework• Past and Current Experience with GrenchMark • A GrenchMark Success Story• Future Work• Conclusions
April 18, 20233
The Generic Problem of Analyzing, Testing, and Comparing Grids
• Use cases for automatically analyzing, testing, and comparing grids• Comparisons for system design and procurement • Functionality testing and system tuning• Performance testing/analysis of grid applications• …
• For grids, this problem is hard !• Testing in real environments is difficult• Grids change rapidly• Validity of tests• …
Gridlarge-scale collection
of federated resources,
e.g., CPUs, HDDs, …
April 18, 20234
A Generic Solution to Analyzing, Testing, and Comparing Grids• “ Generate and run synthetic grid workloads,
based on real and synthetic applications “
• Current alternatives (not covering all problems)• Benchmarking with real/synthetic applications
(representative?) • User-defined test management (statistically sound?)
• Advantages of using synthetic grid workloads• Statistically sound composition of benchmarks• Statistically sound test management• Generic: cover the use cases’ broad spectrum (to be shown)
April 18, 20235
GrenchMark: Towards a Generic Framework for Analyzing, Testing, and Comparing grids• What’s in a name?
grid benchmark → working towards a generic tool for the whole community: help standardizing the testing procedures, but benchmarks are too early; we use synthetic grid workloads instead
• What’s it about?A systematic approach to analyzing, testing, and comparing grid settings, based on synthetic workloads• A set of metrics for analyzing grid settings• A set of representative grid applications
• Both real and synthetic• Easy-to-use tools to create synthetic grid workloads• Flexible, extensible framework
April 18, 20237
GrenchMark: Iterative Research RoadmapSimple functional systemA.Iosup, J.Maassen, R.V.van Nieuwpoort, D.H.J.Epema,
Synthetic Grid Workloads with Ibis, KOALA, and GrenchMark, CoreGRID WS Integrated Research in Grid Computing, Nov 2005.
April 18, 20238
GrenchMark: Iterative Research Roadmap
Open-GrenchMark
CommunityEffort
Complex extensible systemThis work
April 18, 202310
… but Maybe More Complicated Than You Think
• Workload structure• User-defined and statistical models • Dynamic jobs arrival• Burstiness and self-similarity• Feedback, background load• Machine usage assumptions• Users, VOs• Interactive jobs, malleable, moldable, …
• Metrics• A(W) Run/Wait/Resp. Time • Efficiency, MakeSpan• Failure rate, waste
• (Grid) notions• Co-allocation, workflow
• Measurement methods• Long workloads• Saturated / non-saturated system• Start-up, production, and cool-down scenarios• Scaling workload to system
• Applications• Synthetic• Real
• Workload definition language
• Base language layer• Extended language layer
• Other• Can use the same workload for both simulations and real environments
April 18, 202313
Using GrenchMark: Grid System Analysis
• Performance testing: test the performance of an application (for sequential, MPI, Java/Ibis applications)• Report runtimes, waiting times, grid middleware
overhead• Automatic results analysis
• What-if analysis: evaluate potential situations• System change• Grid inter-operability• Special situations: spikes in demand
April 18, 202314
Using GrenchMark: Functionality Testing in Grid Environments
• System functionality testing: show the ability of the system to run various types of applications• Report failure rates
[ arguably, functionality in grids is even more important than performance ! 10% job failure rate in a controlled system like the DAS ]
• Periodic system testing: evaluate the current state of the grid• Replay workloads
April 18, 202315
Using GrenchMark: Comparing Grid Settings
• Single-site vs. co-allocated jobs: compare the success rate of single-site and co-allocated jobs, in a system without reservation capabilities• Single-site jobs 20% better vs. small co-allocated jobs (<32
CPUs), 30% better vs. large co-allocated jobs [setting and workload-dependent !]
• Unitary vs. composite jobs: compare the success rate of unitary and composite jobs, with and without failure handling mechanisms• Both 100% with simple retry mechanism
[setting and workload-dependent !]
April 18, 202316
A GrenchMark Success Story:Releasing the Koala Grid Scheduler on the DAS• Koala [ http://www.st.ewi.tudelft.nl/koala/ ]
• Grid Scheduler with co-allocation capabilities• DAS: The Dutch Grid, ~200 researchers• Initially
• Koala, a tested (!) scheduler, pre-release version• Test specifics
• 3 different job submission modules• Workloads with different jobs requirements,
inter-arrival rates, co-allocated v. single site jobs…• Evaluate: job success rate, Koala overhead and bottlenecks
• Results• 5,000+ jobs successfully run (all workloads); functionality
tests• 2 major bugs first day, 10+ bugs overall (all fixed)• KOALA is now officially released on the DAS
(full credit to KOALA developers, 10x for testing with GrenchMark)
April 18, 202317
A. Iosup, D.H.J. Epema (TU Delft), C.Franke, A. Papaspyrou, L. Schley, B. Song, R. Yahyapour (U Dortmund), On Grid Performance Evaluation using Synthetic Workloads, 12th Workshop on Job Scheduling Strategies for Parallel Processing (JSSPP), held in conjunction with SIGMETRICS, Saint Malo, France, June 2006 (accepted).
GrenchMark’s Current Status: pre-”Open-GrenchMark”
• Already done in Python [http://www.python.org]• Workload Generator• Generic Workload Submitter (Koala, Globus
GRAM, option to extend for JSDL, Condor, PBS, LSF, SGE, …)
• Applications• Unitary, 3 types: sequential, MPI, Ibis (Java)• +35 real and synthetic applications• Composite applications: DAG-based
• Extending modeling capabilities
April 18, 202318
A. Iosup, C. Dumitrescu, D.H.J. Epema (TU Delft), H. Li, L. Wolters (U Leiden), How are Real Grids Used? The Analysis of Four Grid Traces and Its Implications, Tech.Rep. TU Delft/PDS/2006-003, Apr 2006.
Towards Open-GrenchMark: Grid traces, Simulators, Benchmarks
• Distributed testing• Integrate with DiPerF (C. Dumitrescu, I. Raicu, M. Ripeanu)
• Grid traces analysis• Automatic tools for grid traces analysis
• Use in conjunction with simulators• Ability to generate workloads which can be used in
simulated environments (e.g., GangSim, GridSim, …)
• Grid benchmarks• Analyze the requirements for domain-specific grid
benchmarks
April 18, 202319
Conclusion
• GrenchMark generates diverse grid workloadseasy-to-use, flexible, portable, extensible, …
• Experienceused GrenchMark to test KOALA’s functionality and performance. used GrenchMark to analyze, test, and compare grid settings. 15,000+ jobs generated and run … and counting.
• (more) advertisementHave specific grid settings you would like to test? Test with GrenchMark!
April 18, 202320
Thank you!
Questions? Remarks? Observations? All welcome!
GrenchMark http://grenchmark.st.ewi.tudelft.nl/ [10x Paulo]
Alexandru IOSUPTU Delft
http://www.pds.ewi.tudelft.nl/~iosup/index.html [google: “iosup”]
Many thanks to Hashim Mohamed (Koala), Jason Maassen and Rob van Nieuwpoort (Ibis).