green computing omer rana [email protected]
DESCRIPTION
Green Computing Omer Rana [email protected]. The Need. Bill St. Arnaud (CANARIE, Inc). Impact of ICT Industry. Bill St. Arnaud (CANARIE, Inc). Virtualization Techniques. Bill St. Arnaud (CANARIE, Inc). Virtual Machine Monitors (IBM 1960s). - PowerPoint PPT PresentationTRANSCRIPT
![Page 2: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/2.jpg)
The Need
Bill St. Arnaud (CANARIE, Inc)
![Page 3: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/3.jpg)
Impact of ICT Industry
Bill St. Arnaud (CANARIE, Inc)
![Page 4: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/4.jpg)
Virtualization Techniques
Bill St. Arnaud (CANARIE, Inc)
![Page 5: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/5.jpg)
Virtual Machine Monitors (IBM 1960s)
A thin software layer that sits between hardware and the operating system— virtualizing and
managing all hardware resources
IBM Mainframe
IBM VM/370
CMS MVS CMS CMS
App App App App
Ed Bugnion, VMWare
![Page 6: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/6.jpg)
Old idea from the 1960s
• IBM VM/370 – A VMM for IBM mainframe– Multiple OS environments on expensive hardware– Desirable when few machine around
• Popular research idea in 1960s and 1970s– Entire conferences on virtual machine monitor– Hardware/VMM/OS designed together
• Interest died out in the 1980s and 1990s.– Hardware got cheap– Operating systems got more more powerful (e.g multi-
user)
Ed Bugnion, VMWare
![Page 7: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/7.jpg)
A return to Virtual Machines
• Disco: Stanford research project (1996-):– Run commodity OSes on scalable multiprocessors– Focus on high-end: NUMA, MIPS, IRIX
• Hardware has changed:– Cheap, diverse, graphical user interface– Designed without virtualization in mind
• System Software has changed: – Extremely complex– Advanced networking protocols– But even today :
• Not always multi-user• With limitations, incompatibilities, …
Ed Bugnion, VMWare
![Page 8: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/8.jpg)
The Problem Today
Intel Architecture
Operating System
Ed Bugnion, VMWare
![Page 9: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/9.jpg)
The VMware Solution
Intel Architecture
Operating System
Operating System
Intel Architecture
Ed Bugnion, VMWare
![Page 10: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/10.jpg)
VMware™ MultipleWorlds™ Technology
A thin software layer that sits between Intel hardware and the operating system— virtualizing and managing all hardware
resources
Intel Architecture
VMware MultipleWorlds
Win2000
WinNT Linux Win
2000
App App App App
Ed Bugnion, VMWare
![Page 11: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/11.jpg)
MultipleWorlds Technology
A world is an application execution environment with its own operating system
World
Intel Architecture
VMware MultipleWorlds
Win2000
WinNT Linux Win
2000
App App App App
Ed Bugnion, VMWare
![Page 12: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/12.jpg)
Virtual Hardware
Floppy Disks
Parallel Ports Serial/Com Ports
Ethernet
Keyboard
Mouse
Monitor(VMM)
IDE Controller SCSI Controller
Sound Card
Ed Bugnion, VMWare
![Page 13: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/13.jpg)
Attributes of MultipleWorlds Technology
• Software compatibility– Runs pretty much all software
• Low overheads/High performance– Near “raw” machine performance
• Complete isolation– Total data isolation between virtual machines
• Encapsulation– Virtual machines are not tied to physical machines
• Resource management
Ed Bugnion, VMWare
![Page 14: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/14.jpg)
Hosted VMware Architecture
VMware achieves both near-native execution speed and broad device support by transparently switching* between Host Mode and VMM Mode.
Guest OS Applications
Guest Operating System
Host OS Apps
Host OS
PC HardwareDisks Memory CPUNIC
VMware App Virtual Machine
VMware Driver Virtual Machine Monitor
Host Mode VMM Mode
VMware, acting as an application, uses the host to access other devices such as the hard disk, floppy, or network card
The VMware Virtual machine monitor allows each guest OS to directly access the processor (direct execution)
*VMware typically switches modes 1000 times per second
Ed Bugnion, VMWare
![Page 15: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/15.jpg)
Hosted VMM Architecture• Advantages:
– Installs and runs like an application– Portable – host OS does I/O access– Coexists with applications running on the host
• Limits:– Subject to Host OS:
• Resource management decisions• OS failures
• Usenix 2001 paper: J. Sugerman, G. Venkitachalam and B.-H. Lim, “Virtualizing I/O on VMware
Workstation’s Hosted Architecture”.
Ed Bugnion, VMWare
![Page 16: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/16.jpg)
Virtualizing a Network Interface
Host OS
PC HardwarePhysical NIC
VMApp
VMDriver
Guest OS
VMM
Phy
sica
l Eth
erne
t
NIC Driver
NIC Driver
Virtual Bridge
Virtual Network Hub
Ed Bugnion, VMWare
![Page 17: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/17.jpg)
The rise of data centers
• Single place for hosting servers and data• ISP’s now take machines hosted at data
centers• Run by large companies – like BT• Manage
– Power– Computation + Data– Cooling systems– Systems Admin + Network Admin
![Page 18: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/18.jpg)
Data Centre in Tokyo From:Satoshi Matsuoka
![Page 19: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/19.jpg)
http://www.attokyo.co.jp/eng/facility.html
![Page 20: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/20.jpg)
Martin J. Levy (Tier1 Research) and Josh Snowhorn (Terremark)
![Page 21: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/21.jpg)
Martin J. Levy (Tier1 Research) and Josh Snowhorn (Terremark)
![Page 22: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/22.jpg)
Martin J. Levy (Tier1 Research) and Josh Snowhorn (Terremark)
![Page 23: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/23.jpg)
Requirements• Power an important design constraint:
– Electricity costs– Heat dissipation
• Two key options in clusters – enable scaling of:– Operating frequency (square relation)– Supply voltage (cubic relation)
• Balance QoS requirements – e.g.fraction of workload to process locally – with power management
![Page 24: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/24.jpg)
From: Salim Hariri, Mazin Yousif
![Page 25: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/25.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 26: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/26.jpg)
Martin J. Levy (Tier1 Research) and Josh Snowhorn (Terremark)
![Page 27: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/27.jpg)
The case for power management in HPC
• Power/energy consumption a critical issue– Energy = Heat; Heat dissipation is costly
– Limited power supply
– Non-trivial amount of money
• Consequence– Performance limited by available power
– Fewer nodes can operate concurrently
• Opportunity: bottlenecks– Bottleneck component limits performance of other components
– Reduce power of some components, not overall performance
• Today, CPU is:– Major power consumer (~100W),
– Rarely bottleneck and
– Scalable in power/performance (frequency & voltage)
Power/performance“gears”
![Page 28: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/28.jpg)
Is CPU scaling a win?
• Two reasons:1. Frequency and voltage scaling
Performance reduction less than
Power reduction
2. Application throughputThroughput reduction less than
Performance reduction
• Assumptions– CPU large power consumer– CPU driver– Diminishing throughput gains
performance (freq)
pow
er
ap
plic
ati
on t
hro
ug
hput
performance (freq)
(1)
(2)
CPU powerP = ½ CVf2
![Page 29: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/29.jpg)
AMD Athlon-64
• x86 ISA• 64-bit technology• Hypertransport technology – fast memory bus• Performance
– Slower clock frequency– Shorter pipeline (12 vs. 20)– SPEC2K results
• 2GHz AMD-64 is comparable to 2.8GHz P4• P4 better on average by 10% & 30% (INT & FP)
• Frequency and voltage scaling– 2000 – 800 MHz– 1.5 – 1.1 Volts
From: Vincent W. Freeh (NCSU)
![Page 30: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/30.jpg)
LMBench results
• LMBench– Benchmarking suite
– Low-level, micro data
• Test each “gear”
GearFrequency (Mhz)
Voltage
0 2000 1.5
1 1800 1.4
2 1600 1.3
3 1400 1.2
4 1200 1.1
6 800 0.9
From: Vincent W. Freeh (NCSU)
![Page 31: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/31.jpg)
Operating system functions
From: Vincent W. Freeh (NCSU)
![Page 32: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/32.jpg)
Communication
From: Vincent W. Freeh (NCSU)
![Page 33: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/33.jpg)
The problem• Peak power limit, P
– Rack power– Room/utility– Heat dissipation
• Static solution, number of servers is– N = P/Pmax
– Where Pmax maximum power of individual node
• Problem– Peak power > average power (Pmax > Paverage)
– Does not use all power – N * (Pmax - Paverage) unused
– Under performs – performance proportional to N– Power consumption is not predictable
From: Vincent W. Freeh (NCSU)
![Page 34: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/34.jpg)
Safe over provisioning in a cluster• Allocate and manage power among M > N nodes
– Pick M > N• Eg, M = P/Paverage
– MPmax > P
– Plimit = P/M
• Goal– Use more power, safely under limit– Reduce power (& peak CPU performance) of individual nodes– Increase overall application performance
time
pow
er Pmax
Paverage
P(t)
time
pow
er
PlimitPaverage
P(t)
Pmax
From: Vincent W. Freeh (NCSU)
![Page 35: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/35.jpg)
Safe over provisioning in a cluster
• Benefits– Less “unused” power/energy– More efficient power use
• More performance under same power limitation – Let P be performance
– Then more performance means: MP * > NP– Or P */ P > N/M or P */ P > Plimit/Pmax
time
pow
er Pmax
Paverage
P(t)
time
pow
er
PlimitPaverage
P(t)
Pmax
unusedenergy
From: Vincent W. Freeh (NCSU)
![Page 36: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/36.jpg)
When is this a win?
• When P */ P > N/M
or P */ P > Plimit/Pmax
In words:
power reduction more than
performance reduction
• Two reasons:1. Frequency and voltage scaling
2. Application throughput
performance (freq)
pow
er
ap
plic
ati
on t
hro
ug
hput
P * / P
< P av
erag
e/P m
ax
P * / P
> P av
erag
e/P m
ax
performance (freq)
(1)
(2)
From: Vincent W. Freeh (NCSU)
![Page 37: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/37.jpg)
Feedback-directed, adaptive power control
• Uses feedback to control power/energy consumption– Given power goal
– Monitor energy consumption
– Adjust power/performance of CPU
• Several policies– Average power
– Maximum power
– Energy efficiency: select slowest gear (g) such that
From: Vincent W. Freeh (NCSU)
![Page 38: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/38.jpg)
A more holistic approach: Managing a Data Center
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
CRAC: Computer Room Air Conditioningunits
![Page 39: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/39.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 40: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/40.jpg)
Location of Cooling Unitssix CRAC units are serving 1000 servers, consuming 270
KW of power out of a total capacity of 600 KW
http://blogs.zdnet.com/BTL/?p=4022
![Page 41: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/41.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 42: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/42.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 43: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/43.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 44: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/44.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 45: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/45.jpg)
From: Justin Moore, Ratnesh Sharma, Rocky Shih, Jeff Chase, Chandrakant Patel, Partha Ranganathan (HP Labs)
![Page 46: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/46.jpg)
From:Satoshi Matsuoka
![Page 47: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/47.jpg)
From:Satoshi Matsuoka
![Page 48: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/48.jpg)
From:Satoshi Matsuoka
![Page 49: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/49.jpg)
From:Satoshi Matsuoka
![Page 50: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/50.jpg)
From:
Satoshi Matsuoka
![Page 51: Green Computing Omer Rana o.f.rana@csrdiff.ac.uk](https://reader034.vdocuments.us/reader034/viewer/2022051516/56813c49550346895da5c7f8/html5/thumbnails/51.jpg)
From:Satoshi Matsuoka