future of high performance computing

34
National Center for Supercomputing Applications University of Illinois at Urbana-Champaign Future of High Performance Computing

Upload: others

Post on 16-Apr-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Future of High Performance Computing

National Center for Supercomputing Applications

University of Illinois at Urbana-Champaign

Future of HighPerformance Computing

Page 2: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Outline of Presentation

Page 3: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Progress inHigh Performance Computing

Page 4: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Progress in High Performance Computing

Relentless Increase in Performance

B

BBBBBBB

BBBB

BBB

BBB

BBBBB

B

B

BBBBB

BB

109

1010

1011

1012

1013

1014

1015

1016

’92 ’94 ’96 ’98 ’00 ’02 ’04 ’06 ’08 ’10 ’12

Rmax

Page 5: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Progress in High Performance Computing

Relentless Increase in Number of Cores

B

B

B

BBB

B

B

BBBBBBBBBB

BBBBB

B

B

BBBB

B

BB

102

103

104

105

106N

umbe

r of

Cor

es

’92 ’94 ’96 ’98 ’00 ’02 ’04 ’06 ’08 ’10 ’12

Page 6: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions inComputing Technologies

Page 7: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

B

B

B

B

BB

B B B

10

100

1000

10000

1990 1995 2000 2005 2010

Directions in Computing Technologies

Increasing Clock Frequency & PerformanceF

req

uen

cy (

MH

z)

“In the past, performance

scaling in conventional single-

core processors has been

accomplished largely through

increases in clock frequency

(accounting for roughly 80

percent of the performance

gains to date).”

Platform 2015

S. Y. Borkar et al., 2006

Intel Corporation

Intel Pentium

Page 8: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

From Uni-core to Multi-core Processors

AMDUni-, Dual-,

Quad-core,

Processors

IntelMulti-core

Performance

Intel Teraflops Chip

Page 9: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

Multi-core 2009: Intel’s Nehalem

Page 10: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

B

B

B

B

BB

B B B

10

100

1000

10000

1990 1995 2000 2005 2010

Directions in Computing Technologies

Switch to Multicore ChipsF

req

uen

cy (

MH

z)

du

al

co

re

qu

ad

co

re

“For the next several years

the only way to obtain

significant increases in

performance will be through

increasing use of parallelism:

– 8 in 2009

– 16 in 2011

– 32 in 2013

– etc.

Page 11: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

On to Many-core Chips

AMD Firestream

(800 cores)

NVIDIA Telsa

(240 cores)IBM Cell

(1+8 cores)

Page 12: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

New Technologies for HPCGflops

Courtesy of John Owens (UCSD) & Ian Buck (NVIDIA)

2002 2003 2004 2005 2006 2007

3.4 GHzDual-core

2.66 GHzQuad-core

1.35 GHz G80

1.50 GHz G80

NVIDIA (GPU)

INTEL (CPU)

0

50

100

150

200

250

300

350

400

Page 13: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

NVIDIA: Selected Benchmarks

* For GeForce 8800 @ 346 GF (SP), W-m. Hwu et al., 2007

Page 14: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

Benchmarks: Direct SCF Calculations*

* GeForce 8800 @ 346 GF (SP), I. Ufimtsev and T. Martinez, CiSE 10, 26-34 (2008).** Using GAMESS on AMD Opteron 175 CPU.

EnergyTime/Iter (s)

Page 15: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

NVIDIA: Tesla S1070

Page 16: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

Memory Subsystem

Page 17: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Directions in Computing Technologies

Communications Subsystem

Page 18: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Era ofPetascale Computing

Page 19: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Science @ Petascale

Molecular Science Weather & Climate Forecasting

Earth ScienceAstronomy Health

Petascale computing will enable advances in a broad range of

science and engineering disciplines:

Page 20: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Era of Petascale Computing

LANL Roadrunner Computer System

IBM Roadrunner Petascale System

Page 21: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Era of Petascale Computing

ORNL Jaguar Computer System

Cray Jaguar (XT5) Petascale System

Page 22: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Era of Petascale Computing

NSF’s Strategy for High-end Computing

FY’07 FY’11FY’10FY’09FY’08

Sci

ence

an

d E

ng

inee

rin

g C

ap

ab

ilit

y(l

og

ari

thm

ic s

cale

)

Track 1 System

Track 2 Systems

Track 3 Systems

Page 23: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Era of Petascale Computing

NSF’s Track 2 Computing Systems

TACC UT-ORNL PSC

System Attribute Ranger Kraken

Status Operational In progress In progress

Vendor Sun Cray SGI

Processor AMD AMD Intel

Peak Performance (TF) 579 ~1,000

Number Cores/Chip 4 6

Number Processor Cores 62,976 ~100,000 ~100,000

Amount Memory (TB) 123 ~100 ~100

Amount Disk Storage (PB) 1.73 3.3

External Bandwidth (Gbps) 10 20

Page 24: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue WatersPetascale Computing System

Page 25: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue Waters Petascale Computing System

Goals for Blue Waters• Maximize Core Performance

… minimize number of cores needed for a given level of performance aswell as lessen the impact of sections of code with limited scalability

• Maximize Application Scalability… low latency, high-bandwidth communications fabric

• Solve Memory-intensive Problems… large amount of memory

… low latency, high-bandwidth memory subsystem

• Solve Data-intensive Problems… high-bandwidth I/O subsystem

… large quantity of on-line disk, large quantity of archival storage

• Provide Reliable Operation… maximize system integration

… mainframe reliability, availability, serviceability (RAS) technologies

Page 26: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue Waters Petascale Computing System

Blue Waters Computing SystemSystem Attribute Abe Blue Waters*

Vendor Dell IBM

Processor Intel Xeon 5300 IBM Power7

Peak Performance (PF) 0.090

Sustained Performance (PF) 0.005 1

Number of Cores/Chip 4

Number of Processor Cores 9,600 >200,000

Amount of Memory (PB) 0.0144 >0.8

Amount of Disk Storage (PB) 0.1 >10

Amount of Archival Storage (PB) 5 >500

External Bandwidth (Gbps) 40 100-400

* Reference petascale computing system (no accelerators).

Page 27: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue Waters Petascale Computing System

Blue Waters Project

WBS 4.1

IPiPE

WBS 4.2

ISV Forum

WBS 4.3

Workforce Dev.

WBS 4.0Industrial Engagement

WBS 2.1

Petascale Apps

WBS 2.2

Petascale Research

WBS 2.0Applications/Research

WBS 1.0Blue Waters System

WBS 1.1

IBM HW & SW

WBS 1.2

Hardware (non-IBM)

WBS 1.3

Software (non-IBM)

WBS 1.4

Integ., Test & Trans

WBS 3.1

Undergrad. Ed.

WBS 3.2

Graduate Ed.

WBS 3.3

Training

WBS 3.4

Broadening Part.

WBS 3.0ETAO

WBS 5.1

PCF

WBS 5.2

Trans. From ACB

WBS 5.0Facilities

WBS 6.1

Tracking & Controls

WBS 6.2

Finan., Contracts, HR

WBS 6.3

Risk Management

WBS 6.4

Config. Management

WBS 6.0Project Management

WBS 6.5

Reporting & Reviews

WBS 6.6

Communications

Blue Waters ProjectWork Breakdown Structure

WBS 6.7

Cybersecurity

Note: this is not an organizational chartWBS 6.8

MIS Tools

Page 28: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue Waters Petascale Computing System

Great Lakes Consortium for Petascale Computation

Goal: Facilitate the widespread and effective use of petascale computing to address

frontier research questions in science, technology and engineering at research,

educational and industrial organizations across the region and nation.

Charter Members

Page 29: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Blue Waters Petascale Computing System

Petascale Computing Facility

www.ncsa.uiuc.edu/BlueWaters

Partners

EYP MCF/

Gensler

IBM

Yahoo!

Page 30: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Challenges inPetascale Computing

Page 31: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Challenges in Petascale Computing

Accuracy of Computational Models

-30.0

-20.0

-10.0

0.0

10.0

20.0

0 10 20 30 40 50

Err

or

(mE

h)

MPn

Perturbation

expansion not

converging!

Energy of HF(aug-cc-pVDZ): Olsen et al., J. Chem. Phys. 105, 5082 (1996)

Page 32: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Challenges in Petascale Computing

Scalability of Algorithms

(F. Gygi et al., “Large-Scale Electronic Structure Calculations of High-Z Metals on Blue

Gene/L Platform,” Proceedings of Supercomputing, 2006)

108.0 TF

187.7 TF207.3 TF (56% peak)

Weak

Scaling!

Page 33: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Challenges in Petascale Computing

More Challenges

Page 34: Future of High Performance Computing

PICSciE Seminar • 2 February 2009 • Princeton University

Questions?