an overview of high performance computing and future requirements

38
An Overview of High An Overview of High Performance Computing Performance Computing and and Future Requirements Future Requirements Jack Dongarra University of Tennessee Oak Ridge National Laboratory 1

Upload: tobias-raymond

Post on 03-Jan-2016

24 views

Category:

Documents


1 download

DESCRIPTION

An Overview of High Performance Computing and Future Requirements. Jack Dongarra University of Tennessee Oak Ridge National Laboratory. TPP performance. Rate. Size. H. Meuer, H. Simon, E. Strohmaier, & JD - Listing of the 500 most powerful Computers in the World - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: An Overview of High Performance Computing  and Future Requirements

An Overview of High An Overview of High Performance Computing Performance Computing

andandFuture RequirementsFuture Requirements

Jack DongarraUniversity of Tennessee

Oak Ridge National Laboratory

1

Page 2: An Overview of High Performance Computing  and Future Requirements

H. Meuer, H. Simon, E. Strohmaier, & JDH. Meuer, H. Simon, E. Strohmaier, & JD

- Listing of the 500 most powerful Computers in the World - Yardstick: Rmax from LINPACK MPP

Ax=b, dense problem

- Updated twice a yearSC‘xy in the States in NovemberMeeting in Germany in June

- All data available from www.top500.org

Size

Rate

TPP performance

2

Page 3: An Overview of High Performance Computing  and Future Requirements

Performance DevelopmentPerformance Development

1 Gflop/s

1 Tflop/s

100 Mflop/s

100 Gflop/s

100 Tflop/s

10 Gflop/s

10 Tflop/s

1 Pflop/s

100 Pflop/s

10 Pflop/s

59.7 GFlop/s59.7 GFlop/s

400 MFlop/s400 MFlop/s

1.17 TFlop/s1.17 TFlop/s

1.1 PFlop/s1.1 PFlop/s

17.08 TFlop/s17.08 TFlop/s

22.9 PFlop/s22.9 PFlop/s

SUM

N=1

N=500

6-8 years

My Laptop

1993 1995 1997 1999 2001 2003 2005 2007 2009

3

Page 4: An Overview of High Performance Computing  and Future Requirements

Performance of Top25 Over 10 Years

PFlop/s

Rank

Year

4

Page 5: An Overview of High Performance Computing  and Future Requirements

Cores in the Top25 Over Last 10 Years

5

Page 6: An Overview of High Performance Computing  and Future Requirements

Looking at the Gordon Bell Prize(Recognize outstanding achievement in high-performance computing applications and encourage development of parallel processing )

1 GFlop/s; 1988; Cray Y-MP; 8 Processors Static finite element analysis

1 TFlop/s; 1998; Cray T3E; 1024 Processors Modeling of metallic magnet atoms, using a

variation of the locally self-consistent multiple scattering method.

1 PFlop/s; 2008; Cray XT5; 1.5x105 Processors Superconductive materials

1 EFlop/s; ~2018; ?; 1x107 Processors (109 threads)

6

Page 7: An Overview of High Performance Computing  and Future Requirements

Performance Development in Top500

1 Gflop/s

1 Tflop/s

100 Mflop/s

100 Gflop/s

100 Tflop/s

10 Gflop/s

10 Tflop/s

1 Pflop/s

100 Pflop/s

10 Pflop/s SUM

N=1

N=500 GordonBell

Winners

GordonBell

Winners

7

Page 8: An Overview of High Performance Computing  and Future Requirements

Processors Used in Processors Used in SupercomputersSupercomputers

Intel 71%AMD 13%IBM 7%

8

Page 9: An Overview of High Performance Computing  and Future Requirements

Countries / System ShareCountries / System Share

9

Page 10: An Overview of High Performance Computing  and Future Requirements

Industrial Use of SupercomputersIndustrial Use of Supercomputers• Of the 500 Fastest

Supercomputer • Worldwide,

Industrial Use is > 60%

Aerospace Automotive Biology CFD Database Defense Digital Content Creation Digital Media Electronics Energy Environment Finance Gaming Geophysics Image Proc./Rendering Information Processing Service Information Service Life Science Media Medicine Pharmaceutics Research Retail Semiconductor Telecomm Weather and Climate Research Weather Forecasting

10

Page 11: An Overview of High Performance Computing  and Future Requirements

3333rdrd List: The TOP10 List: The TOP10

11

Page 12: An Overview of High Performance Computing  and Future Requirements

3333rdrd List: The TOP10 List: The TOP10

12

Page 13: An Overview of High Performance Computing  and Future Requirements

TFl

op/s

Rank

Distribution of the Top500

11 systems > 250 Tflop/s

78 systems > 50 Tflop/s

223 systems > 25 Tflop/s

17.1 Tflop/s

1.1 Pflop/s

2 systems > 1 Pflop/s

274 systems replaced last time

13

Page 14: An Overview of High Performance Computing  and Future Requirements

LANL Roadrunner LANL Roadrunner A Petascale System in 2008A Petascale System in 2008

“Connected Unit” cluster192 Opteron nodes(180 w/ 2 dual-Cell

bladesconnected w/ 4 PCIe x8

links)

≈ 13,000 Cell HPC chips ≈ 1.33 PetaFlop/s (from Cell)≈ 7,000 dual-core Opterons≈ 122,000 cores

17 clusters

2nd stage InfiniBand 4x DDR interconnect(18 sets of 12 links to 8 switches)

2nd stage InfiniBand interconnect (8 switches)

Based on the 100 Gflop/s (DP) Cell chip

Hybrid Design (2 kinds of chips & 3 kinds of cores)Programming required at 3 levels.

Dual Core Opteron Chip

Cell chip for each core

14

Page 15: An Overview of High Performance Computing  and Future Requirements

ORNL’s Newest System Jaguar XT5 ORNL’s Newest System Jaguar XT5

Office ofScience

Will be upgraded this year to a 2

Pflop/s system with > 224K AMD

Istanbul Cores.

Jaguar Total XT5 XT4

Peak Performance 1,645 1,382 263

AMD Opteron Cores 181,504

150,176

31,328

System Memory (TB) 362 300 62

Disk Bandwidth (GB/s) 284 240 44

Disk Space (TB) 10,750 10,000 750

Interconnect Bandwidth (TB/s) 532 374 15715

Page 16: An Overview of High Performance Computing  and Future Requirements

ORNL/UTK Computer Power Cost Projections 2008-2012

• Over the next 5 years ORNL/UTK will deploy 2 large Petascale systems

• Using 15 MW today

• By 2012 close to 50MW!!

• Power costs close to $10M today.

• Cost estimates based on $0.07 per KwH

Cost Per YearPower becomes the architectural driver for future large systems

> $10M > $20M > $30M

16

Page 17: An Overview of High Performance Computing  and Future Requirements

Something’s Happening Something’s Happening Here…Here…

• In the “old days” it was: each year processors would become faster

• Today the clock speed is fixed or getting slower

• Things are still doubling every 18 -24 months

• Moore’s Law reinterpretated. Number of cores

double every 18-24 months

From K. Olukotun, L. Hammond, H. Sutter, and B. Smith

A hardware issue just became a software problem

17

Page 18: An Overview of High Performance Computing  and Future Requirements

Moore’s Law ReinterpretedMoore’s Law Reinterpreted• Number of cores per chip doubles

every 2 year, while clock speed decreases (not increases). Need to deal with systems with millions of

concurrent threads•Future generation will have billions of threads!

Need to be able to easily replace inter-chip parallelism with intro-chip parallelism

• Number of threads of execution doubles every 2 year

18

Page 19: An Overview of High Performance Computing  and Future Requirements

Today’s MulticoresToday’s Multicores99% of Top500 Systems Are Based on Multicore99% of Top500 Systems Are Based on Multicore

Sun Niagra2 (8 cores)

Intel Polaris [experimental] (80 cores)

IBM BG/P (4 cores)AMD Istambul (6 cores)

IBM Cell (9 cores)

Intel Clovertown (4 cores)

282 use Quad-Core204 use Dual-Core 3 use Nona-core

Fujitsu Venus (8 cores)

IBM Power 7 (8 cores)

19

Page 20: An Overview of High Performance Computing  and Future Requirements

What’s Next?What’s Next?

Many Floating-Point Cores

All Large CoreAll Large CoreMixed LargeMixed LargeandandSmall CoreSmall Core

All Small CoreAll Small Core

Many Small CoresMany Small Cores

Different Classes of Chips Home Games / Graphics Business Scientific

Different Classes of Chips Home Games / Graphics Business Scientific

20

Page 21: An Overview of High Performance Computing  and Future Requirements

CommodityCommodity• Moore’s “Law” favored consumer commodities

Economics drove enormous improvements Specialized processors and mainframes faltered Custom HPC hardware largely disappeared Hard to compete against 50%/year improvement

• Implications Consumer product space defines outcomes It does not always go where we hope or expect Research environments track commercial trends Driven by market economics Think about processors, clusters, commodity storage

21

Page 22: An Overview of High Performance Computing  and Future Requirements

Future Computer SystemsFuture Computer Systems• Most likely be a hybrid design• Think standard multicore chips and

accelerator (GPUs)• Today accelerators are attached• Next generation more integrated• Intel’s Larrabee in 2010

8,16,32,or 64 x86 cores

• AMD’s Fusion in 2011 Multicore with embedded graphics ATI

• Nvidia’s plans?Intel Larrabee

22

Page 23: An Overview of High Performance Computing  and Future Requirements

Architecture of InterestArchitecture of Interest

• Manycore chip• Composed of hybrid cores

Some general purpose Some graphics Some floating point

23

Page 24: An Overview of High Performance Computing  and Future Requirements

Architecture of InterestArchitecture of Interest

• Board composed of multiple chips sharing memory

Mem

ory

24

Page 25: An Overview of High Performance Computing  and Future Requirements

Architecture of InterestArchitecture of Interest

• Rack composed of multiple boards

Mem

ory

25

Page 26: An Overview of High Performance Computing  and Future Requirements

Architecture of InterestArchitecture of Interest• A room full of these racks

• Think millions of cores

Mem

ory

26

Page 27: An Overview of High Performance Computing  and Future Requirements

Major Changes to SoftwareMajor Changes to Software• Must rethink the design of our

software Another disruptive technology

•Similar to what happened with cluster computing and message passing

Rethink and rewrite the applications, algorithms, and software

• Numerical libraries for example will change For example, both LAPACK and

ScaLAPACK will undergo major changes to accommodate this

27

Page 28: An Overview of High Performance Computing  and Future Requirements

Five Important Features to Consider Five Important Features to Consider When Computing at ScaleWhen Computing at Scale

• Effective Use of Many-Core and Hybrid architectures Dynamic Data Driven Execution Block Data Layout

• Exploiting Mixed Precision in the Algorithms Single Precision is 2X faster than Double Precision With GP-GPUs 10x

• Self Adapting / Auto Tuning of Software Too hard to do by hand

• Fault Tolerant Algorithms With 1,000,000’s of cores things will fail

• Communication Avoiding Algorithms For dense computations from O(n log p) to O(log p)

communications GMRES s-step compute ( x, Ax, A2x, … Asx )

28

Page 29: An Overview of High Performance Computing  and Future Requirements

Quasi MainstreamQuasi MainstreamProgramming ModelsProgramming Models

• C, Fortran, C++ and MPI• OpenMP, pthreads• (CUDA, RapidMind, Cn) OpenCL• PGAS (UPC, CAF, Titanium)• HPCS Languages (Chapel, Fortress, X10)• HPC Research Languages and Runtime• HLL (Parallel Matlab, Grid Mathematica, etc.)

29

Page 30: An Overview of High Performance Computing  and Future Requirements

DOE Office of Science Next DOE Office of Science Next SystemSystem

• DOE's requirement for 20-40 PF of compute capability split between the Oak Ridge and Argonne LCF centers

• ORNL's proposed system will be based on accelerator technology   includes software development environment

• Plans are to deploy the system in late 2011 with users getting access in 2012

30

Page 31: An Overview of High Performance Computing  and Future Requirements

Sequoia LLNLSequoia LLNL• Diverse usage models drive

platform and simulation environment requirements Will be 2D ultra-res and 3D high-res

Quantification of Uncertainty engine

3D Science capability for known unknowns and unknown unknowns

• Peak 20 petaFLOP/s• IBM BG/Q• Target production 2011-2016• Sequoia Component Scaling

Memory B:F = 0.08 Mem BW B:F = 0.2 Link BW B:F = 0.1 Min Bisect B:F = 0.03 SAN BW GB/:PF/s = 25.6 F is peak FLOP/s 31

Page 32: An Overview of High Performance Computing  and Future Requirements

NSF University of Illinois; Blue Waters NSF University of Illinois; Blue Waters

Blue Waters NCSA/Illinois 1 Pflop sustained per second

Ranger TACC/U of Texas 504 Tflop/s peak per second

Kraken NICS/U of Tennessee

1 Pflops peak per second

Campuses across the U.S. Several sites 50-100 Tflops peak per second

Blue Waters will be the powerhouse of the National Science Foundation’s strategy to support supercomputers for scientists nationwide

T1

T2

T3

32

Page 33: An Overview of High Performance Computing  and Future Requirements

Blue Waters - Main CharacteristicsBlue Waters - Main Characteristics• Hardware:

Processor: IBM Power7 multicore architecture• 8 cores per chip• 32 Gflop/s per core; 256 Gflop/s chip

More than 200,000 cores will be available Capable of simultaneous multithreading (SMT) Vector multimedia extension capability (VMX) Four or more floating-point operations per

cycle Multiple levels of cache – L1, L2, shared L3 32 GB+ memory per SMP, 2 GB+ per core 16+ cores per SMP 10+ Petabytes of disk storage Network interconnect with RDMA technology

NSF University of Illinois; Blue Waters NSF University of Illinois; Blue Waters

33

Page 34: An Overview of High Performance Computing  and Future Requirements

DARPA’s RFI on Ubiquitous High DARPA’s RFI on Ubiquitous High Performance ComputingPerformance Computing

• 1 PFLOP/S HPL, air-cooled, single 19-inch cabinet ExtremeScale system.

• The power budget for the cabinet is 57 kW, including cooling. • Achieve 50 GFLOPS/W for the High-Performance Linpack (HPL)

benchmark. • The system design should provide high performance for scientific

and engineering applications. • The system should be a highly programmable system that does

not require the application developer to directly manage the complexity of the system to achieve high performance.

• The system must explicitly show a high degree of innovation and software and hardware co-design throughout the life of the program.

• 5 phases; 1) conceptual designs; 2) execution model; 3) full-scale simulation; 4) delivery;

5) modify and refine.

34

Page 35: An Overview of High Performance Computing  and Future Requirements

Exascale ComputingExascale Computing• Exascale systems are likely feasible by 20172 • 10-100 Million processing elements (cores or

mini-cores) with chips perhaps as dense as 1,000 cores per socket, clock rates will grow more slowly

• 3D packaging likely• Large-scale optics based interconnects• 10-100 PB of aggregate memory• Hardware and software based fault management

• Heterogeneous cores• Performance per watt — stretch goal 100 GF/watt of

sustained performance >> 10 – 100 MW Exascale system • Power, area and capital costs will be significantly higher

than for today’s fastest systems

Google: exascale computing study35

Page 36: An Overview of High Performance Computing  and Future Requirements

Conclusions Conclusions • For the last decade or more, the research

investment strategy has been overwhelmingly biased in favor of hardware.

• This strategy needs to be rebalanced - barriers to progress are increasingly on the software side.

• Moreover, the return on investment is more favorable to software. Hardware has a half-life measured in years, while

software has a half-life measured in decades.• High Performance Ecosystem out of balance

Hardware, OS, Compilers, Software, Algorithms, Applications• No Moore’s Law for software, algorithms and applications

36

Page 37: An Overview of High Performance Computing  and Future Requirements

Collaborators / SupportCollaborators / Support

Employment opportunities for post-docs in the ICL group at Tennessee

• Top500– Hans Meuer, Prometeus – Erich Strohmaier, LBNL/NERSC– Horst Simon, LBNL/NERSC

37

Page 38: An Overview of High Performance Computing  and Future Requirements

If you are wondering what’s beyond If you are wondering what’s beyond ExaFlopsExaFlops

Mega, Giga, Tera, Peta, Exa, Zetta …

103 kilo 106 mega 109 giga 1012 tera 1015 peta 1018 exa

1024 yotta 1027 xona 1030 weka 1033 vunda 1036 uda 1039 treda 1042 sorta1045 rinta1048 quexa1051 pepta 1054 ocha 1057 nena 1060 minga 1063 luma

1021 zetta

38