power efficiency metrics for the top500

21
Power Efficiency Metrics for the Top500 Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab

Upload: dangduong

Post on 20-Dec-2016

224 views

Category:

Documents


4 download

TRANSCRIPT

Page 1: Power Efficiency Metrics for the Top500

Power EfficiencyMetrics for the Top500

Shoaib Kamil and John Shalf

CRD/NERSC

Lawrence Berkeley National Lab

Page 2: Power Efficiency Metrics for the Top500

Power for Single Processors

Page 3: Power Efficiency Metrics for the Top500

HPC Concurrency on the RiseTotal # of Processors in Top15

0

50000

100000

150000

200000

250000

300000

350000

Jun-93

Dec-93

Jun-94

Dec-94

Jun-95

Dec-95

Jun-96

Dec-96

Jun-97

Dec-97

Jun-98

Dec-98

Jun-99

Dec-99

Jun-00

Dec-00

Jun-01

Dec-01

Jun-02

Dec-02

Jun-03

Dec-03

Jun-04

Dec-04

Jun-05

Dec-05

Jun-06

List

Pro

cess

ors

Page 4: Power Efficiency Metrics for the Top500

HPC Power Draw on the RiseGrowth in Power Consumption (Top50)

0.00

500.00

1000.00

1500.00

2000.00

2500.00

3000.00

3500.00

4000.00

Jun-00

Sep-00

Dec-00

Mar-01

Jun-01

Sep-01

Dec-01

Mar-02

Jun-02

Sep-02

Dec-02

Mar-03

Jun-03

Sep-03

Dec-03

Mar-04

Jun-04

Sep-04

Dec-04

Mar-05

Jun-05

Sep-05

Dec-05

Mar-06

Jun-06

System

Po

wer (

kW

)

max pwr

avg pwr

Growth in Power Consumption (Top50)Excluding Cooling

0.00

500.00

1000.00

1500.00

2000.00

2500.00

3000.00

Jun-00

Sep-00

Dec-00

Mar-01

Jun-01

Sep-01

Dec-01

Mar-02

Jun-02

Sep-02

Dec-02

Mar-03

Jun-03

Sep-03

Dec-03

Mar-04

Jun-04

Sep-04

Dec-04

Mar-05

Jun-05

Sep-05

Dec-05

Mar-06

Jun-06

System

Po

wer (

kW

)

Avg. Power Top5

Avg. Power Top50

Page 5: Power Efficiency Metrics for the Top500

Broad Objective

Use Top500 List to track power efficiencytrends

Raise Community Awareness of HPCSystem Power Efficiency

Push vendors toward more power efficientsolutions by providing a venue to comparetheir power consumption

Page 6: Power Efficiency Metrics for the Top500

Specific Proposal

Require all Top500 sites to report system powerconsumption under a LINPACK workload Establish “rules of engagement” to govern “fair” collection

of the data

Must establish data collection procedures to make the datacollection easy and have little impact on center operations

We wish to convince you that this can be done withminimal pain If it cannot, we want to hear your input so that the rules can

be drafted in a way that will work for ALL Top500respondents

Page 7: Power Efficiency Metrics for the Top500

What do we mean by “Easy”

We will try to prove to you that one can measurefrom a single node or cabinet and project the powerconsumption for the overall system

At cabinet and node level, You do not have to take your entire system out of service

to measure power under LINPACK

sample a few representative pieces running proportionallysmaller copies of LINPACK and project it to the full systemscale

Can use simpler/less-expensive power measurementapparatus

Page 8: Power Efficiency Metrics for the Top500

Many Ways to Measure Power Clamp meters

+: easy to use, don’t need to disconnect test systems, widevariety of voltages

-: very inaccurate for more than one wire

Inline meters +: accurate, easy to use, can output over serial -: must disconnect test system, limited voltage, limited current

Power panels / PDU panels Unknown accuracy, must stand and read, usually coarse-grained

(unable to differentiate power loads) Sometimes the best or only option: can get numbers for an entire

HPC system

Integrated monitoring in system power supplies (Cray XT) +: accurate, easy to use - : only measures single cabinet. Must know power supply

conversion efficiency to project nominal power use at wall socket

Page 9: Power Efficiency Metrics for the Top500

Testing our Methodology Look at power usage using variety of synthetic and

real benchmarks Memory intensive : STREAM

CPU intensive: HPL/Linpack

IO intensive: IOZone, MADbench

Simulated workloads: NAS PB, NERSC SSP

Compare single node vs cabinet/cluster vs entiresystem Is power consumed when running LINPACK similar to that of

a real workload?

Does power consumed by LINPACK change withconcurrency?

Can we predict full system power from cabinet/node power?

Page 10: Power Efficiency Metrics for the Top500

Single Node Tests: AMD Opteron

Highest power usage is 2x NAS FT and LU

Page 11: Power Efficiency Metrics for the Top500

Similar Results when TestingOther CPU Architectures

Power consumption far less than manufacturer’estimated “nameplate power”

Idle power much lower than active power Power consumption when running LINPACK is very

close to power consumed when running othercompute intensive applications

Core Duo AMD Opteron IBM PPC970/G5

Page 12: Power Efficiency Metrics for the Top500

Entire System Power Usage

0

200

400

600

800

1000

1200

1400

Time-->

Kil

ow

att

s

Catamount

CNL

Full System Test

Tests run across all 19,353 compute cores

Throughput: NERSC “realistic” workload composed of full applications

idle() loop allows powersave on unused processors; (generally moreefficient)

STREAM HPL Throughput

No idle()Idle() loop

Page 13: Power Efficiency Metrics for the Top500

Single Rack Tests

Administrative utility gives rack DC amps & voltage HPL & Paratec are highest power usage

Single Cabinet Power Usage

0

50

100

150

200

250

300

Time -->

Am

ps (

@ 5

2 V

DC

)

STREAM GTC FT.D LU.D CG.D paratec pmemd HPL

Page 14: Power Efficiency Metrics for the Top500

Modeling the Entire System:AC to DC Conversion

Commodity desktop machines are ~75%efficient

Google uses new, 90% efficient powersupplies

Our test system has has ~90% efficientpower supplies

Page 15: Power Efficiency Metrics for the Top500

Modeling the Entire System

Error factor is 0.05 if we assume 90% efficiency

Full System Power Usage, Model Using Actual vs. Single Rack x Num Racks

0.00E+00

2.00E+05

4.00E+05

6.00E+05

8.00E+05

1.00E+06

1.20E+06

1.40E+06

idle STREAM HPL

Wat

ts

Actual, Watts AC

90% Eff

80% Eff

Model, with FS (est. 50 MW)

Page 16: Power Efficiency Metrics for the Top500

Conclusions Power utilization under an HPL/Linpack load is a good

estimator for power usage under mixed workloads for singlenodes, cabinets / clusters, and large scale systems Idle power is not Nameplate and CPU power are not

LINPACK running on one node or rack consumes approxmimatelysame power as the node would consume if it were part of full-sysparallel LINPACK job

We can estimate overall power usage using a subset of theentire HPC system and extrapolating to total number of nodesusing a variety of power measurement techniques And the estimates mostly agree with one-another!

Disk subsystem is a small fraction of overall power (50-60KWvs 1,200 KW) Disk power dominated by spindles and power supplies Idle power for disks not significantly different from active power

Page 17: Power Efficiency Metrics for the Top500

Top500 Power Data Collection Measure System Power when running LINPACK

If measured on circuit for full system (eg. PDU or Panel),then run LINPACK on full system

If cannot differentiate system from other devices sharingsame circuit, then isolate components using line meter orinductive clamp meter (can borrow from local Power Co.)

If measured from line meter or clamp meter, then just runon one rack, measure representative componentscomprising system and extrapolate to entire system

If measured from integral power supply, account for powersupply losses in projections

Target of projections is RMS AC “wall-socket power”consumed by HPC system Must convert measurements of DC power consumption

accordingly

Page 18: Power Efficiency Metrics for the Top500

Top500 Power Data Collection

What to include All components comprising delivered HPC system aside

from external disk subsystem (eg. SAN)

Can extrapolate from measurement of said componentswhile under a LINPACK load (even if load is local)

What to exclude Exclude cooling

Exclude PDU and other power conversion infrastructurelosses that are not part of the deliverd HPC system

Exclude disk subsystem (if not integral): should discussthis further

Page 19: Power Efficiency Metrics for the Top500

Complementary Efforts

Our Effort creates a metric for compute-intensiveparallel scientific workloads

Metrics for I/O intensive workloads: JouleSort by HPLabs

Metrics for transactional workloads: EPA EnergyStarServer Metrics

Re-ranking of Top500 for Power Efficiency:http://www.green500.org/

Page 20: Power Efficiency Metrics for the Top500

Single Node Tests: IO

Highly variable, less than compute-only

Very difficult to assess power draw for I/O

Page 21: Power Efficiency Metrics for the Top500

Modeling the Entire System:Disks

Must take into account disk subsystem

Drive model matters Deskstar 9.6W idle, 13.6W under load

Tonka 7.4W idle, 12.6W under load

Using DDN-provided numbers, estimatedpower draw for model disk subsystem is50KW idle, 60KW active

Observed using PDU panel: ~48KW idle