scalable distributed memory machines

39
EECC756 - Shaaban EECC756 - Shaaban #1 lec # 13 Spring2002 5-2 Scalable Scalable Distributed Memory Distributed Memory Machines Machines Goal: Parallel machines that can be scaled to hundreds or thousands of processors. Design Choices: Custom-designed or commodity nodes? Network scalability. Capability of node-to-network interface (critical). Supporting programming models? What does hardware scalability mean? Avoids inherent design limits on resources. Bandwidth increases with machine size P. Latency should not increase with machine size P. Cost should increase slowly with P.

Upload: norman-bradford

Post on 03-Jan-2016

49 views

Category:

Documents


0 download

DESCRIPTION

Scalable Distributed Memory Machines. Goal: Parallel machines that can be scaled to hundreds or thousands of processors. Design Choices: Custom-designed or commodity nodes? Network scalability. Capability of node-to-network interface ( critical ). Supporting programming models? - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#1 lec # 13 Spring2002 5-2-2002

Scalable Distributed Scalable Distributed Memory MachinesMemory Machines

Goal: Parallel machines that can be scaled to hundreds or thousands of processors.

• Design Choices:– Custom-designed or commodity nodes?– Network scalability. – Capability of node-to-network interface (critical).– Supporting programming models?

• What does hardware scalability mean?– Avoids inherent design limits on resources.– Bandwidth increases with machine size P.– Latency should not increase with machine size P.– Cost should increase slowly with P.

Page 2: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#2 lec # 13 Spring2002 5-2-2002

MPPs Scalability IssuesMPPs Scalability Issues• Problems:

– Memory-access latency.– Interprocess communication complexity or synchronization

overhead.– Multi-cache inconsistency.– Message-passing and message processing overheads.

• Possible Solutions:– Fast dedicated, proprietary and scalable, networks and protocols.– Low-latency fast synchronization techniques possibly hardware-assisted .– Hardware-assisted message processing in communication assists (node-to-

network interfaces). – Weaker memory consistency models.– Scalable directory-based cache coherence protocols.– Shared virtual memory.– Improved software portability; standard parallel and distributed

operating system support.– Software latency-hiding techniques.

Page 3: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#3 lec # 13 Spring2002 5-2-2002

One Extreme:One Extreme: Limited Scaling of a BusLimited Scaling of a Bus

• Bus: Each level of the system design is grounded in the scaling limits at the layers below and assumptions of close coupling between components.

Characteristic Bus

Physical Length ~ 1 ft

Number of Connections fixed

Maximum Bandwidth fixed

Interface to Comm. medium memory inf

Global Order arbitration

Protection Virt -> physical

Trust total

OS single

comm. abstraction HW

Poor Scalability

Page 4: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#4 lec # 13 Spring2002 5-2-2002

Another Extreme:Another Extreme:

ScalingScaling of Workstations in a LAN? of Workstations in a LAN?

• No clear limit to physical scaling, no global order, consensus difficult to achieve.

Characteristic Bus LAN

Physical Length ~ 1 ft KM

Number of Connections fixed many

Maximum Bandwidth fixed ???

Interface to Comm. medium memory inf peripheral

Global Order arbitration ???

Protection Virt -> physical OS

Trust total none

OS single independent

comm. abstraction HW SW

Page 5: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#5 lec # 13 Spring2002 5-2-2002

• Depends largely on network characteristics:– Channel bandwidth.– Static: Topology: Node degree, Bisection width etc.– Multistage: Switch size and connection pattern properties.– Node-to-network interface capabilities.

Bandwidth ScalabilityBandwidth Scalability

P M M P M M P M M P M M

S S S S

Typical switches

Bus

Multiplexers

Crossbar

Page 6: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#6 lec # 13 Spring2002 5-2-2002

Dancehall MP OrganizationDancehall MP Organization

• Network bandwidth?• Bandwidth demand?

– Independent processes?– Communicating processes?

• Latency?

Scalable network

P

$

Switch

M

P

$

P

$

P

$

M M

Switch Switch

Extremely high demands on network in terms ofbandwidth, latency even forindependent processes.

Page 7: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#7 lec # 13 Spring2002 5-2-2002

Generic Distributed Memory Generic Distributed Memory OrganizationOrganization

• Network bandwidth?• Bandwidth demand?

– Independent processes?– Communicating processes?

• Latency? O(log2P) increase?

• Cost scalability of system?

Scalable network

CA

P

$

Switch

M

Switch Switch

Multi-stageinterconnection network (MIN)?Custom-designed?

Node:O(10) Bus-based SMP

Custom-designed CPU?Node/System integration level?How far? Cray-on-a-Chip? SMP-on-a-Chip?

OS Supported?Network protocols?

Communication AssistExtent of functionality?

MessagetransactionDMA?

Global virtual Shared address space?

Page 8: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#8 lec # 13 Spring2002 5-2-2002

Key System Scaling PropertyKey System Scaling Property

• Large number of independent communication paths between nodes.

=> Allow a large number of concurrent transactions using different channels.

• Transactions are initiated independently.

• No global arbitration.

• Effect of a transaction only visible to the nodes involved– Effects propagated through additional transactions.

Page 9: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#9 lec # 13 Spring2002 5-2-2002

Network Latency ScalingNetwork Latency Scaling

• T(n) = Overhead + Channel Time + Routing Delay

• Scaling of overhead?

• Channel Time(n) = n/B --- BW at bottleneck

• RoutingDelay(h,n)

Page 10: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#10 lec # 13 Spring2002 5-2-2002

Network Latency Scaling ExampleNetwork Latency Scaling Example

• Max distance: log2 n

• Number of switches: n log n

• overhead = 1 us, BW = 64 MB/s, 200 ns per hop

• Using pipelined or cut-through routing:• T64(128) = 1.0 us + 2.0 us + 6 hops * 0.2 us/hop = 4.2 us

• T1024(128) = 1.0 us + 2.0 us + 10 hops * 0.2 us/hop = 5.0 us

• Store and Forward• T64

sf(128) = 1.0 us + 6 hops * (2.0 + 0.2) us/hop = 14.2 us

• T1024sf(128) = 1.0 us + 10 hops * (2.0 + 0.2) us/hop = 23 us

O(log2 n) Stage MIN using switches:

Only 20% increase in latency for 16x size increase

Page 11: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#11 lec # 13 Spring2002 5-2-2002

Cost ScalingCost Scaling• cost(p,m) = fixed cost + incremental cost (p,m)

• Bus Based SMP?

• Ratio of processors : memory : network : I/O ?

• Parallel efficiency(p) = Speedup(P) / P

• Similar to speedup, one can define:

Costup(p) = Cost(p) / Cost(1)

• Cost-effective: speedup(p) > costup(p)

Page 12: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#12 lec # 13 Spring2002 5-2-2002

Cost Effective?Cost Effective?

2048 processors: 475 fold speedup at 206x cost

0

500

1000

1500

2000

0 500 1000 1500 2000

Processors

Speedup = P/(1+ logP)

Costup = 1 + 0.1 P

Page 13: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#13 lec # 13 Spring2002 5-2-2002

Parallel Machine Network ExamplesParallel Machine Network Examples

Page 14: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#14 lec # 13 Spring2002 5-2-2002

Physical ScalingPhysical Scaling• Chip-level integration:

– Integrate network interface, message router I/O links.• nCUBE/2, Alpha 21364, IBM Power 4

– IRAM-style Cray-on-a-Chip: V-IRAM

– Memory/Bus controller/chip set: Alpha 21364

– SMP on a chip: Chip Multiprocessor (CMP): IBM Power 4

• Board-level:– Replicating using standard microprocessor cores.

• CM-5 replicated the core of a Sun SparkStation 1 workstation.• Cray T3D and T3E replicated the core of a DEC Alpha workstation.

• System level:• IBM SP-2 uses 8-16 almost complete RS6000 workstations placed in racks.

Page 15: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#15 lec # 13 Spring2002 5-2-2002

Chip-level integration Example:Chip-level integration Example: nCUBE/2 Machine OrganizationnCUBE/2 Machine Organization

• Entire machine synchronous at 40 MHz

Single-chip node

Basic module

Hypercube networkconfiguration

DRAM interface

DM

Ach

anne

ls

Ro

ute

r

MMU

I-Fetch&

decode

64-bit integerIEEE floating point

Operand$

Execution unit

500, 000 transistors (considered large at the time)

64 nodes socketed on a board

13 linksup to 8096nodes possible

Page 16: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#16 lec # 13 Spring2002 5-2-2002

Chip-level integration Example:Chip-level integration Example: Vector Intelligent RAM 2 (Vector Intelligent RAM 2 (V-IRAM-2)V-IRAM-2)

Memory Crossbar Switch

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

M

+

Vector Registers

x

÷

Load/Store

8K I cache 8K D cache

2-way Superscalar VectorProcessor

8 x 64 8 x 64 8 x 64 8 x 64 8 x 64

8 x 64or

16 x 32or

32 x 16

8 x 648 x 64

QueueInstruction

I/OI/O

I/OI/O

Projected 2003. < 0.1 µm, > 2 GHz Projected 2003. < 0.1 µm, > 2 GHz 16 GFLOPS(64b)/64 GOPS(16b)/128MB16 GFLOPS(64b)/64 GOPS(16b)/128MB

Page 17: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#17 lec # 13 Spring2002 5-2-2002

Alpha 21264 core with enhancementsAlpha 21264 core with enhancements Integrated Direct RAMbus memory controller:Integrated Direct RAMbus memory controller:

800 MHz operation, 30ns CAS latency pin to pin, 6 GB/sec read or write bandwidth Directory based cache coherence

Integrated network interface:Integrated network interface: Direct processor-to-processor interconnect, 10 GB/second per processor 15ns processor-to-processor latency, Out-of-order network with adaptive routing Asynchronous clocking between processors, 3 GB/second I/O interface per processor

Chip-level integration Example:Chip-level integration Example: Alpha 21364Alpha 21364

MemoryController

RAMBUS

21264Core

16 L1Miss Buffers

L2Cache

Address Out

Address In

NetworkInterface

NorthSouthEastWestI/O

16 L1Victim Buf 16 L2

Victim Buf

64K Icache

64K Dcache

Page 18: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#18 lec # 13 Spring2002 5-2-2002

Chip-level integration Example:Chip-level integration Example: A Possible Alpha 21364 System A Possible Alpha 21364 System

364M

IO364

M

IO364

M

IO364

M

IO

364M

IO364

M

IO364

M

IO364

M

IO

364M

IO364

M

IO364

M

IO364

M

IO

Page 19: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#19 lec # 13 Spring2002 5-2-2002

• Two tightly-integrated 1GHz CPU cores per 170 Million Transistor chip.

• 128KB L1 Cache per processor

• 1.5 MB On-Chip Shared L2 Cache

• External 32MB L3 Cache: Tags kept on chip.

• 35 Gbytes/s Chip-to-Chip interconnects.

Chip-level integration Example:Chip-level integration Example: IBM Power 4 CMP IBM Power 4 CMP

Page 20: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#20 lec # 13 Spring2002 5-2-2002

Chip-level integration Example:Chip-level integration Example: IBM Power 4 IBM Power 4

Page 21: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#21 lec # 13 Spring2002 5-2-2002

Chip-level integration Example:Chip-level integration Example: IBM Power 4 MCM IBM Power 4 MCM

Page 22: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#22 lec # 13 Spring2002 5-2-2002

Board-level integration Example:Board-level integration Example: CM-5 Machine OrganizationCM-5 Machine Organization

Diagnostics network

Control network

Data network

Processingpartition

Processingpartition

Controlprocessors

I/O partition

PM PM

SPARC

MBUS

DRAMctrl

DRAM DRAM DRAM DRAM

DRAMctrl

Vectorunit DRAM

ctrlDRAM

ctrl

Vectorunit

FPU Datanetworks

Controlnetwork

$ctrl

$SRAM

NI

Design replicated the core of a Sun SparkStation 1 workstation

Fat Tree

Page 23: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#23 lec # 13 Spring2002 5-2-2002

Memory bus

MicroChannel bus

I/O

i860 NI

DMA

DR

AM

IBM SP-2 node

L2 $

Power 2CPU

Memorycontroller

4-wayinterleaved

DRAM

General interconnectionnetwork formed from8-port switches

NIC

System Level Integration Example:System Level Integration Example:

IBM SP-2IBM SP-2

8-16 almostcompleteRS6000workstationsplaced in racks.

Page 24: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#24 lec # 13 Spring2002 5-2-2002

Realizing Programming Models:Realizing Programming Models:

Realized by ProtocolsRealized by Protocols

CAD

Multiprogramming Sharedaddress

Messagepassing

Dataparallel

Database Scientific modeling Parallel applications

Programming models

Communication abstractionUser/system boundary

Compilationor library

Operating systems support

Communication hardware

Physical communication medium

Hardware/software boundary

Network Transactions

Page 25: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#25 lec # 13 Spring2002 5-2-2002

Challenges in Realizing Prog. Models Challenges in Realizing Prog. Models in Large-Scale Machinesin Large-Scale Machines

• No global knowledge, nor global control.

– Barriers, scans, reduce, global-OR give fuzzy global state.

• Very large number of concurrent transactions.

• Management of input buffer resources:

– Many sources can issue a request and over-commit destination before any see the effect.

• Latency is large enough that one is tempted to “take risks”:

– Optimistic protocols.

– Large transfers.

– Dynamic allocation.

• Many more degrees of freedom in design and engineering of these system.

Page 26: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#26 lec # 13 Spring2002 5-2-2002

Network Transaction ProcessingNetwork Transaction Processing

Key Design Issue:

• How much interpretation of the message by CA without involving the CPU?

• How much dedicated processing in the CA?

PM

CA

PM

CA° ° °

Scalable Network

Node Architecture

Communication Assist

Message

Output Processing – checks – translation – formating – scheduling

Input Processing – checks – translation – buffering – action

CA = Communication Assist

Page 27: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#27 lec # 13 Spring2002 5-2-2002

Spectrum of DesignsSpectrum of DesignsNone: Physical bit stream

– blind, physical DMA nCUBE, iPSC, . . .

User/System– User-level port CM-5, *T

– User-level handler J-Machine, Monsoon, . . .

Remote virtual address– Processing, translation Paragon, Meiko CS-2

Global physical address– Proc + Memory controller RP3, BBN, T3D

Cache-to-cache– Cache controller Dash, KSR, Flash

Incr

easi

ng

HW

Su

pp

ort

, S

pec

iali

zati

on

, In

tru

sive

nes

s, P

erfo

rman

ce

(??

?)

Page 28: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#28 lec # 13 Spring2002 5-2-2002

No CA Net Transactions Interpretation:No CA Net Transactions Interpretation: Physical DMA Physical DMA

• DMA controlled by regs, generates interrupts.• Physical => OS initiates transfers.• Send-side:

– Construct system “envelope” around user data in kernel area.• Receive:

– Must receive into system buffer, since no message interpretation in CA.

PMemory

Cmd

DestData

Addr

Length

Rdy

PMemory

DMAchannels

Status,interrupt

Addr

Length

Rdy

sender auth

dest addr

Page 29: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#29 lec # 13 Spring2002 5-2-2002

nCUBE/2 Network InterfacenCUBE/2 Network Interface

• Independent DMA channel per link direction

– Leave input buffers always open.

– Segmented messages.• Routing determines if message is intended for local or remote node

– Dimension-order routing on hypercube.

– Bit-serial with 36 bit cut-through.

Processor

Switch

Input ports

Output ports

Memory

Addr AddrLength

Addr Addr AddrLength

AddrLength

DMAchannels

Memorybus

Os 16 ins 260 cy13 us

Or 18 200 cy15 us

- includes interrupt

Page 30: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#30 lec # 13 Spring2002 5-2-2002

DMA In Conventional LAN DMA In Conventional LAN Network InterfacesNetwork Interfaces

NIC Controller

DMAaddr

len

trncv

TX

RX

Addr LenStatusNext

Addr LenStatusNext

Addr LenStatusNext

Addr LenStatusNext

Addr LenStatusNext

Addr LenStatusNext

Data

Host Memory NIC

IO Busmem bus

Proc

Page 31: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#31 lec # 13 Spring2002 5-2-2002

User-Level PortsUser-Level Ports

• Initiate transaction at user level.• CA interprets delivers message to user without OS intervention.• Network port in user space.• User/system flag in envelope.

– Protection check, translation, routing, media access in source CA– User/sys check in destination CA, interrupt on system.

PMem

DestData

User/system

PMemStatus,interrupt

Page 32: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#32 lec # 13 Spring2002 5-2-2002

User-Level Network Example:User-Level Network Example: CM-5 CM-5• Two data networks and

one control network.

• Input and output FIFO for each network.

• Tag per message:

– Index Network Inteface (NI) mapping table.

• *T integrated NI on chip.

• Also used in iWARP.

Diagnostics network

Control network

Data network

Processingpartition

Processingpartition

Controlprocessors

I/O partition

PM PM

SPARC

MBUS

DRAMctrl

DRAM DRAM DRAM DRAM

DRAMctrl

Vectorunit DRAM

ctrlDRAM

ctrl

Vectorunit

FPU Datanetworks

Controlnetwork

$ctrl

$SRAM

NI

Os 50 cy 1.5 us

Or 53 cy 1.6 us

interrupt 10us

Page 33: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#33 lec # 13 Spring2002 5-2-2002

User-Level HandlersUser-Level Handlers

• Tighter integration of user-level network port with the processor at the register level.• Hardware support to vector to address specified in message

– message ports in registers.

U ser /sy s te m

PM e m

D e stD ata A d dress

PM e m

Page 34: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#34 lec # 13 Spring2002 5-2-2002

iWARPiWARP

• Nodes integrate communication with computation on systolic basis.

• Message data direct to register.

• Stream into memory.

Interface unit

Host

Page 35: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#35 lec # 13 Spring2002 5-2-2002

Dedicated Message Processing Without Dedicated Message Processing Without Specialized Hardware DesignSpecialized Hardware Design

General Purpose processor performs arbitrary output processing (at system level)

General Purpose processor interprets incoming network transactions (at system level)

User Processor <–> Message Processor share memory

Message Processor <–> Message Processor via system network transaction

Network

° ° ° 

dest

Mem

P M P

NI

User System

Mem

P M P

NI

User System

Node:Bus-basedSMP

MPMessageProcessor

Page 36: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#36 lec # 13 Spring2002 5-2-2002

• User Processor stores cmd / msg / data into shared output queue.– Must still check for output queue full (or make elastic).

• Communication assists make transaction happen.– Checking, translation, scheduling, transport, interpretation.

• Effect observed on destination address space and/or events.

• Protocol divided between two layers.

Network

° ° ° 

dest

Mem

P M P

NI

User System

Mem

PM P

NI

Levels of Network TransactionLevels of Network Transaction

Page 37: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#37 lec # 13 Spring2002 5-2-2002

Example: Intel ParagonExample: Intel ParagonNetwork

° ° ° Mem

P M P

NIi860xp50 MHz16 KB $4-way32B BlockMESI

sDMArDMA

64400 MB/s

$ $

16 175 MB/s Duplex

I/ONodes

rteMP handler

Var dataEOP

I/ONodes

Service

Devices

Devices

2048 B

Page 38: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#38 lec # 13 Spring2002 5-2-2002

Dispatcher

User OutputQueues

Send FIFO~Empty

Rcv FIFO~Full

Send DMA

Rcv DMA

DMA done

ComputeProcessorKernel

SystemEvent

Message Processor EventsMessage Processor Events

Page 39: Scalable Distributed Memory Machines

EECC756 - ShaabanEECC756 - Shaaban#39 lec # 13 Spring2002 5-2-2002

• Concurrency Intensive

– Need to keep inbound flows moving while outbound flows stalled.

– Large transfers segmented.

• Reduces overhead but adds latency.

User OutputQueues

Send FIFO~Empty

Rcv FIFO~Full

Send DMA

Rcv DMA

DMA done

ComputeProcessorKernel

SystemEvent

User InputQueues

VAS

Dispatcher

Message Processor AssessmentMessage Processor Assessment