1 intro to multiprocessors [adapted from computer organization and design, patterson & hennessy,...
TRANSCRIPT
.1
Intro to Multiprocessors
[Adapted from Computer Organization and Design, Patterson & Hennessy, © 2005]
.2
The Big Picture: Where are We Now?
Processor
Control
Datapath
Memory
Input
Output
Input
Output
Memory
Processor
Control
Datapath
Multiprocessor – multiple processors with a single shared address space
Cluster – multiple computers (each with their own address space) connected over a local area network (LAN) functioning as a single system
.3
Applications Needing “Supercomputing” Energy (plasma physics (simulating fusion reactions),
geophysical (petroleum) exploration) DoE stockpile stewardship (to ensure the safety and
reliability of the nation’s stockpile of nuclear weapons) Earth and climate (climate and weather prediction,
earthquake, tsunami prediction and mitigation of risks) Transportation (improving vehicles’ airflow dynamics, fuel
consumption, crashworthiness, noise reduction) Bioinformatics and computational biology (genomics,
protein folding, designer drugs) Societal health and safety (pollution reduction, disaster
planning, terrorist action detection)
.5
Encountering Amdahl’s Law
Speedup due to enhancement E is
Speedup w/ E = ---------------------- Exec time w/o E
Exec time w/ E
Suppose that enhancement E accelerates a fraction F (F <1) of the task by a factor S (S>1) and the remainder of the task is unaffected
ExTime w/ E = ExTime w/o E ((1-F) + F/S)
Speedup w/ E = 1 / ((1-F) + F/S)
.7
Examples: Amdahl’s Law
Consider an enhancement which runs 20 times faster but which is only usable 25% of the time.
Speedup w/ E = 1/(.75 + .25/20) = 1.31
What if its usable only 15% of the time?Speedup w/ E = 1/(.85 + .15/20) = 1.17
Amdahl’s Law tells us that to achieve linear speedup with 100 processors, none of the original computation can be scalar!
To get a speedup of 99 from 100 processors, the percentage of the original program that could be scalar would have to be 0.01% or less
Speedup w/ E = 1 / ((1-F) + F/S)
.8
Supercomputer Style Migration (Top500)
In the last 8 years uniprocessor and SIMDs disappeared while Clusters and Constellations grew from 3% to 80%
0
100
200
300
400
500Clusters
Constellations
SIMDs
MPPs
SMPs
Uniproc's
Nov data
http://www.top500.org/lists/2005/11/
Cluster – whole computers interconnected using their I/O bus
Constellation – a cluster that uses an SMP multiprocessor as the building block
.9
Multiprocessor/Clusters Key Questions
Q1 – How do they share data?
Q2 – How do they coordinate?
Q3 – How scalable is the architecture? How many processors can be supported?
.10
Flynn’s Classification Scheme
Now obsolete except for . . .
SISD – single instruction, single data stream aka uniprocessor - what we have been talking about all semester
SIMD – single instruction, multiple data streams single control unit broadcasting operations to multiple datapaths
MISD – multiple instruction, single data no such machine (although some people put vector machines in
this category)
MIMD – multiple instructions, multiple data streams aka multiprocessors (SMPs, MPPs, clusters, NOWs)
.11
SIMD Processors
Single control unit Multiple datapaths (processing elements – PEs) running
in parallel Q1 – PEs are interconnected (usually via a mesh or torus) and
exchange/share data as directed by the control unit Q2 – Each PE performs the same operation on its own local data
PE
PE
PE
PE PE
PE
PE
PE PE
PE
PE
PE PE
PE
PE
PE
Control
.12
Example SIMD Machines
Maker Year # PEs # b/ PE
Max memory
(MB)
PE clock (MHz)
System BW
(MB/s)
Illiac IV UIUC 1972 64 64 1 13 2,560
DAP ICL 1980 4,096 1 2 5 2,560
MPP Goodyear 1982 16,384 1 2 10 20,480
CM-2 Thinking Machines
1987 65,536 1 512 7 16,384
MP-1216 MasPar 1989 16,384 4 1024 25 23,000
.13
Multiprocessor Basic Organizations
Processors connected by a single bus Processors connected by a network
# of Proc
Communication model
Message passing 8 to 2048
Shared address
NUMA 8 to 256
UMA 2 to 64
Physical connection
Network 8 to 256
Bus 2 to 36
.14
Shared Address (Shared Memory) Multi’s
UMAs (uniform memory access) – aka SMP (symmetric multiprocessors)
all accesses to main memory take the same amount of time no matter which processor makes the request or which location is requested
NUMAs (nonuniform memory access) some main memory accesses are faster than others depending on
the processor making the request and which location is requested can scale to larger sizes than UMAs so are potentially higher
performance
Q1 – Single address space shared by all the processors Q2 – Processors coordinate/communicate through shared
variables in memory (via loads and stores) Use of shared data must be coordinated via synchronization
primitives (locks)
.15
N/UMA Remote Memory Access Times (RMAT)
Year Type Max Proc
Interconnection Network
RMAT (ns)
Sun Starfire 1996 SMP 64 Address buses, data switch
500
Cray 3TE 1996 NUMA 2048 2-way 3D torus 300
HP V 1998 SMP 32 8 x 8 crossbar 1000
SGI Origin 3000 1999 NUMA 512 Fat tree 500
Compaq AlphaServer GS
1999 SMP 32 Switched bus 400
Sun V880 2002 SMP 8 Switched bus 240
HP Superdome 9000
2003 SMP 64 Switched bus 275
NASA Columbia 2004 NUMA 10240 Fat tree ???
.16
Single Bus (Shared Address UMA) Multi’s
Caches are used to reduce latency and to lower bus traffic Must provide hardware to ensure that caches and memory
are consistent (cache coherency) Must provide a hardware mechanism to support process
synchronization
Processor Processor Processor
Cache Cache Cache
Single Bus
Memory I/O
.17
Message Passing Multiprocessors
Each processor has its own private address space Q1 – Processors share data by explicitly sending and
receiving information (messages) Q2 – Coordination is built into message passing
primitives (send and receive)
.18
Summary
Flynn’s classification of processors – SISD, SIMD, MIMD Q1 – How do processors share data? Q2 – How do processors coordinate their activity? Q3 – How scalable is the architecture (what is the maximum
number of processors)?
Shared address multis – UMAs and NUMAs Bus connected (shared address UMAs) multis
Cache coherency hardware to ensure data consistency Synchronization primitives for synchronization Bus traffic limits scalability of architecture (< ~ 36 processors)
Message passing multis
.19
Review: Where are We Now?
Processor
Control
Datapath
Memory
Input
Output
Input
Output
Memory
Processor
Control
Datapath
Multiprocessor – multiple processors with a single shared address space
Cluster – multiple computers (each with their own address space) connected over a local area network (LAN) functioning as a single system
.20
Multiprocessor Basics
# of Proc
Communication model
Message passing 8 to 2048
Shared address
NUMA 8 to 256
UMA 2 to 64
Physical connection
Network 8 to 256
Bus 2 to 36
Q1 – How do they share data?
Q2 – How do they coordinate?
Q3 – How scalable is the architecture? How many processors?
.21
Single Bus (Shared Address UMA) Multi’s
Caches are used to reduce latency and to lower bus traffic Write-back caches used to keep bus traffic at a minimum
Must provide hardware to ensure that caches and memory are consistent (cache coherency)
Must provide a hardware mechanism to support process synchronization
Proc1 Proc2 Proc4
Caches Caches Caches
Single Bus
Memory I/O
Proc3
Caches
.22
Multiprocessor Cache Coherency Cache coherency protocols
Bus snooping – cache controllers monitor shared bus traffic with duplicate address tag hardware (so they don’t interfere with processor’s access to the cache)
Proc1 Proc2 ProcN
DCache DCache DCache
Single Bus
Memory I/O
Snoop Snoop Snoop
.23
Bus Snooping Protocols Multiple copies are not a problem when reading Processor must have exclusive access to write a word
What happens if two processors try to write to the same shared data word in the same clock cycle? The bus arbiter decides which processor gets the bus first (and this will be the processor with the first exclusive access). Then the second processor will get exclusive access. Thus, bus arbitration forces sequential behavior.
This sequential consistency is the most conservative of the memory consistency models. With it, the result of any execution is the same as if the accesses of each processor were kept in order and the accesses among different processors were interleaved.
All other processors sharing that data must be informed of writes
.24
Handling Writes
Ensuring that all other processors sharing data are informed of writes can be handled two ways:
1. Write-update (write-broadcast) – writing processor broadcasts new data over the bus, all copies are updated All writes go to the bus higher bus traffic Since new values appear in caches sooner, can reduce latency
2. Write-invalidate – writing processor issues invalidation signal on bus, cache snoops check to see if they have a copy of the data, if so they invalidate their cache block containing the word (this allows multiple readers but only one writer) Uses the bus only on the first write lower bus traffic, so better
use of bus bandwidth
.26
A Write-Invalidate CC Protocol
Shared(clean)
Invalid
Modified(dirty)
write-back caching protocol in black
read (miss)
write (h
it or m
iss)
read (hit or miss)
read (hit) or write (hit or miss)
writ
e (m
iss)
send in
validate
receives invalidate(write by another processor
to this block)w
rite-
back
due
to
read
(m
iss)
by
anot
her
proc
esso
r to
thi
s bl
ock
send
inva
lidat
e
signals from the processor coherence additions in redsignals from the bus coherence additions in blue
State machinefor CPU requests for cache block
State machine for bus requests for each cache block
.28
Write-Invalidate CC Examples I = invalid (many), S = shared (many), M = modified (only one)
Proc 1
A S
Main Mem A
Proc 2
A I
1. read miss for A
2. read request for A
3. snoop sees read request for
A & lets MM supply A
4. gets A from MM & changes its state
to S
Proc 1
A S
Main Mem A
Proc 2
A I
1. write miss for A
2. writes A & changes its state
to M
Proc 1
A M
Main Mem A
Proc 2
A I
1. read miss for A3. snoop sees read request for A, writes-
back A to MM
2. read request for A
4. gets A from MM & changes its state
to S
3. P2 sends invalidate for A
4. change A state to I
5. change A state to S
Proc 1
A M
Main Mem A
Proc 2
A I
1. write miss for A
2. writes A & changes its state
to M
3. P2 sends invalidate for A
4. change A state to I
.29
Block Size Effects
Writes to one word in a multi-word block mean either the full block is invalidated (write-invalidate) or the full block is exchanged between processors (write-update)
- Alternatively, could broadcast only the written word
Multi-word blocks can also result in false sharing: when two processors are writing to two different variables in the same cache block
With write-invalidate false sharing increases cache miss rates
Compilers can help reduce false sharing by allocating highly correlated data to the same cache block
A B
Proc1 Proc2
4 word cache block
.30
Other Coherence Protocols
There are many variations on cache coherence protocols
Another write-invalidate protocol used in the Pentium 4 (and many other micro’s) is MESI with four states:
Modified – same Exclusive – only one copy of the shared data is allowed to be
cached; memory has an up-to-date copy- Since there is only one copy of the block, write hits don’t need to
send invalidate signal
Shared – multiple copies of the shared data may be cached (i.e., data permitted to be cached with more than one processor); memory has an up-to-date copy
Invalid – same
.32
Commercial Single Backplane Multiprocessors
Processor # proc. MHz BW/ system
Compaq PL Pentium Pro 4 200 540
IBM R40 PowerPC 8 112 1800
AlphaServer Alpha 21164 12 440 2150
SGI Pow Chal MIPS R10000 36 195 1200
Sun 6000 UltraSPARC 30 167 2600
.33
Summary Key questions
Q1 - How do processors share data? Q2 - How do processors coordinate their activity? Q3 - How scalable is the architecture (what is the maximum
number of processors)?
Bus connected (shared address UMA’s(SMP’s)) multi’s Cache coherency hardware to ensure data consistency Synchronization primitives for synchronization Scalability of bus connected UMAs limited (< ~ 36 processors)
because the three desirable bus characteristics- high bandwidth
- low latency
- long length
are incompatible
Network connected NUMAs are more scalable