introduction to multicore

Upload: bhargav-tangirala

Post on 05-Apr-2018

228 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/2/2019 Introduction to Multicore

    1/55

    Multicores, Multiprocessors, andClusters

  • 8/2/2019 Introduction to Multicore

    2/55

    Introduction

    Goal: connecting multiple computersto get higher performance Multiprocessors

    Scalability, availability, power efficiency

    Job-level (process-level) parallelism High throughput for independent jobs

    Parallel processing program

    Single program run on multiple processors Multicore microprocessors

    Chips with multiple processors (cores)

  • 8/2/2019 Introduction to Multicore

    3/55

    Hardware and Software

    Hardware Serial: e.g., Pentium 4 Parallel: e.g., quad-core Xeon e5345

    Software Sequential: e.g., matrix multiplication

    Concurrent: e.g., operating system

    Sequential/concurrent software can run

    on serial/parallel hardware Challenge: making effective use of parallel

    hardware

  • 8/2/2019 Introduction to Multicore

    4/55

    Parallel Programming

    Parallel software is the problem Need to get significant performance

    improvement

    Otherwise, just use a faster uniprocessor,since its easier!

    Difficulties

    Partitioning

    Coordination

    Communications overhead

  • 8/2/2019 Introduction to Multicore

    5/55

    Amdahls Law

    Sequential part can limit speedup Example: 100 processors, 90

    speedup?

    Tnew = Tparallelizable/100 + Tsequential

    Solving: Fparallelizable = 0.999

    Need sequential part to be 0.1% oforiginal time

    90/100F)F(1

    1Speedupableparallelizableparalleliz

  • 8/2/2019 Introduction to Multicore

    6/55

    Scaling Example

    Workload: sum of 10 scalars, and 10 10matrix sum Speed up from 10 to 100 processors

    Single processor: Time = (10 + 100) tadd 10 processors

    Time = 10 tadd + 100/10 tadd = 20 tadd Speedup = 110/20 = 5.5 (55% of potential)

    100 processors Time = 10 tadd + 100/100 tadd = 11 tadd

    Speedup = 110/11 = 10 (10% of potential) Assumes load can be balanced across

    processors

  • 8/2/2019 Introduction to Multicore

    7/55

    Scaling Example (cont)

    What if matrix size is 100 100?

    Single processor: Time = (10 + 10000) tadd 10 processors

    Time = 10 tadd + 10000/10 tadd = 1010 tadd

    Speedup = 10010/1010 = 9.9 (99% of potential)

    100 processors

    Time = 10 tadd + 10000/100 tadd = 110 tadd

    Speedup = 10010/110 = 91 (91% of potential) Assuming load balanced

  • 8/2/2019 Introduction to Multicore

    8/55

    Strong vs Weak Scaling

    Strong scaling: problem size fixed As in example

    Weak scaling: problem size proportional

    to number of processors 10 processors, 10 10 matrix Time = 20 tadd

    100 processors, 32 32 matrix

    Time = 10 tadd + 1000/100 tadd = 20 tadd

    Constant performance in this example

  • 8/2/2019 Introduction to Multicore

    9/55

    Shared Memory

    SMP: shared memory multiprocessor

    Hardware provides single physicaladdress space for all processors

    Synchronize shared variables using locks

    Memory access time UMA (uniform) vs. NUMA (nonuniform)

  • 8/2/2019 Introduction to Multicore

    10/55

    Example: Sum Reduction

    Sum 100,000 numbers on 100 processorUMA Each processor has ID: 0 Pn 99

    Partition 1000 numbers per processor

    Initial summation on each processor

    sum[Pn] = 0;for (i = 1000*Pn;

    i < 1000*(Pn+1); i = i + 1)sum[Pn] = sum[Pn] + A[i];

    Now need to add these partial sums Reduction: divide and conquer

    Half the processors add pairs, then quarter,

    Need to synchronize between reduction steps

  • 8/2/2019 Introduction to Multicore

    11/55

    Example: Sum Reduction

    half = 100;

    repeatsynch();

    if (half%2 != 0 && Pn == 0)

    sum[0] = sum[0] + sum[half-1];

    /* Conditional sum needed when half is odd;

    Processor0 gets missing element */half = half/2; /* dividing line on who sums */

    if (Pn < half) sum[Pn] = sum[Pn] +sum[Pn+half];

    until (half == 1);

  • 8/2/2019 Introduction to Multicore

    12/55

    Message Passing

    Each processor has private physicaladdress space

    Hardware sends/receives messagesbetween processors

  • 8/2/2019 Introduction to Multicore

    13/55

    Loosely Coupled Clusters

    Network of independent computers

    Each has private memory and OS

    Connected using I/O system E.g., Ethernet/switch, Internet

    Suitable for applications with independenttasks

    Web servers, databases, simulations,

    High availability, scalable, affordable

    Problems

    Administration cost (prefer virtual machines)

    Low interconnect bandwidth

    c.f. processor/memory bandwidth on an SMP

  • 8/2/2019 Introduction to Multicore

    14/55

    Sum Reduction (Again)

    Sum 100,000 on 100 processors

    First distribute 100 numbers to each

    The do partial sums

    sum = 0;

    for (i = 0; i

  • 8/2/2019 Introduction to Multicore

    15/55

    Sum Reduction (Again) Given send() and receive() operations

    limit = 100; half = 100;/* 100 processors */

    repeathalf = (half+1)/2; /* send vs. receive

    dividing line */if (Pn >= half && Pn < limit)send(Pn - half, sum);

    if (Pn < (limit/2))sum = sum + receive();

    limit = half; /* upper limit of senders */until (half == 1); /* exit with final sum */

    Send/receive also provide synchronization Assumes send/receive take similar time to addition

  • 8/2/2019 Introduction to Multicore

    16/55

    M l i h di

  • 8/2/2019 Introduction to Multicore

    17/55

    Multithreading

    Performing multiple threads of execution in parallel Replicate registers, PC, etc. Fast switching between threads

    In multiple-issue dynamically scheduled processor Schedule instructions from multiple threads

    Instructions from independent threads execute whenfunction units are available

    Within threads, dependencies handled by scheduling andregister renaming

    Example: Intel Pentium-4 HT

    Two threads: duplicated registers, shared function unitsand caches

  • 8/2/2019 Introduction to Multicore

    18/55

    Future of Multithreading

    Will it survive? In what form? Power considerations simplified

    microarchitectures

    Simpler forms of multithreading Tolerating cache-miss latency

    Multiple simple cores might share

    resources more effectively

    S

  • 8/2/2019 Introduction to Multicore

    19/55

    Instruction and Data Streams

    An alternate classificationData Streams

    Single Multiple

    Instruction

    Streams

    Single SISD:

    Intel Pentium 4

    SIMD: SSE

    instructions of x86Multiple MISD:

    No examples todayMIMD:Intel Xeon e5345

    SPMD: Single Program Multiple Data A parallel program on a MIMD computer

  • 8/2/2019 Introduction to Multicore

    20/55

    SIMD

    Operate elementwise on vectors of data E.g., MMX and SSE instructions in x86

    Multiple data elements in 128-bit wide registers

    All processors execute the same

    instruction at the same time Each with different data address, etc.

    Simplifies synchronization

    Reduced instruction control hardware Works best for highly data-parallel

    applications

  • 8/2/2019 Introduction to Multicore

    21/55

    Vector Processors

    Highly pipelined function units

    Stream data from/to vector registers to units

    Data collected from memory into registers

    Results stored from registers to memory

    Example: Vector extension 32 64-element registers (64-bit elements)

    Vector instructions lv, sv: load/store vector

    addv.d: add vectors of double

    addvs.d: add scalar to each element of vector of double

    Significantly reduces instruction-fetchbandwidth

  • 8/2/2019 Introduction to Multicore

    22/55

    Example: (Y = a X + Y)

    Conventional MIPS code

    l.d $f0,a($sp) ;load scalar aaddiu r4,$s0,#512 ;upper bound of what to

    loadloop: l.d $f2,0($s0) ;load x(i)

    mul.d $f2,$f2,$f0 ;a x(i)l.d $f4,0($s1) ;load y(i)add.d $f4,$f4,$f2 ;a x(i) + y(i)s.d $f4,0($s1) ;store into y(i)addiu $s0,$s0,#8 ;increment index to xaddiu $s1,$s1,#8 ;increment index to ysubu $t0,r4,$s0 ;compute boundbne $t0,$zero,loop ;check if done

    Vector MIPS code

    l.d $f0,a($sp) ;load scalar alv $v1,0($s0) ;load vector xmulvs.d $v2,$v1,$f0 ;vector-scalar multiplylv $v3,0($s1) ;load vector yaddv.d $v4,$v2,$v3 ;add y to productsv $v4,0($s1) ;store the result

    F E l S t

  • 8/2/2019 Introduction to Multicore

    23/55

    Four Example Systems

    2 quad-coreIntel Xeon e5345(Clovertown)

    2 quad-core

    AMD Opteron X4 2356(Barcelona)

  • 8/2/2019 Introduction to Multicore

    24/55

    IBM C ll B db d E i

  • 8/2/2019 Introduction to Multicore

    25/55

    IBM Cell Broadband Engine

    Abbreviations

    PPE: PowerPC Engine

    SPE: SynergisticProcessing Element

    MFC: Memory FlowController

    LS: Local Store

    SIMD: Single InstructionMultiple Data

    32KB,I, D

    512KBU

    300-600cycles

    128 , 128-bit registers128-bit vector/cycle

    CELL BE P i M d l

  • 8/2/2019 Introduction to Multicore

    26/55

    CELL BE Programming Model

    No direct access to DRAM from LS of SPE,Buffer size: 16KB

  • 8/2/2019 Introduction to Multicore

    27/55

    G hi i th S t

  • 8/2/2019 Introduction to Multicore

    28/55

    Graphics in the System

    GPU A hit t

  • 8/2/2019 Introduction to Multicore

    29/55

    GPU Architectures

    Processing is highly data-parallel GPUs are highly multithreaded Use thread switching to hide memory latency

    Less reliance on multi-level caches

    Graphics memory is wide and high-bandwidth

    Trend toward general purpose GPUs Heterogeneous CPU/GPU systems CPU for sequential code, GPU for parallel code

    Programming languages/APIs

    DirectX, OpenGL C for Graphics (Cg), High Level Shader Language(HLSL)

    Compute Unified Device Architecture (CUDA)

    Classifying GPUs

  • 8/2/2019 Introduction to Multicore

    30/55

    Classifying GPUs

    Dont fit nicely into SIMD/MIMD model Conditional execution in a thread allows an

    illusion of MIMD But with performance degredation

    Need to write general purpose code with care

    Static: Discoveredat Compile Time

    Dynamic: Discoveredat Runtime

    Instruction-Level

    Parallelism

    VLIW Superscalar

    Data-LevelParallelism

    SIMD or Vector Tesla Multiprocessor

    Programming Model

  • 8/2/2019 Introduction to Multicore

    31/55

    Programming Model

    CUDAs design goals

    extend a standard sequential programming language,specifically C/C++, focus on the important issues of parallelismhow to craft efficient

    parallel algorithmsrather than grappling with the mechanics ofan unfamiliar and complicated language.

    minimalist set of abstractions for expressing parallelism highly scalable parallel code that can run across tens of

    thousands of concurrent threads and hundreds of processor

    cores.

    Tesla C870 GPU

  • 8/2/2019 Introduction to Multicore

    32/55

    Tesla C870 GPU

    8 KB / multiprocessor

    8 KB / multiprocessor

    1.5 GB per GPU

    16 KB

    up to 768 threads (21 bytes of shared memory and 10 registers/thread)

    (8192)

    CUDA Program Organization

  • 8/2/2019 Introduction to Multicore

    33/55

    CUDA Program Organization

    Host program and GPU One or more parallel kernels

    Kernel scalar sequential program on a set of parallel threads. Threads

    The programmer organizes grid of thread blocks.

    Threads of a Block allowed to synchronize via barriers have access to a high-speed, per-block shared on-chip memory for interthread

    communication. Threads from different blocks in the same grid

    coordinate only via shared global memory space visible to all threads.

    CUDA requires that thread blocks be independent, meaning: kernel must execute correctly no matter the order in which blocks are run, provides scalability

    Main consideration in decomposing workload into kernels the need for global communication synchronization amongst threads

    Tesla C870 Threads

  • 8/2/2019 Introduction to Multicore

    34/55

    Tesla C870 Threads

    Tesla C870 Memory Hierarchy

  • 8/2/2019 Introduction to Multicore

    35/55

    Tesla C870 Memory Hierarchy

    texture and constant caches: utilized by all the threads within the GPU

    shared memory: individual and protected region available to threads ofa multiprocessor

    From Software Perspective

  • 8/2/2019 Introduction to Multicore

    36/55

    From Software Perspective

    A kernel is a SPMD-style

    Programmer organizes threads into thread blocks;

    Kernel consists of a grid of one or more blocks

    A thread block is a group of concurrent threads that cancooperate amongst themselves through barrier synchronization

    a per-block private shared memory space Programmer specifies

    number of blocks

    number of threads per block

    Each thread block = virtual multiprocessor each thread has a fixed register each block has a fixed allocation of per-block shared memory.

    SIMT Architecture

  • 8/2/2019 Introduction to Multicore

    37/55

    SIMT Architecture

    Threads of a block grouped into warps containing 32 threads each.

    A warps threads free to follow arbitrary and independent execution paths (DIVERGE)

    collectively execute only a single instruction at any particular instant.

    Divergence and reconvergence

    wait in turn while other threads of the warp execute the instructions managed in hardware.

    Ideal match for typical data-parallel programs,

    SIMT execution of threads transparent to the programmer

    a scalar computational model.

    Example

  • 8/2/2019 Introduction to Multicore

    38/55

    Example

    B blocks of T threads

    Which element am I processing?

    Data parallel vs. Task parallel kernels !Expose more fine-grained parallelism

    OpenMP vs CUDA

  • 8/2/2019 Introduction to Multicore

    39/55

    OpenMP vs CUDA

    for (i = 0; i < 32; i++){

    for (j = 0; j < 32; j++)value=some_function(A[i][j]}

    #pragma omp parallel for shared(A) private(i,j)

    Efficiency Considerations

  • 8/2/2019 Introduction to Multicore

    40/55

    Efficiency Considerations

    Avoid execution divergence

    threads within a warp follow different execution paths. Divergence between warps is ok

    Allow loading a block of data into SM

    process it there, and then write the final result back out to

    external memory.

    Coalesce memory accesses

    Access executive words instead of gather-scatter

    Create enough parallel work 5K to 10K threads

    Potential Tricks

  • 8/2/2019 Introduction to Multicore

    41/55

    Potential Tricks

    Large register file

    primary scratch space for computation. Small blocks of elements to each thread,

    rather than a single element to each thread

    The nonblocking nature of loads software prefetching for hiding memory latency.

    White Spaces After Digital TV

  • 8/2/2019 Introduction to Multicore

    42/55

    White Spaces After Digital TV

    In telecommunications, white spaces refer to vacant

    frequency bands between licensed broadcastchannels or services like wireless microphones.

    After the transition to digital TV in the U.S. in June2009, the amount of white space exceeded theamount of occupied spectrum even in major cities.

    Utilization of white spaces for digital communicationsrequires propagation loss models to detect occupied

    frequencies in near real-time for operation withoutcausing harmful interference to a DTV signal, or otherwireless systems operating on a previously vacantchannel.

    Challenge

  • 8/2/2019 Introduction to Multicore

    43/55

    Irregular Terrain Model (ITM), also known as theLongley-Rice model, is used to make predictions

    of radio field strength based on the elevationprofile of terrains between the transmitter and thereceiver. Due to constant changes in terrain topography and variations in radio propagation,

    there is a pressing need for computational resources capable of running hundredsof thousands of transmission loss calculations per second.

    Challenge

    ITM

  • 8/2/2019 Introduction to Multicore

    44/55

    Given the path length (d) for a radio

    transmitter T, a circle is drawn around Twith radius d. Along the contour line, 64 hypothetical

    receivers (Ri) are placed with equaldistance from each other.

    Vector lines from Tto each Riare furtherpartitioned into 0.5 km sectors (Sj). Atmospheric and geographic conditions

    along each sector form the profile of thatterrain (used 256K profiles).

    For each profile, ITM involves independentcomputations based on atmospheric andgeographic conditions followed bytransmission loss calculations.

    ITM

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    45/55

    GPU Strategies for ITM

    8 KB /multiprocessor

    8 KB /multiprocessor

    1.5 GB perGPU

    16

    KB

    ITM requires 45 registers Each profile is 1KB ( radio frequency, path length, antenna heights, surface transferimpedance, plus 157 elevation points)

    How many threads / MP?

    reduces register count to 37

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    46/55

    GPU Strategies for ITM

    128*16 threads 16*16 threads 192*16 threads

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    47/55

    GPU Strategies for ITM

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    48/55

    GPU Strategies for ITM

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    49/55

    GPU Strategies for ITM

  • 8/2/2019 Introduction to Multicore

    50/55

    Configurations

  • 8/2/2019 Introduction to Multicore

    51/55

    Configurations

    Comparison

  • 8/2/2019 Introduction to Multicore

    52/55

    Comparison

    Productivity from code development perspective Based on personal experience of a Ph.D. student

    with C/C++ knowledge and the serial version of the ITM code in hand, without prior background on the Cell BE and GPU programming environments.

    Data logged for the learning curve and design and debugging times individually.

    Fermi: Next Generation GPU

  • 8/2/2019 Introduction to Multicore

    53/55

    32 cores, 1536 threads per core

    No complex mechanisms speculation, out of order execution, superscalar, instruction

    level parallelism, branch predictors

    Column access to memory faster search, read

    First GPU with error correcting code registers, shared memory, caches, DRAM

    Languages supported C, C++, FORTRAN, Java, Matlab, and Python

    IEEE 75408 Double-precision floating point

    fused multiply-add operations, Streaming and task switching (25 microseconds)

    Launching multiple kernels (up to 16) simultaneously visualization and computing

    Fermi: Next Generation GPU

    GeForce GTX580 (512 streaming processors). It costs exactly $499 on newegg

  • 8/2/2019 Introduction to Multicore

    54/55

    Concluding Remarks

  • 8/2/2019 Introduction to Multicore

    55/55

    Concluding Remarks

    Amdahls Law doesnt apply to parallel computers

    Goal: higher performance by using multipleprocessors

    Difficulties

    Developing parallel software Devising appropriate architectures

    Many reasons for optimism

    Changing software and application environment

    Chip-level multiprocessors with lower latency, higherbandwidth interconnect

    An ongoing challenge for computer architects!