high performance storage service virtualization scott baker university of arizona

Post on 20-Dec-2015

213 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

High Performance Storage Service Virtualization

Scott Baker

University of Arizona

What is Virtualization?

C lie nt S e rve rS e rvic e

C lie nt S e rve rS e rvic e

V irtua lize rS e rvic e '

A virtualizer is inserted between the client to provide a better service:

Consider an existing client/server scenario:

Why Virtualization ?

• Why not create a newer/better service ?– Modified clients / servers– Lengthy standardization process– Slow to integrate into existing infrastructure

• Virtualization offers– Unmodified clients/server– No standardization requirements– Rapid integration

Types of Virtualizations

• Mutation– Change a service into something different

• Aggregation– 2+ servers 1 big server

• Replication– 2+ servers 1 more reliable server

• Fortification– Vulnerable server more secure server

Mutation (Gecko)

• Web uses HTTP protocol• Traditional programs use file system semantics

– open, close, read, write, etc

• Inconvenient to modify existing body of applications to use the web

• WWW8, SP&E papers

L in u x o rW in d o ws C lien t

H T T PG e c k o

N F S W o rld W ideW e b

Aggregation (Mirage)

• Combine 2+ NFS file systems to create one big file system

• Clients are unaware multiple servers exist• IEEE LCN paper

C lie nt

N F S S e rve r

M ira g eN F S

N F S S e rve r

N F S

N F S

Replication (Mirage)

• Unmodified primary server• Specialized backup server• Asymmetric design

– Commercial primary, commodity backup

• Logging

C lie nt P rim a ry S e rve rM ira g eN F S

B a c k up S e rve r

N F S

N F S + B P

Fortification (Mirage)

• Fortify a server against DoS attacks

• Several ideas– Prevent faulty requests from reaching servers– Use scheduling to ensure fairness– Push authentication to the border of the

network

• Currently work in progress

Mirage in the Network

• Mirage could be located in:– Client (patch OS, or user-mode daemon)– Server (patch app, or user-mode daemon)– Router (unmodified clients, servers)

• Mirage is a router not an application– Rewrite packets on-the-fly & forward– Benefits: stateless, low overhead

NFS Basics

• NFS uses “handles” to identify objects– Lookup (par_handle, “name”) chi_handle– Read (chi_handle) data

• Handles are opaque to clients

• Handles are 32-bytes (NFSv2) or 64-bytes (NFSv3)

Aggregation Issues

• Make 2+ servers look like 1 big server

• Key problem– 2 servers may generate the same handle for

different objects– Client will be confused

• Solution– Virtual handles

Virtual and Physical Handles

• Virtual handles exist between clients and Mirage• Physical handles exist between Mirage and

servers

C lient M irage

S erver # 1

S erver # 2

V irtual H and les P hys ic al H and les

VFH Contents

• Mirage decides what to put in VFH• VFH is composed of

– PIN (Physical Inode Number)– PFS (Physical File System Number)– SID (Server ID)– VIN (Virtual Inode Number)– HVC (Handle Verification Checksum)– MCH (Mount Checksum)

• (PIN, PFS, SID) uniquely identifies a file

Data Structures

• Transaction Table (TT)– Entry created during request– Entry deleted during reply– Remembers NFS Proc Number, Client ID

• Handle Table (HT)– VFH PFH mappings

• Tables are Soft State

Request / Reply Processing

• On requests,– Lookup VFH in HT yields PFH– Rewrite VFH in request with PFH– Forward to server (SID tells which one)

• On replies,– Lookup PFH in HT yields VFH

• Create new mapping if necessary

– Rewrite PFH in reply with VFH– Forward to client

Router Failure / Recovery

• If router fails, TT and HT are lost• Clients will retry any ops in progress

– TT state regenerated automatically

• Recover HT state from fields in VFH– Extract (PIN, PFS, SID)– Search servers for (PIN, PFS, SID) to get

PFH– Similar to BASE

• Periodically checkpoint HT to servers

Prototypes

• User Mode Process– Linux operating system / commodity HW– Proof of concept– Demonstrates aggregation & replication– UDP Sockets

• IXP2400 Network Processor– High performance– Possible production system– Subject of ongoing/future work

IXP2400 Overview

• 1 StrongArm CPU (general purpose, Linux OS)• 8 Microengine CPUs (packet processing)

B U S

m e 4

m e 7

m e 6

m e 5

m e 0

m e 3

m e 2

m e 1

G B E TH

G B E TH

Str o ng Ar mC P U

D R AM M e m o r ySR AM M e m o r y

S c ratc h P ad

H as h

C ap

Microengine CPU Properties

• Lots of registers– 256 GPR, 128 NN, 512 memory-i/o

• Special packet-processing instruction set

• Multithreading support– 8 threads per microengine– Zero context-switch overhead

• Asynchronous memory I/O

• Fast-path processing

Memory

• DRAM: 64 MB / 300 cycles– Direct IO to and from network interface

• SRAM: 8 MB / 150 cycles– Support atomic operations– Built-in “queues” w/ atomic dequeue, get, put

• Scratchpad:16 KB / 60 cycles– Supports atomic operations– Built-in “rings” with atomic get/put

• Local per-microengine: 2560 B / 3 cycles

IXP Issues

• Divide Mirage functionality across Microengines• Control interface between StrongArm and

Microengines• Optimize Microengine code

R e c e i ve r Tr ans m i tte rC l as s i fi e r

N fs R e q

N fs R e p

Str o ng Ar mC P U

P ac ke tsIn

P ac ke tsO ut

Benchmark Configuration

• Two IXP boards: Benchmark and Mirage• Attempt throughput and measure actual

F l o o d

F akeSe r ve r

C o unt

N F S R e q

N F S R e p

R eq u es t

R e w r itten R e q u e st

R ep ly

R ew r itte n R e p ly

B e nc hm ar k IX P B o ar d M i r ag e IX P B o ar d

(note: transmit, receive, classifier microengines not shown)

Loopback Configuration

• Simulates a router without Mirage

F l o o d

F akeSe r ve r

C o unt

R e que s t

R eq u es t

R ep ly

R e pl y

B e nc hm ar k IX P B o ar d M i r ag e IX P B o ar d

IXP Performance

0

100000

200000

300000

400000

500000

600000

Attempted Throughput (Packets / Sec)

Actu

al

Th

rou

gh

pu

t (P

ackets

/ S

ec)

getattr-loop

getattr-mirage

write-loop

write-mirage

Analysis

• User-mode Mirage– 40,000 packets/second– Read/Write bandwidth at 320 Mbps

• IXP Mirage– 290,000 packets/second– Read/write bandwidth exceeds gigabit line

speed (In theory, approx 2.4 Gbps)

Status

• Completed– User-mode Mutation (Gecko), Aggregation,

Replication, Fortification– IXP Aggregation

• To-do– IXP performance tuning– Finish IXP benchmarks– IXP Replication ?– IXP Gecko ?– SOSP Paper

Publications• Scott Baker, John Hartman, “The Gecko NFS Web Proxy,”

Proceedings of the Eighth International Conference on World Wide Web. 1999.

• Scott Baker, Bongki Moon, “Distributed Cooperative Web Servers,” Proceedings of the Eighth International Conference on World Wide Web. 1999.

• Scott Baker, John Hartman, “The design and implementation of the Gecko NFS Web Proxy,” Software Practice and Experience, June 2001.

• Scott Baker,  John Hartman, and Ian Murdock, "Swarm: Agent-Based Storage," The 2004 International Conference on Software Engineering Research and Practice. Las Vegas, Nevada. June, 2004.

• Scott Baker and John Hartman, "The Mirage NFS Router," The 29th IEEE Conference on Local Area Networks. Tampa, FL. November, 2004.

top related