an adaptable benchmark for mpfs performance testing a master thesis presentation yubing wang...

27
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Post on 19-Dec-2015

220 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

An Adaptable Benchmark for MPFS Performance

Testing

A Master Thesis Presentation

Yubing Wang

Advisor: Prof. Mark Claypool

Page 2: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Outline of the Presentation

• Background

• MPFS Benchmarking Approaches

• Benchmarking Testbed

• Performance Data

• Conclusion

• Future Work

Page 3: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

SAN File System

• Storage Area Networks (SAN) NAS + Fibre Channel + Switch + HBA (Host Based Adapter). Architecture

• SAN File Systems An architecture for distributed file systems based on shared

storage. Fully exploits the special characteristics of Filbre Channel-based

LANs. Key feature is that clients transfer data directly from the device

across the SAN Advantages include: availability, load-balancing and scalability

Page 4: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

MPFS

• Drawbacks of Conventional Network File Sharing Server is the bottleneck. Store-and-forward model results in higher response

time.

• MPFS Architecture Server only involves in control data (metadata)

operations while file data operations are performed directly between clients and disks

MPFS uses standard protocols such as NFS and CIFS for control and metadata operations.

Potential advantages include better scalability and higher availability.

Page 5: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

File System Benchmarks

• SPEC SFS Only measure the server performanceOnly generate RPC loadNFS protocol onlyUnix only

• NetBench Windows onlyCIFS protocol only

Page 6: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Ideal MPFS Benchmark

• Help in understanding MPFS performance.

• Be relevant to a wide range of applications.

• Be scalable and target both large and small files.

• Provide workloads across various platforms.

• Allow for fair comparisons across products.

Page 7: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Motivations

• Current File System benchmarks are not suitable for the MPFS performance measurementThey only measure the server’s performance.They only target some specific file access protocols.

• MPFS is a new file access protocol and demands new file system benchmarkThe split-data-metadata architecture will prevail in the

SAN industry.Performance is critical to SAN file system.

Page 8: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Performance Metrics

• Throughput I/O rate, measured in operations/second Data rate, measured in bytes/seconds

• Response Time Overall average response time for all mixed operations. Average response time for individual operation. Measured in Msec/Op.

• Scalability Number of client hosts supported by the system with acceptable

performance

• Sharing System throughput and response time when multiple clients access the

same data

Page 9: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

MPFS Benchmark Overview• Application groups target the critical performance characteristics of

MPFS.

• Application mix percentages are derived from the low-level NFS or CIFS operation mix percentages.

• The file set is scalable and targets both big files and small files.

• The file access pattern is based on an earlier file access trace study.

• The load-generating processes in each load-generator is Poisson distributed.

• The embedded micro-benchmarks measure how MPFS performs

under intensive I/O traffics. • The huge file set and random file selection avoids caching effect.

Page 10: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Application Groups

• The application group is a mix of system calls that mimic MPFS applications.

• The applications are selected by profiling some real-world MPFS applications.

• The applications include both I/O operations and metadata operations.

• The operation groups for Windows NT follow the file I/O calls used in the Disk Mix test of NetBench.

Page 11: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Application Mix Tuning

• The application mix percentage is derived from the low-level NFS or CIFS operation mix percentage.

• The default NFS operation mix percentage we use is the NFS version 3 mix published by SPEC SFS 2.0.

• The default CIFS operation mix percentage is the CIFS operation mix used in NetBench.

• We allow user to specify the mix percentage for their specific applications.

Page 12: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

File Set Construction

• We build three types of file set with different file size distribution.

• We have small, medium and large file sets.• The small file set comprises 88% of small files (<= 16 KB). • The large file set comprises 18% of large files (>= 128MB).• We build huge file set to avoid caching effect.• The number of files and amount of data in our file set is

scaled to the target load levels.

Page 13: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

File Access Pattern

• Based on an empirical file system workload study.• File Access Order: sequential access or random

access • File access locality: the same files tend to get the

same type of access repeatedly. • File access burst: certain file access pattern occurs

in bursts. • Overwrite/Append Ratio: pre-fetching and space

allocation

Page 14: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Work Load Management

• Think time follows the exponential distribution.

• Operation selection is based on the specified mix percentage the operation context and file access patterns.

• Operation context is determined by profiling the MPFS applications.

Page 15: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

File Sharing

• Mainly measure how the locking mechanism affects the performance.

• Include read and write sharing.

• Multiple processes in a single client access the same file simultaneously.

• Multiple clients access the same file simultaneously.

Page 16: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Embedded Micro-benchmarks

• Measure the I/O performance of MPFS.

• Include sequential read, sequential write, random read, random write and random read/write

• Report the throughput measured in megabytes/second for each I/O test

Page 17: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Caching

• A larger client cache or more effective client caching may greatly affect the performance measurement since our benchmark is in the application level.

• Huge file set and random file selection help to avoid the caching effect.

Page 18: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Testbed Configuration

Fibre Channel

Unix/NTClient

Unix/NTClient

Unix/NTClient

Unix/NTClient

Data Mover

Unix/NTClient

Unix/NTClient

Unix/NTClient

Unix/NTClient

Page 19: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

System Monitors

• Network Monitor: - monitors the network states

- collects the network traffic statistics

• I/O Monitor: - monitors the disk I/O activities

- collects the I/O statistics

• CPU Monitor: - monitors the CPU usage

• Protocol Statistic Monitor: - collects the MPFS/NFS/CIFS statistics

Page 20: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Web Interface

Page 21: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Throughput and Response Time

0

1

2

3

4

5

6

7

8

9

0 1000 2000 3000 4000 5000

Generated Load (Ops/Sec)

Res

po

nse

Tim

e (M

sec/

Op

)

Generated Load Vs. Response Time for the MPFS Benchmarking testbed with 8 Solaris Clients

Page 22: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Scalability

0

500

1000

1500

2000

2500

3000

3500

4000

4500

5000

0 2 4 6 8 10 12

Number of Solaris Clients

Max

Mea

sure

d A

ggre

gate

Th

roug

hput

(Ops

/Sec

)

Measured Maximum Aggregate Throughput versus Number of Solaris Clients

Page 23: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Change of the Mix Percentage

0

2

4

6

8

10

12

0 1000 2000 3000 4000 5000

Generated Load (Ops/Sec)

Res

po

nse

tim

e (M

sec/

Op

)

Meta and Read/WriteMixRead/Write Mix

Meta Mix

Generated Load Vs. Response Time for different operation group mixes

Page 24: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Comparison between NFS and MPFS

0

2

4

6

8

10

12

14

0 1000 2000 3000 4000 5000

Generated Load (Ops/Sec)

Res

po

nse

Tim

e (M

sec/

Op

)

10 MPFS Clients

5 MPFS Clients &5 NFS Clients10 NFS Clients

Generated Load Vs. Response Time for three different MPFS and

NFS Solaris client combinations

Page 25: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Conclusion (1)

Our benchmark achieves four major goals:• Helps in understanding MPFS performance

Measure throughput and response time Measure the scalability Measure the performance for each individual operation Compare the performance of MPFS with that of NFS or CIFS.

• Generates realistic workload Operations are selected by profiling the real-life MPFS

applications. File access patterns are derived from an empirical file system

workload study. File set construction mimics the real-world environment.

Page 26: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Conclusion (2)

• File set is scalable and target both large and small filesThe number of files and amount of data in our file set is

scaled to the target load levels.The file sets are of different file size distribution.

• Provide workloads across various platforms.Our benchmark supports both Unix and Windows NT

systems.

Page 27: An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool

Future Work

• Create more realistic workload Build up a large set of MPFS trace archives Develop a profiling model to characterize the traces

• Improve the scalability measurement Our benchmark uses the number of clients (load generators) to represent the

scalability. The mapping between the load generator and client in a real-world application

is subject to further investigation.

• Develop a more general workload model for SAN file systems Different SAN file systems may have different implementation. A general benchmark should be independent of the implementation