towards energy efficient data management in hpc: the open ... · 11/14/16 11 oed use cases...
Post on 24-Aug-2020
0 Views
Preview:
TRANSCRIPT
Towards Energy Efficient Data Management in HPC: The Open Ethernet Drive Approach
Anthony Kougkas, Anthony Fleck, Xian-He Sun
PDSW-DISCS’16Monday, November 14th
Salt Lake City, USA
11/14/16 2
Outline
● Introduction● Background● Evaluation results● Conclusions● Future directions
11/14/16 3
Introduction
● What is an Open Ethernet Drive (OED)?● Who makes them?● Why do we need one?
11/14/16 4
Open Ethernet Drive
● An “intelligent” storage device in a 3.5” form factor● ARM-based CPU● Fixed-size RAM● Ethernet card● ...and a disk drive.
11/14/16 5
Open Ethernet Drive ecosystem
● Kinetic Open Storage Project (8/2015) created by– Seagate– Western Digital (HGST)– Toshiba
● Joined byCisco Cleversafe
(IBM)DELL
DigitalSense NetAppOpen
vStorage
RedHat Scality
11/14/16 6
Why an Open Ethernet Drive in HPC?● Two main reasons:
– Optimize global I/O performance– Reduce energy consumption
11/14/16 7
I/O optimization using OED● Processor-per-disk database machines (1983), perform simple
queries on disk exploiting locality.● Active Storage (1998), proposed to offload some computations
to storage servers.● Decoupled Execution Paradigm (2013), specialized data nodes
perform computations to minimize the data movement.● Active Burst Buffer (2016) perform in-situ visualization and/or
analysis.● OED encapsulates a lot of the necessary tech in a small,
affordable device that will enable extra functionality.
11/14/16 8
Energy and cost savings● Designed with low-powered mobile components.● OED small factor requires less space.● And thus, more efficient cooling.● Less and easy maintenance.
11/14/16 9
Outline
● Introduction● Background● Evaluation results● Conclusions● Future directions
11/14/16 10
OED architecture● Designed to bring computation closer to the data.● Presented in enclosures of multiple such drives.● Enclosures have an embedded switched fabric (60Gbit/s). ● Runs Linux OS (Debian 8.0).● Internal components are subject to each implementation.
11/14/16 11
OED use cases● Mirantis, collaborated with HGST to deploy Openstack’s
Swift object store, Ceph’s OSDs and GlusterFS bricks.● Cloudian, deployed its own Hyperstore service on an
enclosure of 60 OED drives.● Skylable, deployed their object store service SkylableSX.● All of the above concluded that OED is the perfect
building block for an energy efficient and horizontally scalable storage cluster.
Can we bring it to HPC and harness its strengths?
11/14/16 12
Outline
● Introduction● Background● Evaluation results● Conclusions● Future directions
11/14/16 13
Test environment● Three categories:
– Hardware components with benchmarks
– Overall device with real applications
– Energy consumption (Watts)
● Software used:– Stress-ng
– SysBench
– Iperf
– Out-of-core sorting
– Vector addition
– Descriptive statistics
11/14/16 14
CPU performanceStress-ng
Sysbench
16x slower than personal computer 9x slower than server node
50x slower than personal computer 30x slower than server node
11/14/16 15
RAM performanceStress-ng
Sysbench
12x slower than personal computer 5x slower than server node
11x slower than personal computer 7x slower than server node
11/14/16 16
Disk performance
Stress-ng Sysbench
2.3x faster than personal computer 1.7x faster than server node
4.5x faster than personal computer 3.5x faster than server node
11/14/16 17
Ethernet performance
Stress-ng Iperf
2-6x slower than personal computer 1-4x slower than server node
3x slower than personal computer 2x slower than server node
11/14/16 18
Real Applications
Sorting Desc. Statistics
Let’s just say OEDs are currently slower :(
Vector Addition
11/14/16 19
Energy consumption● Higher Performance comes
with a cost.● OED needs 1/10th of the
power compared to an average node.
● Sorting integers took 3x more time on the OED but consumed 1/14th of watts needed per sorting unit.
● Sorting 4GB of integers: ● OED → 1380w● Server → 3800w
11/14/16 20
Outline
● Introduction● Background● Evaluation results● Conclusions● Future directions
11/14/16 21
Conclusions● This 1st generation of OED technology is not yet on par
with the average server node in terms of performance.● Energy savings seem promising.● OEDs could be used to run parallel file system servers for
an archival and energy efficient storage solution.● As OED technology progresses, data-intensive operations
can be accelerated by offloading computation on OEDs.
11/14/16 22
Outline
● Introduction● Background● Evaluation results● Conclusions● Future directions
11/14/16 23
Future work● Installed MPICH and OrangeFS storage system on an
enclosure of 60 OED drives.● Initial IOR benchmarks were successful. ● The 2nd generation of OED looks very promising.● Planning to explore the use of OED as specialized data
nodes that can run operations on local data– Compression / decompression
– Deduplication
– Statistics
11/14/16 24
In the meantime...
Q & A
Towards Energy Efficient Data Management in HPC: The Open Ethernet Drive Approach
Anthony Kougkas
akougkas@hawk.iit.edu
The authors would like to acknowledge Los Alamos Lab for providing us with the prototype devices.
top related