exascale storage
TRANSCRIPT
By2016,server-basedstoragesolu6onswilllowerstoragehardwarecostsby50%ormore…
Storage@aTippingPoint…
Whatdoesthismeanforus?
By2018,3ofthetop7general-purposediskarrayvendorswilleitherbeacquiredorexitthestoragehardwarebusiness…
LeverageonFrameworkIntegra6on,ManagementSoMware&Automa6on, strongCustomer Supportcapabili6esandevolvingoverallstorageecosystemasaandrobustglobaloffering
Opportunityupforgrabs…
Assess emerging storage architectures,technologies&approachestocreateacombinedstrategytomeetspecificworkloadrequirements
Makingthemostofit…
NewbreedofStorageServices…
Designedfor…
• Web-Scale–Scale-Up&Scale-Out• Mul6-Tenancy–Mul6-Customer/Container• Hyper-Access–Millionsofend-consumers• Resilience–Ge]ngoverPrac6calLimita6ons
StoreGlobal–AccessLocal!
OSSAddressestheNeed…
OSSSDSSolu@ons
ü Nutanix
ü Gluster
Enterprise
Hyper-Scale
Transac@onal
Distributed
Appliance
COTS/HPC
OneWorkloaddoesn’tfitall…
Architecturestofitvariousworkloads
Type1:ClusteredArchitecture
‘Federated Model’ layered a top‘scale-up’ architecture makesthemmore‘scale-out’typefromamanagementstandpoint.
Tends to ‘bounce the IO’ un6l itgetstothebrain(header)thathasthe data. ’Federated’ models usedata mobi l i ty approach torebalance between brains &persistence pools, leading to lowlatencyonwrites
Brains(HAHeader)
PersistentPool
Type2:TightlyCoupled,Scale-Out
Uses shared memory (cache andmetadata)betweennodes,andthedataitself isdistributedacross somenumbernodes.Thisarchitecturedealswithlargeamountofinter-nodecommunica6on
The defining element of sharedmemory models is cri6cal tothese des igns . I t enables‘symmetric’ IO paths through allbrains. It isdesignedsothat infailure (planned or unplanned)modes, IO opera6ons wouldremainrela6velybalanced.
Brains(HAHeader)
PersistentPool
IOPath(SharedMemory)
Type3:LooselyCoupled,Scale-Out
Thismodeldoesnotusing sharedmemory between nodes, but thedata itself is distributed acrossmul6ple nodes. It deals with alarger amount of inter-nodecommunica6on on writes (IOintensive) as data is distributed.Asitistransac6onalthewritesaredistributed&alwayscoherent
Thedesign–• Simpleinopera6onsandscaling.• Verygooddistributedreadsasdata
isservicedbymul6plenodes.• Not ‘HA’. The resilience comes from
datacopies&distribu6on.
Brains(mul@-node)
DistributedPool
Type4:Distributed,ShareNothing
TheDesign–• NoSharedMemory• Non-transac6on,Lazydata• Distributedreadscanbeachieved• No‘HA’.Theresilienceofspecified
datacancomefromdistribu6on.
TheArchitecture–• The‘MostScalable’Architecture• Super-Simpleimplementa6on• HighlyCOTSreliable,Lowcost• Mostly‘SoMwareOnly’design• Object & non-POSIX support on
basefilesystem
WorkloadbasedArchitecture…
On-Premise Hosted/Cloud(Private) Hosted/Cloud(Public)
GlusterStoragehasafullysupportedintegra6on• HadoopDataPlaiorm2.1• AmbariManagementSuite
Thisintegra6oncanrunvariousHadoopjobswith• accomplishedfilesystemplug-in• reliableenterprisegradestorageback-end• standardprotocol-basedingestop6ons• nosinglepointoffailure
GlusterStorageisaverified,highperformanceback-endforSplunk'scoldstorage6er,usedforvastmachinedataanalysis.Web-scale object storage solu6ons for archival & rich media, areCloudStackofferingsonCephStorage
ISVMaturityfocused…
Exascale…
Scale-outStackDesign
SingleglobalnamespaceAggregatesdiskandmemoryresourcesintoasingletrustedstoragepool.SecuritySupport SELinux enforcing mode with SSL-basedin-flightencryp@onObjectaccesstofilestorageFilestorecanbeaccessedusingobject-API.ErasurecodingEnhance data protec@on by us inginforma@on stored in the system toreconstructlostorcorrupteddata.Bit-rotdetecXonHelppreservetheintegrityofdataassetsbydetec@ngsilentcorrup@on.
TieringAutoma@callymovedatabetweenfast(SSD-based)andslow(HDD)@ersbasedonaccessfrequency.ReplicaXonSupports synchronous replica@on within adata center and asynchronous [email protected]@onthroughcluster-widefilesystemsnapshots.Useraccessibleforeasyrecoveryoffiles.ElasXchashingalgorithmNometadataserverlayereliminatesperformanceboYlenecksandsinglepointsoffailure.
FeatureGlance…
IndustryStandardClientSupport
• NFS, SMB protocols for file-basedaccess
• NFSv4 mul@-headed support forenhancedsecurity&resilience
• OpenStackSwi]supportforObjectaccess
• GlusterFS na@ve client for highlyparallelizedaccess
DeepHadoopIntegraXon
• HDFS-compa@blefilesystem• Nosinglepointoffailure• NFSandFUSEbaseddatainges@on
IntegraXonwithRHEV
• Centralized visibility and unifiedmanagement of storage and virtualinfrastructures through RHEVManagerconsole.
• Livemigra@onofvirtualmachines
FeatureGlance…Easyonlinemanagement
• Web-basedmanagementconsole• Powerful and intui@ve CLI for Linux
admins• Monitoring(Nagios-based)• Expand/shrinkstoragecapacitywithout
down@me
Scale-outWrite…• Theclientini6ateanIOandtransmitsittothenode it's communica6ngwith. For all-in-one style architectures, this is a VMnode that's co-located with the client onthesamehardware
• Once the node receives the writeacknowledgement from theothernode(s),i t r e s pond s b a c k t o t h e c l i e n tacknowledgingthewrite.
• Depending on the array plaiorm, otherthings can be done with the write likeinlinededuplica6on,compression,etc.
• Some arrays that implement flash-basedwritecachingcanstagethewritestoflashtocleartheRAMformoreincomingwrites.
• Thewriteiseventuallyflushedtodisk(SSDor Magne6c) on each node that receivedthewrite
Scale-outRead….• The client ini6ates an IO request andt r a n sm i t s i t t o t h e n o d e i t scommunica6ng with. For all-in-one stylearchitectures,thisisaVMnodethat'sco-located with the client on the samehardware.
• ThenodereceivesthatIO,checksitsreadcache in RAM for the data and then(depending on the array) checks SSDcacheforthedata.
• If the data isn't in either loca6on, thenode checks itsmetadata table to locatethedataondisk (localoranothernode/nodes). Data is read directly from theunderlying disks if local or is requestedfrom containing node across the inter-nodelink.
• The node places a copy of the read incacheandrespondstotheclientwiththerequesteddata.
Scale-outResilience….DistributedClustered
• UseofSSD(&Magne6c)acrosstheenvironmentasonesharedreadcache• Speed comparable with an All-Flash Array; All VM IO will be from flash, while
backupwillbefromSSD-SSD-Magne6c• ScalingofcapacityandperformanceachievedbyaddingmoreSSDs• LimitsfailureimpactofSSD.IOavailableforrebuild&hotcacheforLive-Migra6on
TheBo]omLine…
SoMwareDefined Storage (SDS) canachieve ‘Exascale’ propor6ons,which todatehasbeendifficulttomanagecost-effec6vely,evenatEnterpriselevels.Wide-spreadadop6onofWeb-ScaleandDistributedApplica6onarchitecture,by Enterprises, poses significant opportuni6es for SDS usage to gomainstream. Enterprises essen6ally would look at Service Providers toprovisionthishyper-scaleinfrastructure,whiletheyfocusonamoreengagingBusinessApp&Dev-Ops.Bearinmind,however,thattheastrategicallyposi6onedSDSServiceporioliomayrequiresubstan6alspecialistskillsandresources inareassuchassizing,integra6on, tuning, maintenance and support, a packaged Service offeringfromtheServiceProviderisamuchan6cipatedmove.
GlusterFSCurrentFeatures&Roadmap:hYp://gluster.readthedocs.org/en/latest/presenta@onsGlusterFS_Current_Features_and_Roadmap.pdf
AddiXonalReading…
GartnerDocID:G00255093hYp://www.gartner.com/technology/reprints.do?id=1-23NR9T2&ct=141027&st=sb
RedHatGlusterStoragehYp://www.redhat.com/en/files/resources/en-rhst-gluster-datasheet-INC0210625.pdf
UnderstandingStorageArchitecturehYp://virtualgeek.typepad.com/virtual_geek/2014/01/understanding-storage-architectures.html
DistributedFileSystemhYp://cecs.wright.edu/~pmate@/Courses/7370/Lectures/DistFileSys/distributed-fs.html
Discussion&Huddle…
[email protected]+919619455020