metrocluster in clustered data ontap
TRANSCRIPT
GAJAH ANNUAL REPORT 2015 | 1
Next up: MetroCluster in Clustered Data ONTAP
Webinar
Follow along on Twitter!@FastLaneUS | #FLMC16
GAJAH ANNUAL REPORT 2015 | 2
MetroCluster in Clustered Data ONTAPPresented by Tia Williams
Follow along on Twitter!@FastLaneUS | #FLMC16
FAST LANE 2016 | 3
Why MetroCluster for Clustered Data ONTAPMetroCluster ArchitectureTwo-Node MetroClusterMetroCluster Non-Disruptive OperationsTransitioning MetroCluster to Clustered Data ONTAP 8.3
Agenda
@FastLaneUS | #FLMC16
FAST LANE 2016 | 4
Why MetroCluster for Clustered Data ONTAP Need for Continuous Availability Solution
Types of outages
Approximately 85% are planned events
Approximately 15% are unplanned events (1% natural disasters)
With shared infrastructure, negotiating a downtime is next to impossible
Approximately 70% of unplanned events are due to internal data center failures
Mission-critical applications demand no data loss
Downtime equals loss of revenue and reputation
85%Planned events
14%
Reasons for Storage Outage
Other
Superstorm Sandy
Vendor Patch
Vendor Software
Human Error
Power Failure
Vendor Hardware
18%
9%
9%
9%
18%
18%
18%
Source: The InfoPro, Storage Wave-17, 2013
@FastLaneUS | #FLMC16
FAST LANE 2016 | 5
Why MetroCluster for Clustered Data ONTAP
The Clustered Data ONTAP® operating system provides NDO within the data center
Ability to withstand component failures Ability to perform maintenance operations without
disruption Ability to perform technology refresh without
disruptionCluster Data ONTAPData Center A
MetroCluster™ technology enables business continuity and continuous availability beyond the data center
MetroCluster
Cluster B inData Center B
Cluster A inData Center A
MetroCluster Extends Non-disruptive Operations Beyond the Data Center
@FastLaneUS | #FLMC16
FAST LANE 2016 | 6
MetroCluster maintains the availability of your storage infrastructure
Why MetroCluster for Clustered Data ONTAP
Non-disruptive operations leading to zero data loss
Set-it-once simplicity
Zero change management
Lower cost and complexity of competitive solutions
Seamless integration with storage efficiency, SnapMirror,
NDO, virtualized storage
Unified: supports both SAN and NAS
Native Continuous Availability for Business-Critical Applications
Up to 200km
@FastLaneUS | #FLMC16
FAST LANE 2016 | 7
Why MetroCluster for Clustered Data ONTAP
Control failure
Storage or rack failure
Network failure
Local data center failure
Complete site failure
MetroCluster Protects Against:
“Zero minutes of planned and unplanned downtime since 2009.”
Jack Wolfskin
Site/Bldg A
Site/Bldg B
Up to 200 km
@FastLaneUS | #FLMC16
FAST LANE 2016 | 8
Why MetroCluster for Clustered Data ONTAP
Non-disruptive operations
Data ONTAP upgrade or platform
refresh does not require an outage
Site switchover required only for disasters
and site-wide events
All local component failures handled
locally
Most workflows do not require site-
level switchover
All nodes actively serve data to applications
MetroCluster Leverages Local HA Failover
@FastLaneUS | #FLMC16
FAST LANE 2016 | 9
Why MetroCluster for Clustered Data ONTAP MetroCluster for VMware Environments
Failed Server
VMware ESXi VMware ESXi
No Reboot,Seamless Cutover
NetApp® MetroCluster™
Fault Tolerance
Operating Server
Site 1 Site 2
Virtualization makes the infrastructure mission critical
Completes VMware HA/FT Same levels of availability
for storage that VMware®
HA and FT provide for VMs
Simplifies operations Zero interdependencies No application or OS agents
Deploys with confidence Tested and documented interoperability
On vMSC HCL since 5.0 NFS, iSCSI, FC, FCoE Only certified NAS solution
@FastLaneUS | #FLMC16
FAST LANE 2016 | 10
Why MetroCluster for Clustered Data ONTAP
Comprehensive Protection with MetroCluster, SnapMirror and SnapVault
MetroCluster™ with SnapMirror® and
SnapVault® provides continuous availability
within the data center and disaster recovery
protection at unlimited distances. It also provides the ability to remotely back up and archive to tape for
a fully integrated zero-data-loss 3-way DR
solution.
SnapMirror®Unlimited Distance
Disaster Recovery Site
Backup and Recovery Site
SnapVault®Unlimited Distance
Unlimited Distance
Up to 200 km
MetroCluster™
Multiple Recovery Points with Snapshot™ copies
Local Datacenter, Campus, Metro Area
@FastLaneUS | #FLMC16
FAST LANE 2016 | 11
MetroCluster Architecture
Two separate two-node clusters, one on each site, separated by up to 200 km Two clusters are connected through redundant fabrics NVRAM is mirrored to the local HA partner and DR partner on the remote site, sharing the same ISL fabric as the
storage replication Data written to the primary copy and synchronously replicated to secondary copy at the remote site
Works on an aggregate level and each aggregate consists of two “plexes”, one local and the other remote Writes are performed synchronously to both plexes and reads are performed from the local storage (by default)
Cluster peering interconnect mirrors cluster configurations
Cluster A Data Center A
Cluster BData Center B
200 km
Synchronous mirroring
NVRAM mirroringNode A1
Node A2
Node B1
Node B2
NVRAM mirroring
Cluster peering
ISL NVRAM mirroring
@FastLaneUS | #FLMC16
FAST LANE 2016 | 12
MetroCluster Architecture MetroCluster Replication Mechanism
Three different replication streams between two HA pairs across sites
NVRAM mirrored to HA partner and DR partner
All disk traffic mirrored at the aggregate level
Dedicated switch fabric and ISLs required
Cluster configuration replicated via peered network• All cluster configuration information is mirrored to the remote site• Can leverage existing, shared network infrastructure
Fibre Channel
IP Network
@FastLaneUS | #FLMC16
FAST LANE 2016 | 13
MetroCluster Architecture Cabling Overview
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS 2246
600GB
600GB
600GB
600G
B
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS 2246
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS 2246
600GB
600GB
600GB
600G
B
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
ATTO FibreBridges2 per stack
4 FC switches + ISLs
Storage shelves
HA pairSwitchless or switched cluster interconnects
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS 2246
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
2 FC initiators + 1 16Gb FCVI from each controller to each switch
Cluster AData Center A
Cluster BData Center B
Ethernet SASFC - ISL linkFC
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS2246
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600G
B
600GB
600G
B
600GB
600GB
600GB
600GB
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS2246
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600GB
600G
B
600GB
600G
B
600GB
600GB
600GB
600GB
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS2246
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
4 5 6 70 1 2 3 12 13 14 158 9 10 11 20 21 22 2316 17 18 19
DS2246
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
600G
B
Cluster peering(any routed IP network)
FC-VI
@FastLaneUS | #FLMC16
FAST LANE 2016 | 14
MetroCluster Architecture
Primary
Secondary
FC-VI 2 FC Initiators
Node A1
Node A2
Node B1
Node B2
Cluster Interconnect
ATTO: FC-SAS Bridge
ATTO: FC-SAS Bridge
DS4243
Backend Fabric
Backend Fabric
Cluster Peering Network
Site 1Primary
Site 1 Secondary
Site 2Primary
Site 2Secondary
Site 1 Site 2
3 Links
FC
FC - ISL linkSAS
FC-VI
Ethernet
@FastLaneUS | #FLMC16
FAST LANE 2016 | 15
MetroCluster Architecture
Local failover/failback
For workflows such as tech refreshes and
Data ONTAP upgrades
For component failures
Simple, non-disruptive switchover/switchback
No application/host scripting/action required
Planned or unplanned
One-command switchover
Three-command switchback SO/SB
MetroCluster Failover Characteristics
@FastLaneUS | #FLMC16
FAST LANE 2016 | 16
MetroCluster Architecture MetroCluster Simplifies Management
Site 1 Site 2
New Volume
MetroCluster
Others
Automatically protected
ExchangeExchange
SharePointSharePoint
OracleOracle
New Volume
Create LUN Create replicaLUN
Aggregate Mirroring
Create volume
Set up replication
@FastLaneUS | #FLMC16
FAST LANE 2016 | 17
MetroCluster Architecture MetroCluster Simplifies Management
Site 1 Site 2
New Volume
MetroCluster
Others
CFOD Command
ExchangeExchange
SharePointSharePoint
OracleOracle
New Volume
Aggregate SyncMirror®
1. Break Mirror 2. Bring Online
3. Break Mirror 4. Bring Online
5. Break Mirror 6. Bring Online
7. Break Mirror 8. Bring Online
@FastLaneUS | #FLMC16
FAST LANE 2016 | 18
MetroCluster Requirements
FCPFCVI ISL Links
FC Switches FC Switches
Cluster Peering Interconnect IP
Switches: Dedicated switches (Cisco®, Brocade) ISL/FC Link: Dedicated fibre or dedicated wavelength (with DWDM)Cluster peering: IP networkPlatform rules: MetroCluster™ only supported in midrange and high-end controllers; FlexArray supported All nodes in a MetroCluster DR group need to be identical (platform, storage, switches)
@FastLaneUS | #FLMC16
FAST LANE 2016 | 19
Two-Node MetroCluster
Distances of up to 200 km between sites
All storage is fabric-attached and visible to both nodes
Same level of protection of an HA-pair
Random read/write performance with inline compression enabled
Switchover and switchback transfer the cluster’s entire workload between sites
Support stretch MetroCluster (500 meters)
*Four-node MetroCluster highly recommended because of local-HA failover support
@FastLaneUS | #FLMC16
FAST LANE 2016 | 20
MetroCluster Supported Two-Node MetroCluster Configurations
@FastLaneUS | #FLMC16
FAST LANE 2016 | 21
Two-Node MetroCluster Two-Node Automatic Unplanned Switchover
Two-Node Automatic Unplanned Switchover (AUSO) is default in two node
MetroCluster configuration
Automatic failover triggered by node panic, reboot, power loss, power down
Disk ownership is transferred to the DR partner
Manual switchback required to return to normal operations
*Not available in 4-Node MetroCluster
@FastLaneUS | #FLMC16
FAST LANE 2016 | 22
MetroCluster Non-Disruptive Operations
For any failures (unplanned) on a single node in a cluster an automatic local HA failover is performed Local HA failover is also performed for planned events
Nondisruptive software upgrades Nondisruptive controller refreshes Addition of new HBAs, Flash Cache™ intelligent caching, etc.
Cluster A Data Center A
Cluster BData Center B
Synchronous MirroringISL
@FastLaneUS | #FLMC16
FAST LANE 2016 | 23
Transitioning MetroCluster to Clustered Data ONTAP Data ONTAP 8.3.0 MetroCluster Considerations
MetroCluster™ system size with Data ONTAP® 8.3 operating system:
4 nodes (2 nodes per site)
All aggregates have to be synchronously mirrored
You cannot convert clustered Data ONTAP into MetroCluster with data in place
Site switchover is for the entire cluster
Most failure scenarios are covered by local HA
@FastLaneUS | #FLMC16
FAST LANE 2016 | 24
Transitioning MetroCluster to Clustered Data ONTAP
Technology used: NetApp transition tool
SnapMirror®-based transition (7MTT) Downtime of minutes plus time to stop and restart apps on the “active” nodes
(similar to other clustered Data ONTAP® transitions) Customer-selected tools
Application-level migration tools Hypervisor-level migration tools Operating system volume manager tools
MetroCluster Transition Overview
@FastLaneUS | #FLMC16
FAST LANE 2016 | 25
Transitioning MetroCluster to Clustered Data ONTAP MetroCluster Transition Process
Primary Site Secondary Site
Data ONTAP® 8.3Existing
SnapM
irror
ISLs can be shared during transition (limits apply)
SnapM
irror
@FastLaneUS | #FLMC16
FAST LANE 2016 | 26
Transitioning MetroCluster to Clustered Data ONTAP MetroCluster Consolidation Savings
Consolidate two existing instances into 4-node MetroCluster™ DR group
Opex savings50% savings from ISL sharing
(approximately $100K annually)Non-disruptive upgrades/tech
refreshes (approximately $3K annually)Space, power, and cooling
savings from consolidation
Capex savingsExtend useful life of hardware
(approximately $10K annually)
FAS80XX
FAS32XX/62XX
@FastLaneUS | #FLMC16
FAST LANE 2016 | 27
Transitioning MetroCluster to Clustered Data ONTAP MetroCluster 7-Mode and Clustered Data ONTAP Comparison
Feature Prior to Data ONTAP 8.3
Data ONTAP 8.3
Synchronous data protection Yes YesNondisruptive component failure and replacement Yes Yes
HA and DR Yes (2 nodes) Yes (2 or 4 nodes)“Set and forget” ease of use Yes YesCross-site switchover Single command Single commandSupport/Compatible with all key Data ONTAP® features
Dedupe, SM, SV, tape, etc.
Same + QoS, VolMove, SVM
Clustered Data ONTAP value proposition No Data mobility, NDOTelecom costs No ISL sharing 4 nodes, ISL sharingLocal HA No YesMaximum distance 200 km 200 km
@FastLaneUS | #FLMC16
FAST LANE 2016 | 28
MetroCluster in Clustered Data ONTAP MetroCluster Resources
NetApp TR-4375 MetroCluster for Clustered Data ONTAP 8.3.1 Clustered Data ONTAP 8.3 MetroCluster Installation and Configuration Guide MetroCluster Service Guide
@FastLaneUS | #FLMC16
FAST LANE 2016 | 29
Summary
Why MetroCluster for clustered Data ONTAP
MetroCluster Architecture
Two Node MetroCluster
MetroCluster Non-Disruptive Operations
Transitioning MetroCluster to clustered Data ONTAP 8.3
@FastLaneUS | #FLMC16
FAST LANE 2016 | 30
Related Courses Clustered Data ONTAP
CIFS Administration NFS Administration Data ONTAP Cluster Administration SAN Scaling and Architecting
(CIFS)(NFS)(DCADM)(SANSA)
2 days1 day5 days2 days
@FastLaneUS | #FLMC16
FAST LANE 2016 | 31
Join our Loyalty Program!
Fast Lane Receive Prizes and Other Offers…
Check out our current promotions!
@FastLaneUS | #FLMC16
GAJAH ANNUAL REPORT 2015 | 32
THANK YOU
@FastLaneUS | #FLMC16