Active-Active DC
Active-Active DC
Active-Active DC
min, so you can have “even more fun” after this full session
2. Some slides have a icon on the top right corner. It means they relate to an interesting reference or
associated information for the sake of completeness. I won’t cover them during the session.
3. Some slides have a icon on the top right corner. We will “fly over” those, if any.
Everybody wantsthat
Then try to figure thisout
… Then,
… and feeldivide & conquer
tired (or panic )
4
Objectives Legend
1. Understand the Active/Active Data
Centre requirements and considerations
Load SSL
2. Provide considerations for Active/Active Balancer Offloader APIC
DC Design – inclusive for Metro areas -
from storage, DCI, LAN extension, ACI
Application
and network services perspectives. SVI / HSRP Policy
IDS / IPS
3. Analyse the Active/Active aspects of Default Gw Infrastructure
cloud workloads and cloud-native Controller
applications with containers and micro-
services WAN
Accelerator Firewall
4. Deep dive on ACI Fabric extension,
stretch fabrics, application portability,
VM mobility, policy synchronisation, etc.
5. Share Experiences with State-full
Devices placements and their impact
within DCI environment
5
Agenda
• Active-Active (A/A) Data Centre:
– Market & Business Drivers
– Terminology, Criticality levels and Solutions Overview
• Q&A
Industry Evolution & Data Centres
Digitisation and IoT/IoE
7
Global – Data Centre / Cloud Traffic Forecast (2014-2019)
Data Centre Traffic Highlights Cloud Traffic Highlights
Globally, data centre traffic will reach 10.4 Zettabytes per year Globally, cloud data centre traffic will represent 83% of total data
(863 Exabytes per month) by 2019, up from 3.4 Zettabytes per centre traffic by 2019, compared to 61% in 2014.
year (287 Exabytes per month) in 2014.
Globally, cloud data centre traffic will reach 8.6 Zettabytes per
Globally, data centre traffic will grow 3.0-fold by 2019, at a year (719 Exabytes per month) by 2019, up from 2.1 Zettabytes
CAGR of 25% from 2014 to 2019. per year (176 Exabytes per month) in 2014.
Globally, data centre traffic grew 35% in 2014, up from 2.6 Globally, cloud data centre traffic will grow 4.1-fold by 2019, at
Zettabytes per year (213 Exabytes per month) in 2013. a CAGR of 33% from 2014 to 2019.
Globally, 73.1% of data centre traffic will remain within the data Globally, cloud data centre traffic grew 52% in 2014, up from 1.4
centre by 2019, compared to 75.4% in 2014. Zettabytes per year (116 Exabytes per month) in 2013.
Globally, 18.2% of data centre traffic will travel to end users by Globally, consumer will represent 69% of cloud data centre traffic
2019, compared to 17.8% in 2014. by 2019, compared to 63% in 2014.
Globally, 8.7% of data centre traffic will travel between data
centres by 2019, compared to 6.8% in 2014.
http://www.cisco.com/c/dam/assets/sol/sp/gci/global-cloud-index-infographic.html
Two Market Transitions – One DC Network
Applications PaaS
Infrastructure HyperScale
Traditional Application Centric Data Centres
Data Centre Infrastructure (ACI)
Networking
Network
Network + Services
DC Abstraction & Automation
Switching Apps Policy
The App Market Transition – From Traditional to Cloud-native
IaaS PaaS
Terminology
• The Terminology around Workload and Business Availability / Continuity is not
always consistent
• Some examples:
“Availability Zone”
• AWS - Availability Zones are distinct locations within a region that are engineered to be
isolated from failures in other Availability Zones
• OpenStack - An availability zone is commonly used to identify a set of servers that have a
common attribute. For instance, if some of the racks in your data centre are on a
separate power source, you can put servers in those racks in their own availability zone.
Availability Zone & Regions – AWS Definitions
Time to Recover
Data Lost 7
p.m.
Application Resiliency and Business Criticality Levels
Defining How A Service Outage Impacts Business Will Dictate A Redundancy Strategy (And Cost)
Each Data Centre should accommodate all levels… Cost is important factor
Business GSLB
Stateless
Recovery Disaster Recovery Plan Geo-clusters
Network Service
HA Cluster
Operation Cost Data Centre Maintenance / DCI LAN extension
LAN Extension
Containment Migration / Consolidation DCI Layer 3
Stateful
Disaster Avoidance Metro Virtual DC
Business Continuity Bandwidth
Virtualisation
Resource & Optimisation Workload Elasticity Hair-pinning
VM mobility
Latency
Inter-Cloud Networking VM Mobility
Cloud Services Flexibility Virtualisation
XaaS Automation
Application Centric View of Data Centre Interconnect (DCI)
• Applications consume resources across the Cloud DC infrastructure
• If an Application moves between sites, each element of the Application
Environment must also adjust to the new location
• DCI extends the Application Environment between Geographic sites within
Private Clouds and Public Clouds
• Critical IT Use Cases including Business Continuity, Workload Mobility, and
Disaster Recovery within Public and Private Clouds, impact each element of the
Application Environment
Application Environment
Cisco Public 24
Ex.2: Requirements for Metro/Geo Data Centres – Cold Migration
Move a Stopped Virtualised Workload across Metro/Geo DCs, Reboot Machine at New Site
Cisco Public
Cloud-Native Apps: Why does it Matter to Customers ?
Flexibility as Cloud native applications are fully portable (not dependent of a particular
infrastructure or cloud implementation)
Why / Where are customers Struggling on this Transition ?
Operations
Traditional Cloud-Native
“Reliable” “Agile”
Policy-based Management (Contiv)
http://www.contiv.io
Ex.3: Multi-site Abstraction and Portability of Network Metadata and
Cloud-native Applications Based on Micro-services (with Docker Containers)
• Q&A
Data Centre Interconnect
SAN Extension
• Asynchronous Data replication: The Application receives the acknowledgement for I/O complete as soon as the
primary disk is updated while the copy continues to the remote disk.
• Unlimited distances
Synchronous Asynchronous
Data Replication Data Replication
4 1 2 1
2 3
3
Storage Deployment in DCI
Option 1 - Shared Storage
Core Network
DC 1 DC 2
Initiator
ESX-A source ESX-B target
Virtual Center
Volumes
Target
Storage Deployment in DCI
Shared Storage Improvement Using Cisco IOA
Core Network
DC 1 DC 2
Virtual Centre
Core Network
DC 1 DC 2
NAS Read
Write ?
2 Temp
data
Read
Write
ESX-A source data 3 Cache ESX-B target
data
ACK 1
4
2 data
ACK
Virtual Center
FlexCache does NOT act as a write-back cache
FlexCache responds to the Host only if/when the original subsystem ack’ed to it
No imperative need to protect a Flexcache from a power Failure
http://www.cisco.com/en/US/docs/solutions/Enterprise/Data_Center/DCI/4.0/Netapp/dciNetapp.html
Storage Deployment in DCI
Option 3 - EMC VPLEX Metro (Active/Active)
Synchronous Latency
Distributed Virtual Volume
Fibre Channel
DC A DC B
Storage Deployment in DCI
Option 3 - EMC VPLEX Metro (Active/Active)
Core Network
DC 1 DC 2
Initiator
ESX-A source ESX-B target
Virtual Centre
Target
From the Storage
LUNv LUNv
EMC
Initiator CLARiiON
EMC
VMAX
VPLEX Synchronous Latency requiments 5ms max VPLEX
Target Engine Engine
http://www.cisco.com/en/US/docs/solutions/Enterprise/Data_Center/DCI/4.0/ EMC/dciEmc.html
Storage Deployment in DCI
Option 4 - EMC VPLEX Geo (Active/Active)
Active/Active Storage Virtualisation Platform for the Private and
Hybrid Cloud
Enables workloads Mobility over Long distance at Asynchronous
distances using Microsoft Hyper-V.
DC A DC B
http://www.emc.com/collateral/hardware/white-papers/h8214-application-mobility-vplex-geo-wp.pdf
Evolution of Storage @ the Tale of 2 ITs
SPEED
EFFICIENCY
Operational Simplicity
Applications Operations Infrastructure
Traditional Emerging Mode 1 Mode 2 Array-Based Server-Based
VM VM VM VM VM
+ APP
+ +
3 Simplified Consumption
Cisco UCS Director Storage Orchestration
UCS Manager Storage Profiles
UCS M-Series
Fourth Generation UCS HyperFlex Systems Modular Servers
UCS C3160
INTEGRATED HYPERCONVERGED
UCS MINI UCS B-SERIES SCALE OUT
INFRASTRUCTURE
UCS as the common platform for heterogeneous workloads (with Storage) in the Data Centre
Contiv Storage Provides Policy-Rich Container Storage that
Leverages Ceph/NFS Underneath - volplugin
"volplugin", despite the name, is actually a suite of
components:
https://docs.docker.com/engine/extend/plugins/
Agenda
• Active-Active (A/A) Data Centre:
– Market & Business Drivers
– Terminology, Criticality levels and Solutions Overview
• Q&A
Extending Virtual Tenant Space outside the Fabric
Logical view of Multi-tier Applications
IP network IP network
Maintain the L2 segmentation End
to End toward the remote DC
Maintain the L3
segmentation of the
Layer 3 Edge public networks (VRF) Layer 3 Edge Gateway
Gateway to the outside world
Web Web
App App
DB DB
MPLS Transport
EoMPLS
Transparent point to point
VPLS SP style
MPLS
Large scale & Multi-tenants, Point to Multipoint
PBB-EVPN
Large scale & Multi-tenants, Point to Multipoint
IP Transport
OTV
Enterprise style Inter-site MAC Routing IP style
IP LISP
For Subnet extension and Path Optimisation
VXLAN/EVPN – evolving
Emerging A/A site interconnect (Layer 2 only or with Anycast L3 gateway)
DCI LAN Extension
IP-Based Solution
OTV
Overlay Transport Virtualisation
Technology Pillars
OTV is a “MAC in IP” technique to extend Layer 2 domains
OVER ANY TRANSPORT
4
VLAN MAC IF
1 100 MAC A IP A
3 New MACs are
learned on VLAN 100
OTV updates exchanged via 100 MAC B IP A
the L3 core
Vlan 100 MAC A 3 300 MAC C IP A
300 MAC C IP A
South 58
Overlay Transport Virtualisation
Inter-Sites Packet Flow
4
MAC TABLE Transport MAC TABLE
VLAN MAC IF Infrastructure VLAN MAC IF
Decap
100 MAC 1 Eth 2 IP A IP B 100 MAC 1 IP A
3 5
100 OTV MAC 2 Eth 1 OTV
Encap OTV
100 MAC 2 OTV
IP A 6
100 MAC 3 IP B MAC 1 MAC 3 IP A IP B
MAC 1 MAC 3 IP A IP B Layer 2
100 MAC 3 Eth 3
Lookup
100 MAC 4 IP B 100 MAC 4 Eth 4
MAC 1 MAC 3
MAC 1 MAC 3 West East
Server 1 Site Site Server 3
1 7
59
Placement of the OTV Edge Device
Option 1 | 2: OTV in the DC Aggregation
L2-L3 boundary at aggregation
DC Core performs only L3 role
STP and L2 broadcast Domains isolated between PODs
Intra-DC and Inter-DCs LAN extension provided by OTV
SVIs SVIs SVIs SVIs
vPC vPC
The Default Gateway (SVI) are distributed among the Leafs (Anycast Gateway)
The Firewalls host the Default Gateway
No SVIs at the Aggregation Layer or DCI Layer
No Need for the OTV VDC
Spine
L3
Border-Leaf
L2
Core
Leaf
OTV OTV OTV OTV
Def Def
L3 GWY GWY
Anycast L3 GWY
OTV DCI Layer L2
Aggregation
Firewall Firewall
OTV
Summary
Extensions over any transport (IP, MPLS)
Failure boundary preservation
Only 5 CLI
Site independence commands
Single Home Device (SHD) Dual Home Device (DHD) Dual Home Device (DHD) Dual Home Network (DHN)
Single Home Network (SHN) Active / Active Per-Flow LB Active / Active Per-Service LB Active / Active Per-Service LB
PE1
PE1 PE1
Bundle-
CE1 Eth25
Bundle-
Bundle- Eth25
Eth25 VID X
PE1 VID X
VID X Bundle- CE1
CE1 Eth1 CE1 MST-AG / REP-AG
MPLS MPLS MPLS MPLS
MST on N-PE
Core Core Core Core
G.8032
P-mLACP
VID X
VID X VID Y
Bundle- Bundle-
Eth25 Eth25
Bundle-
CE2 Eth25
PE2 PE2 PE2
69
Optimise L2 Forwarding with PBB-EVPN
PBB-EVPN
IP-VPN/L3
7K/6K/5K
FEX 2K
2-layers architecture, high control plane scale, small-medium physical ports scale
More ToR, more Tenant scale
ASR 9K MACSec: 100G Linerate encryption
•Linerate link level MACSec encryption on every 100G port on the linecard: Industry best solution
•Supports the max security protection available commercially: NSA suite-B standards
•Supports both P2P and P2MP encryption over the cloud
•Other devices in the cloud could be MACSec agnostic (act as pass through) and still get end to end encryption
benefits
ASR9000
MPLS Cloud PE
CE
ASR9000 PE
CE
EoMPLS/VPLS, E-VPN
MACSec
PE
Payload VLAN MACSec Inner D/S Payload VLAN MACSec Inner D/S MPLS Outer D/S Payload VLAN MACSec Inner D/S
Header MAC Header MAC Labels MAC Header MAC
Hardware based:
o One global L3 only fabric
o Anycast VTEP L2 or L3 gateway distributed
o VxLAN EVPN (ToR)
Fabric
o ACI (ToR)
Metro Distance – Dark fibre / DWDM
o VLAN translation with local VLAN significant
Network Dual-Fabric
Considerations
Any Distance
N-S Traffic localisation is a choice between efficiency (latency-sensitive Application) and elaboration
Connecting a VXLAN
fabric to Outside
Overview and Models
Host Reachability is End-to-End
L3 Spine
VXLAN Leaf
Intra/Inter-DC L3 Spine
End-nodes
VXLAN Leaf
vPC Domain
DCI Leaf nodes
L2 extension End-2-End
VxLAN/EVPN Fabric 1 VxLAN/EVPN Fabric 2
Why to use it ?
o Scalability
o Delineation for the Fabric interconnection Dot1Q Layer 3 Dot1Q
DCI/L2
What is the Cisco Value of it ? Host Reachability is local to each site Host Reachability is local to each site
o Multi-site Anycast L3 Gateway
o Storm Control, BPDU Guard
o DCI technology agnostic (Over IP, Over MPLS, DWDM…)
* Do NOT terminate the overlay tunnel
Dual VXLAN Fabrics interconnection
Inter-Fabric Connectivity using DCI solution
V V V V V V
91
Dual VXLAN Fabrics interconnection
Inter-Fabric Connectivity using OTV solution
EVPN
VXLAN/EVPN
iBGP
EVPN
• Initiated from the Spine or the Border Leaf
iBGP
VNI 30000 VNI 31000
V V V V V V
92
Agenda
• Active-Active (A/A) Data Centre:
– Market & Business Drivers
– Terminology, Criticality levels and Solutions Overview
• Q&A
Differentiation for Nexus/ACI Solutions with Contiv
App1 App2
Application Composition
+ • Contiv.io is an open-source project that creates a
Policy Intent policy framework in different domains of containers
• Network Policies: Policies for Application Security,
Contiv Master Prioritisation, and Network Resource Allocation
Docker | Kubernetes | • Network Services for Apps (Virtual or Physical
Mesos Plugin Agents Service appliances)
• Analytics/Diagnostics
• Integrates with Cisco ACI, Nexus, and UCS
Solutions
Node 1 Node2 Node-n
• Status: Beta
Policy-Based Container Networking with Project Contiv
– Netplugin (1/2)
• Contiv netplugin is a generic networking plugin that is designed to provide multi host policy based networking for containerised
applications.
• Netplugin is designed to be used with docker containers and with cluster schedulers like Swarm, Kubernetes and Mesos
https://docs.docker.com/engine/extend/plugins/
Policy-Based Container Networking with Project Contiv
– Netplugin (2/2)
Contiv Object Model:
• Contiv object model provides a way for users to specify their Intent.
• Netmaster provides a REST API to view and modify contiv object model.
• Netctl command-line tool is a convenient utility to interact with the object model
Network Related
How Does it Work? ACI and Contiv Plugin
Agenda
• Active-Active (A/A) Data Centre:
– Market & Business Drivers
– Terminology, Criticality levels and Solutions Overview
• Q&A
Fabric Infrastructure
Important Concepts – Inside and Outside APIC
‘Outside’ EPG associated Forwarding Policy for ‘inside’ EPG’s defined by associated
with external network Bridge Domain network policies
policies (OSPF, BGP, …
peering)
Web App DB
L2/L3
POD ‘A’ IP Network POD ‘n’ Site ‘A’ IP Network Site ‘n’
MP-BGP - EVPN
… MP-BGP - EVPN
APIC Cluster
Stretched ACI Fabric
ACI Stretched Fabric
DC Site 1 DC Site 2
vCenter
Fabric stretched to two sites works as a Work with one or more transit leaf per site any
single fabric deployed within a DC leaf node can be a transit leaf
One APIC cluster one management and Number of transit leaf and links dictated by
configuration point redundancy and bandwidth capacity decision
Anycast GW on all leaf switches
Stretched ACI Fabric
Support for 3 Interconnected Sites Site 2
Site 1
Site 3
Transit Leaf
2x40G or 4x40G
ACI Multi-POD Solution
Overview Inter-POD Network
MP-BGP - EVPN
Multiple ACI PODs connected by an IP Inter-POD Forwarding control plane (IS-IS, COOP) fault
L3 network, each POD consists of leaf and spine isolation
nodes Data Plane VXLAN encapsulation between
Managed by a single APIC Cluster PODs
Single Management and Policy Domain End-to-end policy enforcement
ACI Multi-POD Solution
Use Cases
Handling 3-tiers physical POD
cabling layout Inter-POD
Leaf Nodes Network
Cable constrain (multiple buildings,
campus, metro) requires a second
tier of “spines” Spine Nodes
Preferred option when compared to
ToR FEX deployment
Software
The solution will be available from Q3CY16 SW Release
Hardware
The Multi-POD solution can be supported with all currently shipping Nexus
9000 platforms
The requirement is to use multicast in the Inter-POD Network for handling BUM
(L2 Broadcast, Unknown Unicast, Multicast) traffic across PODs
ACI Multi-POD Solution
Supported Topologies
Intra-DC Two DC sites connected
back2back
10G/40G/100G
40G/100G 40G/100G
POD 1 POD n POD 1 40G/100G 40G/100G
POD 2
Dark fibre/DWDM (up to
10 msec RTT)
…
APIC Cluster
DB Web/App Web/App DB
APIC Cluster
Web/App Web/App
L3
40G/100G 40G/100G
40G/100G (up to 10 msec RTT)
POD 3
ACI Multi-POD Solution
Inter-POD Network (IPN) Requirements
Not managed by APIC, must be pre-configured
IPN topology can be arbitrary, not mandatory to POD ‘A’ 40G/100G 40G/100G
POD ‘B’
connect to all spine nodes
MP-BGP - EVPN
Main requirements:
• 40G/100G interfaces to connect to the spine nodes DB APIC Cluster
Web/App Web/App
• OSPF to peer with the spine nodes and learn VTEP reachability
Web/App Web/App
DB DB
Single APIC Cluster
ACI Multi-POD Solution
Summary
L2/L3
DCI
Independent ACI Fabrics interconnected via L2 Data Plane VXLAN encapsulation terminated
and L3 DCI technologies at the edge of each Fabric
Each ACI Fabric is independently managed by a VLAN hand-off to the DCI devices for providing
Layer 2 extension service
separate APIC cluster
Requires to classify inbound traffic for
Separate Management and Policy Domains
providing end-to-end policy extensibility
Multi-Site Fabrics
Reachability
Fabric ‘A’ Fabric ‘B’
mBGP - EVPN
Web1 Web2 Import Web & App Export Web & App Web1 Web2
from Fabric ‘B’ to Fabric ‘A’
Web1 Web2 Import Web & App Export Web & App Web1 Web2
from Fabric ‘B’ to Fabric ‘A’
WAN
‘Register for EP
notification in ‘WEB1’ EPG Ext-WEB1 App
APP1 WEB1 ‘Ext-WEB1 EPG created in
Subnet/BD ‘A’ ACI remote L3Out
Toolkit
ACI Toolkit Multisite peers with local and remote APIC clusters and specifies:
What local EPG needs to be “exported” to a remote site (‘WEB1’) and its name in the
remote location (‘Ext-WEB1’)
What contracts will be consumed/provided by that EPG in the remote site
The L3Out in the remote sites where to program host routes
Stretched Application Deployment with CliQr CloudCenter & ACI
CliQr CloudCenter
Northbound API
Cisco ACI
http://goo.gl/CU6MSb
Four Deployment Topologies
Single Pod Multi Site
1. 2. 3.
Four Deployment Topologies
Single Pod Multi Site
1. 2. 3.
Database Application
• Q&A
• Historically this has been
Network Service Placement for Metro Distances well accepted for most of
Metro Virtual DC (Twin-
A/S Stateful Devices Stretched Across 2 Locations – Nominal Workflow DC)
• Almost 80% of Twin-DC
follows this model
L3 Core
• Network Services are usually active on
primary DC
• Distributed pair of Act/Sby FW & SLB on
Outside VLAN
each location
• Additional VLAN Extended for state
FW FT and session synch
synchronisation between peers
Inside VLAN
• Source NAT for SLB VIP
Src-NAT
scenario is limited to 2 sites
SLB session synch
Front-end VLAN
Subnet A
Subnet A Back-end VLAN
Subnet A
Subnet A Back-end VLAN
L3 Core
• FW failover to remote site
• Front-end server farm moves to
remote site
Outside VLAN
• Source NAT for SLB VIP maintains
the return path thru the Active SLB
• Partial move of a server farm is not
optimised
Inside VLAN
• Understand and identify the multi-
tier frameworks
VIP VIP VLAN
Src-NAT
Front-end VLAN
Subnet A Subnet A
Back-end VLAN
Src-NAT
HSRP
Front-end VLAN Filter
Subnet A Subnet A
Back-end VLAN
L3 Core
Front-End VLAN
Subnet A Subnet A
Back-End VLAN
Subnet A / 24
Subnet A / 25
Subnet B / 25
Subnet A / 24
Subnet B / 24
Subnet A / 25
M-DB
2 - Request is attracted by DC-1 (best metric) for subnet A
(/25) and DC-2 is primary DC for Subnet B
3 – Apps migrates and restarts on DC-2 (Cold move)
4 – LISP IGP Assist detects the new EID and map-notifies its
Map-Register local LISP XTR (DC-2)
CCL CCL
5 –LISP FHR DC-2 Redistribute LISP routes into IGP
6 – Installs a /32 LISP Interface Route of the host (EID)
7 – Sends a Map-register of new EID to the MS
Redistribute 8 – MS sends a Map-notifies to LISP FHR (DC-1)
Map-Notify Redistribute
LISP route into LISP route into
9 – LISP FHR DC-1 Redistribute LISP routes into IGP and
• Q&A
Q&A
Complete Your Online Session Evaluation
Give us your feedback and receive a
Cisco 2016 T-Shirt by completing the
Overall Event Survey and 5 Session
Evaluations.
– Directly from your mobile device on the Cisco Live Mobile
App
– By visiting the Cisco Live Mobile Site http://showcase.genie-
connect.com/ciscolivemelbourne2016/
– Visit any Cisco Live Internet Station located throughout
the venue
Learn online with Cisco Live!
T-Shirts can be collected Friday 11 March Visit us online after the conference
for full access to session videos and
at Registration presentations.
www.CiscoLiveAPAC.com
Thank you