Documente Academic
Documente Profesional
Documente Cultură
On your browser:
https://clsandiego15.cnf.io/ On your browser:
Search for BRKACI-3503
Extending ACI to Multiple Sites
Dual Site Deployment Deep Dive
Santiago Freitas (safreita@cisco.com), Customer Solutions Architect
Patrice Bellagamba (pbellaga@cisco.com), Distinguished Systems Engineer
BRKACI-3503
Multi-Fabric Design Options
Single APIC Cluster / Single Domain Multiple APIC Clusters / Multiple Domains
DB Web
App
L2 L2
DCI DCI
DB Web
App
Stretched Fabric
Supported Distances and Interconnection Technologies
Stretched ACI Fabric
• Single fabric stretched to two sites. Works the same way as Single fabric deployed within a single DC
• One APIC cluster. One management and configuration point.
• Anycast GW on all leaf switches.
• Work with one or more transit leaf per site. Any leaf can be transit leaf.
• Number of transit leaf and links is redundancy and bandwidth capacity decision
Supported Distances and Interconnection Technologies
Dark Fiber
IS-IS adj
• Port mode EoMPLS used to stretch the ACI fabric over long distance.
• DC Interconnect links could be 10G (minimum) or higher with 40G facing the Leafs / Spines
• DWDM or Dark Fiber provides connectivity between two sites.
• 1.0(3f) release or later, Max 10ms RTT between sites.
• Under normal conditions 10 ms allows us to support two DCs up to 800 KMs apart.
• Other ports on the Router used for connecting to the WAN via L3Out
Please provide your input to the questions asked on the mobile App.
• ASR9K performs
• EoMPLS port Xconnect
• Speed Adaptation with QoS
mpls ldp
log
hello-adjacency
graceful-restart
!
router-id 5.5.5.1
interface TenGigE0/2/1/0
Speed Adaptation with QoS
• Due to the difference of speed (40G => 10G), QoS is a must on DCI
• Fabric Control Traffic must be protected
• Demonstration of stability versus link overload.
class APIC+Traceroute_Traffic
priority level 2 interface TenGigE0/2/1/0
police rate percent 15 description To-ASR9k-4
cdp
class MPLS_CE_BGP+ASA+vASA+vF5_HA_Traffic mtu 9216
bandwidth 500 mbps service-policy output QoS_Out_to_10G_DCI_Network
queue-limit 40 kbytes ipv4 address 5.5.2.1 255.255.255.252
load-interval 30
class User_Data_Traffic_1
bandwidth 3200 mbps
queue-limit 40 kbytes
class User_Data_Traffic_2
bandwidth 3200 mbps
queue-limit 40 kbytes
class SPAN_Traffic
bandwidth 100 mbps
queue-limit 40 kbytes
class class-default
VMM Integration
• EPG to EPG atomic counter works when the EPG is not present on transit leaf.
• Other Atomic Counters works fine: ALE2 = N9396PX, N9396TX,
• Leaf-to-Leaf (or TEP-to-TEP) works with ALE2 based Nexus 9300. N93128TX and N93128PX with 6-port
GEM N9K-6PQ, N9372TX, N9372PX
• Between endpoints (EP to EP). and N9332PQ.
Transit Leaf and WAN Traffic
BD “Blue”
EPG A
BD setup in Layer 2.
Leaf 3, DC1
Leaf 5, DC 2
Logical Topology Deep Dive
WAN-CE to ASA, BGP peering through the Fabric
WAN EPG with L2 BD with static binding towards ASA and WAN CE Even numbered tenants use the
primary path into/out of the fabric
ASA/T4/act(config)#route-map set-localpref-200-inprefixes permit 10
ASA/T4/act(config-route-map)# set local-preference 200
via DC2 and odd tenants use the
primary path into/out of the fabric
ASA/T4/act(config-if)# interface TenGigabitEthernet0/7.1041
ASA/T4/act(config-if)# nameif outside
via the “left side” DC1
ASA/T4/act(config-if)# ip address 10.1.1.254 255.255.255.0 standby 10.1.1.253
RealWebEPG
Spine 1 == DC 1
Spine 3 == DC 2
The fabric uses MP-BGP to distribute external routes within ACI fabric.
Current SW Release supports a max of two MP-BGP route reflectors.
E-W Flows
1. 10 IP endpoints per EPG
2. WebEPG AppEPG and AppEPG DbEPG
3. Stateless traffic
4. Flows spread across LEAF switches as shown above
5. Data rate (10 Gbps in DC1, 10 Gbps in DC2 and 10 Gbps across DCs)
Traffic Flow – Multicast, Test Traffic
Multicast Flows
1. Per Tenant (T3 and T4), 500 Groups and 500 Flows
2. Traffic Rate: Sent at 500 Mbps
3. Stateless Traffic
4. Intra-EPG traffic
VMotion
VMs on
same DCs
VMs on
different
DCs
Firewall and Load Balancer failover
Improving ASA failover time
- Virtual F5 failure
- Failure: 8 seconds
- Failback: no losses
Intra-DC Link Failure
Spine 1 Failed/Restored
Powered Off
RP/0/RSP0/CPU0:DC1-ASR9K-1#sh log
(snip)
LC/0/2/CPU0:Apr 14 10:17:43.609 : vic_0[365]: %L2-ETHERNET-3-TX_DISABLE :
Interface FortyGigE0/2/0/0, link no longer forced down due to remote signalling
• APIC’s in DC1 takes around 10-15 seconds to realize loss of reachability to all
APIC/fabric nodes in DC2.
• DC1 APIC controllers can execute policy read and write operations.
• DC2 fabric nodes were able to learn endpoints in data plane. No disruption.
- DC2 APIC Controller doesn’t show the learnt endpoint.
• APIC controllers in DC1 and DC2 synced up and APIC controller in DC2
started showing the learnt endpoint.
• The IBGP sessions between MPLS CE routers and ASA’s going over DCI
PW links were recovered.
• Overload the fabric with user traffic, i.e. COS 0 or COS 1, by sending more than DCI links
can handle.
QoS on ASR 9K engaged to protect SUP_Traffic (COS 5), APIC+Traceroute_Traffic (COS 3 and 6),
MPLS_CE_BGP+vServiceNodes_HA_Traffic (COS 2) and limit SPAN_Traffic (COS 4 and 7).
• When site 1 goes down, user can access and monitor the ACI fabric via the controller in
site 2 but user can’t make configuration changes.
Data Center Failure
Restoring ability to make configuration changes
DEMO available
Stretched Fabric Link failures – https://www.youtube.com/watch?v=xgxPQNR_42c
vMotion over Stretched Fabric with EoMPLS - https://www.youtube.com/watch?v=RLkryVvzFM0
ACI Multi-Site
Multiple APIC Clusters / Multiple Domains
Disclaimer
The solutions presented from this slide onwards are still under testing /
validations.
Target: Q4CY2015.
Please contact the presenters if you need to perform a Proof of
Concept earlier.
Dual-Fabric Design Scenarios
• Two independent ACI fabrics.
Two management and configuration
domains.
• Design Goals:
• Active/Active workload.
• Extend L2 and subnet across sites.
• Anycast GW on both fabrics
• Interconnect Technologies:
• Dark Fiber or DWDM (back to back vPC)
• VXLAN/OTV/VPLS/PBB for L2 extension over IP
Dual-Fabric with Common Anycast GW IP
• Multiple Anycast GW IP assigned on ACI for same subnet
• Unique Primary IP and common secondary IP for same subnet between Fabrics
• Different GW MAC per Site.
• Unique SVI MAC and common virtual MAC (roadmap Q4CY2015)
• On the Bridge Domain, ARP and L2 Unknown Unicast Flood must be enabled.
vCenter vCenter
Server VLAN Server
VLAN VLAN 300 VLAN
VLAN
100 100 200
200
vCenter vCenter
Server VLAN Server
VLAN VLAN 300 VLAN
VLAN
100 100 200
200
APIC APIC
EPG static
binding
L3
DCI ESX-DC2
ESX-DC1 VLAN to
EPG static
VXLAN
DVS-DC1 Nexus 9300 binding
mapping DVS-DC2
Server 2
NX-OS Mode
Server 1
10.1.5.81 10.1.5.92
EPG static
VXLAN overlay
withbinding
BGP-EVPN
Tech Preview
VXLAN Overlay – BGP EVPN Peering
For Layer 2 DCI Extension
VTEP, anycast IP
facing vPC edge
• Anycast VTEP
• Virtual Tunnel End Point
Nexus
• VXLAN Src/Dest 9300
• Next-Hop in BGP EVPN
Address-Family
VXLAN encap
MP-BGP
EVPN
Cross Fabric L3 Extension
• Not all EPGs have
to be extended
• Some subnets are
local to a fabric.
• L3 Peering
between the
Fabrics is required.
• ACI support iBGP
or OSPF with 11.0
release
• eBGP on 11.1
EPG to EPG Policy Synchronization across sites
Policy ACI Toolkit Policy
• You can submit an entry for more than one of your “favorite” speakers
• Don’t forget to follow @CiscoLive and @CiscoPress
• View the official rules at http://bit.ly/CLUSwin
Complete Your Online Session Evaluation
• Give us your feedback to be
entered into a Daily Survey
Drawing. A daily winner
will receive a $750 Amazon
gift card.
• Complete your session surveys
though the Cisco Live mobile
app or your computer on
Cisco Live Connect.
Don’t forget: Cisco Live sessions will be available
for viewing on-demand after the event at
CiscoLive.com/Online
Continue Your Education
• Demos in the Cisco campus
• Walk-in Self-Paced Labs
• Table Topics
• Meet the Engineer 1:1 meetings
• Related sessions
Thank you