Sunteți pe pagina 1din 102

Data Centre

Networking—
Architecture and
Design Guidelines
BRKDCT-2001

Ian Bond

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 1
HOUSEKEEPING
ƒ We value your feedback, don’t forget to complete your online session
evaluations after each session and complete the Overall Conference
Evaluation which will be available online from Friday.

ƒ Visit the World of Solutions on Level -01!

ƒ Please remember this is a ‘No Smoking’ venue!

ƒ Please switch off your mobile phones!


ƒ Please remember to wear your badge at all times including the Party!
ƒ Do you have a question? Feel free to ask them during the Q&A section or
write your question on the Question form given to you and hand it to the
Room Monitor when you see them holding up the Q&A sign.

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 2
Before We Get Started:
ƒ Q and A at end of session
ƒ Intermediate level session focused on data centre front
end architecture for the transactional model
ƒ Other recommended sessions:
BRKDCT-2002 : Data Centre IP Front-End : Solution for Business Continuance
BRKDCT-2003 : High-Density Server Farms
BRKDCT-2004 : Data Centre Fibre Channel Back End Infrastructure: Solutions
for Disaster Recovery
BRKDCT-2005 : Design and Deployment of Layer 2 Clusters and Geoclusters
BRKDCT-2006 : Introduction to High-Performance Computing
BRKDCT-2007 : Data Centre Optical Infrastructure for the Enterprise
BRKDCT-3008 : Advanced SAN Fabrics and Storage Virtualisation
BRKDCT-3009 : Advanced Understanding of Infiniband Technology
TECDCT-1001 : Architecting Data Centre Infrastructure and Services

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 3
Agenda
Data Centre Infrastructure
ƒ Overview of Cisco Data Centre Network Architecture
ƒ Core Layer Design
ƒ Aggregation Layer Design
ƒ Access Layer Design
ƒ Density and Scalability Implications
ƒ Scaling Bandwidth and Density
ƒ Spanning Tree Design and Scalability
ƒ Increasing HA in the DC
ƒ A look at the Future…

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 4
Cisco Data
Centre
Networking
Architecture

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 5
Cisco Data Centre Network Architecture –
Intelligent Information Network and
Service Oriented Network Architecture

Cisco 3-5 Year


Technology Vision IIN Why?

Architecture Framework
for implementing IIN in SONA What?
the Enterprise

Enterprise Data MAN / Branch


Architectures Campus Branch How?
Centre WAN of One
based on SONA

Technologies Security, IPC, Application Networking,


Storage, Routing, Switching, Network
Mgmt
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 6
Cisco Data Centre Network Architecture
Instant Unified Meeting
PLM
Business
CRM ERP
Messaging Collaboration
Messaging Place

HCM Applications
Procurement SCM IPCC Applications
IP Phone Video
Delivery
Traditional Architecture / Service Oriented Architecture

Advanced Analytics and Decision Support


Services Management

ANALYTICS & ADAPTIVE


ApplicationNetworking
Application Networking Services
Services
WAAS, App Acceleration,
Application Delivery Services
INTERACTIVE

Optimisation,
Optimisation, Security and Server Offload
SERVICES
LAYER

Infrastructure Enhancing Services

POLICY
Firewalls, Intrusion Protection,
Security
Security Services
Agents
Infrastructure
RDMA,
Enhancing Services
Virtualisation, Replication,
Compute
Low LatencyServices
Clustering Storage Fabric Services
Virtual Fabrics
Infrastructure Management

Compute Network Storage Network


INFRASTRUCTURE

Network Virtualisation Services


NETWORKED

Server Server Storage Data Centre


LAYER

Fabric Data
Switching Enterprise
Switching Interconnect
Campus Branch Places in the Network WAN/MAN Teleworker
Centre
Modular EdgeDirector DWDM,
Infiniband Rack SONET,
Switching Fabric
Blade SDH, FCIP
Server Storage Clients
SFS Family Catalyst Family MDS Family ONS Family

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 7
Data Centre Architecture Overview
Layers of the Enterprise Multi-Tier Model

ƒ Layer 2 and layer 3 access Enterprise Core


topologies
ƒ Dual and single attached 1RU DC Core
and blade servers
ƒ Multiple aggregation modules DC Aggregation
ƒ Web/app/database multi-tier
environments
ƒ L2 adjacency requirements DC Access
ƒ Mix of over-subscription
requirements
ƒ Environmental implications Blade Chassis w/
Integrated Switch
ƒ Stateful services for security and
load balancing
Blade Chassis w/ Mainframe
Pass Thru w/ OSA
L2 w/ Clustering L3 Access
and NIC Teaming

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 8
Core Layer
Design

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 9
Core Layer Design
Requirements
Scaling
ƒ Is a separate DC Core Layer Campus Access Layer
required?
Campus
ƒ Consider: Distribution
10GigE port density
Administrative domains
Anticipate future requirements Campus Core

ƒ Key core characteristics


Distributed forwarding architecture DC Core
Low latency switching
Aggregation
10GE scalability
Advanced ECMP
Advanced hashing algorithms
Scalable IP multicast support

Server Farm Access Layer

Scaling

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 10
Core Layer Design
L2/L3 Characteristics

ƒ Layer 2/3 boundaries:


Campus Core
All Layer 3 links at core,
L2 contained at/below
aggregation module
L2 extension through core is DC Core
CEF Hash
not recommended
Applied to
ƒ CEF hashing algorithm Packets on
Equal Cost
Default hash is on L3 IP Routes
addresses only L3
L3 + L4 port is optional and L2
may improve load distribution
Aggregation
CORE1(config)#mls ip cef load full
Leverages automatic source port
randomisation in client TCP stack Access

Web Application Database


Servers Servers Servers

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 11
Core Layer Design
Routing Protocol Design: OSPF
ƒ NSSA helps to limit LSA propagation,
but permits route redistribution (RHI)
Campus Core
ƒ Advertise default into NSSA,
summarise routes out
ƒ OSPF default reference b/w is 100M, Area 0
use “auto-cost reference-bandwidth” NSSA DC Core
set to 10G value L0=10.10.1.1
Default Default
L0=10.10.2.2

ƒ VLANs on 10GE trunks have OSPF DC Subnets


cost = 1G (cost 1000), adjust (Summarised)
bandwidth value to reflect 10GE for L3 vlan-ospf

interswitch L3 vlan
Aggregation
ƒ Loopback interfaces simplify
troubleshooting (neighbor ID)
L0=10.10.3.3 L0=10.10.4.4
ƒ Use passive-network default: open
up only links to allow Access
ƒ Use authentication: more secure and
avoids undesired adjacencies
ƒ Timers spf 1/1, interface hello-dead =
1/3 Web Application Database
Servers Servers Servers

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 12
Core Layer Design
Routing Protocol Design: EIGRP

ƒ Advertise default into DC with


Campus Core
interface command on core:
ip summary-address eigrp 20 0.0.0.0 0.0.0.0 200
Cost of 200 required to be preferred
route over the NULL0 route DC Core
installed by EIGRP
Default Default
DC Subnets
ƒ If other default routes exist (from (Summarised)
internet edge for example), may L3 vlan-eigrp

need to use distribute lists to


filter out Aggregation

ƒ Use passive-network default


ƒ Summarise DC subnets to core Access
with interface command on agg:

ip summary-address eigrp 20 10.20.0.0 255.255.0.0


Web Application Database
Servers Servers Servers

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 13
Aggregation Layer
Design

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 14
Aggregation Layer Design
Spanning Tree Design

ƒ Rapid-PVST+ (802.1w) or Core


MIST (802.1s), Root Secondary
ƒ Choice of .1w/.1s based on scale Root Primary
HSRP Secondary
Standby Context
of logical+virtual ports required HSRP Primary
Active Context
ƒ R-PVST+ is recommended and
best replacement for 802.1d
Fast converging: inherits proprietary
enhancements (Uplink-fast,
Backbone-fast)
Access layer uplink failures: ~300ms –
2sec
Most flexible design options
Combined with RootGuard,
BPDUGuard, LoopGuard ,and UDLD
achieves most stable STP
environment
UDLD global only enables on Fiber
ports, must enable manually on
copper ports
Rootguard
LoopGuard
BPDU Guard
UDLD Global

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 15
Aggregation Layer Design
Integrated Services
Services: Firewall, Load Balancing, SSL
Encryption/Decryption

+
ƒ L4-L7 services integrated in Cisco Catalyst® 6500
ƒ Server load balancing, firewall and SSL services may be
deployed in:
Active-standby pairs (CSM, FWSM 2.X)
Active-active pairs (ACE, FWSM 3.1)
ƒ Integrated blades optimise rack space, cabling, mgmt, providing flexibility
and economies of scale
ƒ Influences many aspects of overall design
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 16
Aggregation Layer Design
Service Module Placement Consideration
Cisco Catalyst 6500 Switch Fabric Channels Cisco Catalyst 6513
6-Slot 9-Slot 13-Slot

Dual Dual Single Single Channel Fabric Attached Modules


Dual Dual Single Sup720, ACE, 6724, 6516
FWSM, SSLSM, NAM-2, IDSM-2
Dual Dual Single

Dual Dual Single Classic Bus Modules (No Channel)


Dual Dual Single CSM, IDSM-1, NAM-1, 61xx-64xx series
Single
* Dual * Dual (6704, 6708, 6748 Not Permitted
Dual Single
in These Slots)
Dual * Single
Dual Dual
Dual Channel Fabric Attached Modules
Dual
6748, 6704, 6708
Dual

Dual (Supports All Single Channel and Classic


Bus Modules Also)
Dual

The Choice Between 6509 and 6513 Usually Comes Down to:
6513: Supports Larger Number of Service Modules
6509: Supports Larger Number of 10GE Ports * Primary Sup720 Placement

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 17
Aggregation Layer Design
Active-Standby Service Design

ƒ Active-standby services
Content Switching Module
Core
Firewall Service Module (2.x)
SSL Module Root Primary Root Secondary
HSRP Primary HSRP Secondary
ƒ Under utilises access Active Context Standby Context

layer uplinks
ƒ Under utilises service
modules and switch fabrics
ƒ Multiple service module pairs
does permit active-active
design but….

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 18
Aggregation Layer Design
Active-Active Service Design

ƒ Active-Active Service Design


Application Control Engine
Core
SLB+SSL+FW
4G/8G/16G switch fabric options VLAN5: VLAN6:
Root Primary Root Primary
Active-standby distribution HSRP Primary HSRP Primary
per context Active Context Active Context
Firewall Service Module (3.x)
Two active-standby groups
permit distribution of
contexts across two FWSM’s VLAN 6: VLAN 5:
(not per context) Root Secondary Root Secondary
HSRP Secondary HSRP Secondary
ƒ Permits uplink load balancing Standby Context Standby Context

while having services applied


ƒ Increases overall service
performance vlan6 vlan6

vlan5 vlan6 vlan6 vlan5

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 19
Aggregation Layer Design
Establishing Path Preference
Establish Route Preference for
Service Enabled Applications
ƒ Use Route Health Injection Core
feature of ACE and CSM 3. Route Map on Host
Route Sets Preferred
ƒ Introduce route-map to Metric of Route
route-map RHI permit 10 4. If Context Failover

RHI injected route to set match ip address 44


set metric-type type-1
Occurs, RHI and Route
Preference Follow

desired metric
ƒ Aligns advertised route
of VIP with active context on 2. If Healthy, Installs
ACE, CSM, FWSM and SSL Host Route to VIP
on Local MSFC

service modules
ƒ Avoids unnecessary use
of inter-switch link and
vlan6 vlan6
asymmetrical flows
vlan5 vlan6 vlan6 vlan5
1. ACE Probes to
Real Servers in VIP
to Determine Health

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 20
Core and Aggregation Layer Design
STP, HSRP and Service Context Alignment

ƒ Align server access to


primary components in Core
aggregation layer:
Root Primary Root Secondary
–vlan root HSRP Primary HSRP Secondary
–primary HSRP instance Active Context Standby Context

–active service context


ƒ Provides more
predictable design
Simplifies troubleshooting

ƒ More efficient traffic flow


Decreases chance of flow ping-
pong across inter-switch link

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 21
Aggregation Layer Design
Scaling the Aggregation Layer

ƒ Aggregation modules provide:


Spanning tree scaling
HSRP Scaling Campus Core
Access layer density
10GE/GEC uplinks
DC Core
Application services scaling
SLB/firewall
Fault domain sizing
ƒ Core layer provides inter-agg
module transport:
Low latency distributed forwarding Aggregation Aggregation Aggregation
architecture (use DFC’s) Module 1 Module 2 Module 3
100,000’s PPS forwarding rate Server Farm Access Layer
Provides inter-agg module transport
medium in Scaling
multi-tier model

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 22
Aggregation Layer Design
Using VRFs in the DC (1)

ƒ Enables virtualisation/ MPLS or


partitioning of network VRF-Green other Core
resources (MSFC, VRF-Blue DC Core
ACE, FWSM)
VRF-Red
ƒ Permits use of application services
with multiple access topologies Agg1 Agg2

ƒ Maps well to path isolation Alternate Primary


MAN/WAN designs such as Firewall and SLB Contexts on Agg1
Contexts for Green, and 2 to Achieve
with MPLS Blue, and Red Active-Active Design

ƒ Security policy VLANs Isolate


Contexts on Access
management and deploy-
802.1Q
ment by user group/vrf Trunks

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 23
Aggregation Layer Design
Using VRFs in the DC (2)
Red VRF Red VRF
ƒ Maps the virtualised DC Green VRF Blue VRF Green VRFBlue VRF

to the MPLS Campus WAN/Branch

MAN/WAN cloud
Core: P-Nodes

PE PE
DC Core
Agg module 1 Agg module 2

802.1Q 802.1Q
Trunks Trunks

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 24
Access Layer
Design

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 25
Access Layer Design
Defining Layer 2 Access

ƒ L3 routing is first performed in


the aggregation layer DC Core
ƒ L2 topologies consist of looped
and loop-free models Agg1 Agg2
L3 Inter-Switch Link
ƒ Stateful services at Agg can be L2
provided across the L2 access Primary Root Secondary Root
Primary HSRP Secondary HSRP
(FW, SLB, SSL, etc.) Active Services Standby Services

nks
T ru
ƒ VLANs are extended across

.1q
inter-switch link trunk

802
for looped access
ƒ VLANs are not extended across
inter-switch link trunk for loop-
free access

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 26
Access Layer Design
Establish a Deterministic Model
ƒ Align active components in traffic path on
common Agg switch:
Primary STP root DC Core
Aggregation-1(config)#spanning-tree vlan 1-10 L3+L4 Hash
root primary Path
Pref
Primary HSRP (outbound) L3 Agg1 Agg2
standby 1 priority X L2
Primary Root
Active Service modules/contexts Primary HSRP
Active Services

nks
Path Preference (inbound) Def gwy

T ru
Use RHI & Route map

.1q
802
Route-map preferred-path
match ip address x.x.x.x
set metric -30
(also see slide 15 )

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 27
Access Layer Design
Balancing VLANs on Uplinks: L2 Looped
Access
ƒ Distributes traffic load across
uplinks
DC Core
ƒ STP blocks uplink path L3+L4 Hash
for VLANs to secondary Agg1 Agg2
root switch L3
ƒ If active/standby service L2
Blue-Primary Root Red-Primary Root
modules are used; consider Blue-Primary HSRP Red-Primary HSRP
Red-Sec Root Blue-Sec Root
inter-switch link utilisation to Red-Sec HSRP Def gwy Def gwy Blue-Sec HSRP
reach active instance
Multiple service modules may be
distributed on both agg switches to 802.1q Trunks
achieve balance
Consider b/w of inter-switch link in a
failure situation
ƒ If active/active service modules
are used;
(ACE and FWSM3.1)
Balance contexts+hsrp across agg
switches
Consider establishing inbound path
preference
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 28
Access Layer
Design:
L2 Looped Design
Model

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 29
Access Layer Design
Looped Design Model

ƒ VLANs are extended between Primary STP Root Secondary STP Root
aggregation switches, creating Primary HSRP Secondary HSRP
the looped topology Active Services Standby Services
ƒ Spanning Tree is used to L3 Inter-Switch Link
prevent actual loops (Rapid L2
PVST+, MST) F F
.1Q Trunk
ƒ Redundant path exists through a
F F F F F F F
second path that is blocking
ƒ Two looped topology designs:
Triangle and square
FB FB F F
ƒ VLANs may be load balanced B
across access layer uplinks
ƒ Inter-switch link utilisation must
be considered as this may be
used to reach Looped Triangle Looped Square
active services

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 30
Access Layer Design
L2 Looped Topologies

L2 Looped Triangle L2 Looped Square


L3 L3
L2 L2

VLAN VLAN
10 10

VLAN VLAN VLAN VLAN


10 20 10 20

Row 2, Cabinet 3 Row 9, Cabinet 8 Row 2, Cabinet 3 Row 9, Cabinet 8

Looped Triangle Access Looped Square Access


ƒ Supports VLAN extension/L2 adjacency across
access layer ƒ Supports VLAN extension/L2 adjacency across access
layer
ƒ Resiliency achieved with dual homing and STP
ƒ Resiliency achieved with dual homing and STP
ƒ Quick convergence with 802.1W/S
ƒ Quick convergence with 802.1W/S
ƒ Supports stateful services at aggregation layer
ƒ Supports stateful services at aggregation layer
ƒ Proven and widely used
ƒ Active-active uplinks align well to ACE/FWSM 3.1
ƒ Achieves higher density access layer, optimising 10GE
aggregation layer density
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 31
Access Layer Design
Benefits of L2 Looped Design
ƒ Services like firewall
and load balancing can
easily be deployed at the DC Core
aggregation layer and
shared across multiple L3
access layer switches L2
ƒ VLANs are primarily
contained between pairs
of access switches but─
ƒ VLANs may be extended
to different access
switches to support
NIC teaming
Clustering L2 adjacency
Administrative reasons
Geographical challenges
Row A, Cab 7 Row K, Cab 3

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 32
Access Layer Design
Drawbacks of Layer 2 Looped Design

ƒ Main drawback: if a loop occurs


the network may become
unmanageable due to the infinite
replication
of frames 0000.0000.3333

ƒ 802.1w Rapid PVST+ combined


with STP related features and
best practices improve stability DST MAC 0000.0000.4444
and help to prevent loop
3/2 3/2
conditions
UDLD
Loopguard 3/1 3/1
Rootguard Switch 1 Switch 2
BPDUguard DST MAC 0000.0000.4444
Limited domain size
Stay under STP watermarks for logical
and virtual ports

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 33
Access Layer
Design:
L2 LoopFree
Design Model

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 34
Access Layer Design
Loop-Free Design
ƒ Alternative to looped design
ƒ 2 Models: LoopFree U and LoopFree
Inverted U DC Core
ƒ Benefit: Spanning Tree
is enabled but no port is blocking so L3
all links are forwarding L2
L3
ƒ Benefit: less chance of
L2
loop condition due to
misconfigurations or other anomolies
ƒ L2-L3 boundary varies
by loop-free model used:
U or Inverted-U
ƒ Implication considerations with
service modules, L2 adjacency, and
single attached servers

LoopFree U LoopFree
Inverted U

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 35
Access Layer Design
Loop-Free Topologies

L2 Loop-Free U L2 Loop-Free Inverted U


L3
L2
L3
L2

VLAN
10

VLAN VLAN VLAN VLAN


10 20 10 20

Loop-Free U Access Loop-Free Inverted U Access


ƒ VLANs contained in switch pairs ƒ Supports VLAN extension
(no extension outside of switch pairs) ƒ No STP blocking; all uplinks active
ƒ No STP blocking; all uplinks active ƒ Access switch uplink failure black holes
ƒ Autostate implications for certain service single attached servers
modules (CSM) ƒ Supports all service module
ƒ ACE supports autostate and per context implementations
failover

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 36
Access Layer Design
Loop-Free U Design and Service Modules (1)
Agg1, Active HSRP Agg2, Standby
Services Services
HSRP Secondary on
L3 Agg2 Takes Over as def-
L2 gwy, But How to Reach
Active Service Modules?

MSFC on Agg1

SVI for Interface vlan 10


goes to down state if
uplink is the only
interface in vlan 10

VLAN 10 VLAN 11
VLAN 20, 21
ƒ If the uplink connecting access and aggregation goes down, the VLAN interface on the MSFC
goes down as well due to the way autostate works
ƒ CSM and FWSM has implications as autostate is not conveyed (leaving black hole)
ƒ Tracking and monitoring features may be used to allow failover of service modules based on
uplink failure but would you want a service module failover for one access switch uplink failure?
ƒ Not recommended to use loop-free L2 access with active-standby service module
implementations. (See slide on ACE next)

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 37
Access Layer Design
Loop-Free U Design and Service Modules (2)
ACE Context 1 ACE Context 1
HSRP

L3
L2 HSRP Secondary on Agg2
Takes over as def-gwy

VLAN 10, 11 VLAN 20, 21


With ACE:
ƒ Per context failover with autostate
ƒ If uplink fails to Agg1, ACE can switchover to Agg2 (under 1sec)
ƒ Requires ACE on access trunk side for autostate failover
ƒ May be combined with FWSM3.1 for active-active design
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 38
Access Layer Design
Drawbacks of Loop-Free Inverted-U Design

ƒ Single attached servers


are black-holed if access DC Core
switch uplink fails
L3
ƒ Distributed EtherChannel® L2
can reduce chance of
black holing
ƒ NIC teaming improves
resiliency in this design
ƒ Inter-switch link scaling
needs to be considered when
using active-
standby service modules

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 39
Access Layer Design
Using FlexLinks in the Data Centre
ƒ Flexlinks are an active-standby pair
defined on a common DC Core
access switch
ƒ Flexlink pairs have STP turned “off” so no
BPDU’s are propagated
ƒ Failover in 1-2 second range
ƒ An interface can belong to only 1 FlexLink
pair of same or different interface
types (GE, 10GE, port-channel, etc.)
ƒ Agg switch is not aware of FlexLinks
configured on access switch, links are
up and STP logical/virtual ports are
active/allocated
Router# configure terminal
ƒ Supported as of 12.2.18SXF Router(conf)# interface Gigabit1/1
Router(conf-if)# switchport backup interface Gigabit1/2
ƒ An alternative to triangle Router(conf-if)# exit
loop design but must consider risk of
possible loop (see next slides) Router# show interface switchport backup
Router Backup Interface Pairs:
Active Interface Backup Interface State
------------------------------------------------------------------------
Gigabit1/1 Gigabit1/2 Active Up/Backup Standby
Port-channel1 Gigabit7/1 Active Up/Backup Standby
Gigabit7/2 Gigabit7/3 Active Up/Backup Standby

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 40
Access Layer Design
Considerations when using FlexLinks (1)
ƒ Loop Considerations: 10.20.15.57 10.20.16.27

STP is disabled so it is not a


backup safety mechanism for
Core
possible loop conditions
Root Primary Root Secondary
HSRP Primary HSRP Primary
If a loop is accidentally placed between Active Services Backup Services
access switch ports in same vlan—
BPDUGuard will catch:
Apr 13 16:07:33: %PM-SP-4-ERR_DISABLE:
bpduguard error detected on Gi2/2, putting Po-2 T7/4
Gi2/2 in err-disable state
T7/4 Po-3

NOTE: Without BPDU Guard on


access ports, the Aggregation switch Access1 Access2
will be negatively affected with high G2/1 G2/1
CPU condition, hsrp flapping and other
negative conditions oops!

oops
VLAN 6
10.20.6.67 10.20.6.56
00:A0:D1:E1:DB:7F 00:A0:D1:E1:DB:FD

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 41
Access Layer Design
Considerations when using FlexLinks (2)
ƒ Loop Considerations: 10.20.15.57 10.20.16.27

If a loop is accidentally placed between


the access switch and the aggregation
switch a loop condition will occur. Core

NOTE: This will create negative affects


with high CPU condition, hsrp flapping Root Primary Root Secondary
HSRP Primary HSRP Primary
and other conditions Active Services Backup Services

ƒ Other Considerations
No preempt: need to consider inter- Po-2 T7/4
T7/4 Po-3
switch link b/w for failover
situations
oops!
Access1 Access2
G2/1 G2/1
Backup link is unused and in
standby state which doesn’t align
to active/active service module oops
designs 10.20.6.67
VLAN 6
10.20.6.56
00:A0:D1:E1:DB:7F 00:A0:D1:E1:DB:FD

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 42
Access Layer Design
Comparing Looped, Loop-Free and Flexlinks

VLAN Service Single


Uplink vlans
Extension Module Black- Attached Access Switch Must Consider
on Agg Switch
Supported Holing on Server Black- Density per Inter-Switch
in Blocking or (3)
Across Uplink Failure Holing on Agg Module Link Scaling
Standby State
Access Layer (5) Uplink Failure

(3)
Looped
Triangle
- + + + - +

Looped
Square
+ + + + + -
(4)

(4)
Loop-Free U + - - + + +
(1,2)
Loop-Free (1,2)
Inverted U
+ + + +/- + -

FlexLinks - + + + - +
1. Use of Distributed EtherChannel Greatly Reduces Chances of Black Holing Condition
2. NIC Teaming Can Eliminate Black Holing Condition
3. When Service Modules Are Used and Active Service Modules Are Aligned to Agg1
4. ACE Module Permits L2 Loopfree Access with per Context Switchover on Uplink failure
5. Applies to when using CSM or FWSM in active/standby arrangement

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 43
Access Layer
Design:
L3 Design Model

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 44
Access Layer Design
Defining Layer 3 Access

ƒ L3 access switches connect


to aggregation with a
dedicated subnet
DC Core DC Core
ƒ L3 routing is first performed in the
access switch itself
ƒ .1Q trunks between pairs of L3
access switches support L2 DC Aggregation
adjacency requirements (limited to
access switch pairs)
ƒ All uplinks are active up to
ECMP maximum of 8, no spanning DC Access
tree blocking occurs L3
L2
ƒ Convergence time is usually better
than Spanning Tree (Rapid PVST+ is
close)
ƒ Provides isolation/shelter for hosts
affected by broadcasts

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 45
Access Layer Design
Need L3 for Multicast Sources?

ƒ Multicast sources on L2
access works well when
IGMP snooping is available
DC Core DC Core
ƒ IGMP snooping at access
switch automatically limits
multicast flow to interfaces
with registered clients DC Aggregation
in VLAN
ƒ Use L3 when IGMP snooping
is not available or when
particular L3 administrative DC Access
functions are required L3
L2

L3 Access with L3 Access with


Multicast Sources Multicast Sources

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 46
Access Layer Design
Benefits of L3 Access
ƒ Minimises broadcast domains
attaining high level of stability
ƒ Meet server stability DC Core
DC Core
requirements or isolate
particular application
environments
ƒ Creates smaller failure DC Aggregation
domains, increasing stability
ƒ All uplinks are available
paths, no blocking (up to DC Access
ECMP maximum) L3
L2
ƒ Fast uplink convergence:
failover and fallback, no arp
table to rebuild for agg
switches

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 47
Access Layer Design
Drawbacks of Layer 3 Design
ƒ L2 adjacency is limited to
access pairs (clustering and
NIC teaming limited) DC Core DC Core
ƒ IP address space
management is more difficult
than L2 access DC Aggregation
ƒ If migrating to L3 access, IP
address changes may be
difficult on servers (may DC Access
break apps) L3
L2
ƒ Would normally require
services to be deployed at
each access layer pair to
maintain L2 adjacency with
server and provide stateful
failover
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 48
Access Layer Design
L2 or L3? What Are My Requirements?
Aggregation Aggregation
L3

OSPF, EIGRP
L2

Rapid PVST+ or MST


L3
Access L2 Access

Layer 2 Layer 3
The Choice of One Design Versus the Other One Has to Do With:
ƒ Difficulties in managing loops ƒ Ability to extend VLANs
ƒ Staff skillset; time to resolution ƒ Specific application requirements
ƒ Broadcast domain sizing
ƒ Convergence properties
ƒ Oversubscription requirements
ƒ NIC teaming; adjacency
ƒ Link utilisation on uplinks
ƒ HA clustering; adjacency ƒ Service module support/placement

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 49
Access Layer
Design:
BladeServers
and VMs

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 50
Blade Server Requirements
Connectivity Options

Using Pass-Through Modules Using Integrated Ethernet Switches

Aggregation
Layer

External L2
Switches

Integrated L2
Switches

Interface 1
Interface 2
Blade Server Chassis Blade Server Chassis

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 51
Blade Server Requirements
Connectivity Options
ƒ Layer 3 access tier may be used
to aggregate blade Aggregation
Layer
server uplinks L3
L2
–Permits using 10GE uplinks
into agg layer 10GE
ƒ Avoid dual tier Layer 3 Layer 2
Access Access
Layer 2 access designs
L3
STP blocking L2

Over-subscription
Integrated L2
Larger failure domain Switches

ƒ Consider “Trunk Failover”


feature of integrated switch

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 52
Blade Server Requirements
Trunk Failover Feature
ƒ Switch takes down server Aggregation
interfaces if corresponding uplink
fails, forcing NIC teaming failover
ƒ Solves NIC teaming limitations;
prevents black-holing of traffic
ƒ Achieves maximum bandwidth
utilisation:
No blocking by STP, but STP is
enabled for loop protection
Can distribute trunk failover groups
across switches Integrated L2
switches
ƒ Dependent upon the NIC feature
set for NIC Teaming/failover

Blade Server Chassis


Interface 1
Interface 2

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 53
Virtual Machine Principles &
Considerations

ƒ Virtual Machines enable a single


“real” server to host multiple
O/S+applications
1:1 Æ 1:n mapping
ƒ Each Virtual Machine has a
unique MAC + IP address
network identity Hypervisor
ƒ Rack-optimised servers, Blade
Servers and Virtual Machines are Port Address
driving very dense edge 2/1 aa.aa.aa
environments 2/1 bb.bb.bb
2/1 cc.cc.cc
2/1 dd.dd.dd
2/1 ee.ee.ee
2/1 11.11.11

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 54
Virtual Machine Considerations
ƒ Virtual Machines
multiply network
resource utilisation
ƒ Increased Control &
Data Plane demands
Virtual
Servers

Real Servers

Port Address
Port Address
ƒ Scalable Data & Control Port
Port
Port Address
Port 2/1 Address
Address
2/1Address
aa.aa.aa
aa.aa.aa
2/1 aa.aa.aa
Plane Attributes…Hardware Port 2/1 Address
Port 2/1 Address
2/1
2/1
2/1
aa.aa.aa
aa.aa.aa
aa.aa.aa
bb.bb.bb
bb.bb.bb
2/1 bb.bb.bb
MAC learning, Dedicated 2/1
2/1 2/1
2/1
2/1 aa.aa.aa
aa.aa.aa
2/1
2/1
bb.bb.bb
bb.bb.bb
bb.bb.bb
cc.cc.cc
cc.cc.cc
2/1
bb.bb.bb cc.cc.cc
Control buffers, Cache 2/1 2/1
2/1 2/1 bb.bb.bb
2/1
cc.cc.cc
dd.dd.dd
2/1cc.cc.cc
2/1 dd.dd.dd
cc.cc.cc
dd.dd.dd
2/1cc.cc.cc
2/1
based forwarding, Broadcast 2/1 2/1
2/1
2/1
2/1
2/1
dd.dd.dd
2/1cc.cc.cc
dd.dd.dd
dd.dd.dd
ee.ee.ee
ee.ee.ee
ee.ee.ee
2/1 2/1 dd.dd.dd
Suppression, Layer-2 trace, 2/1 2/1 dd.dd.dd
2/1
2/1
2/1
2/1
ee.ee.ee
ee.ee.ee
ee.ee.ee
11.11.11
11.11.11
11.11.11
2/1 2/1 ee.ee.ee
etc 2/1 2/1 ee.ee.ee
2/1
2/1
11.11.11
11.11.11
11.11.11
11.11.11
2/1 11.11.11
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 55
Density and
Scalability
Implications in the
Data Centre

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 56
Density and Scalability Implications
Modular or 1RU Access Layer Switching
Models
ƒ Where are the issues?
Cabling
Power
Cooling
Spanning Tree Scalability
Management
Oversubscription
Sparing
Redundancy
ƒ The right solution is usually
based on business
requirements
ƒ Hybrid implementations can
and do exist
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 57
57
Density and Scalability Implications
Server Farm Cabinet Layout
~30-40 1RU Servers per Rack

…… N- Racks per Row


…… Considerations:

How Many Interfaces per Server


Top of Rack Switch placement or
End of Row/Within the Row
Separate Switch for OOB Network
Cabling overhead vs under floor
Patch systems
Cooling capacity
Power distribution/ Redundancy
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 58
Density and Scalability Implications
Cabinet Design with 1RU Switching
Servers Connect Directly to a 1RU Switch

ƒ Minimises the number of cables to


run from each cabinet/rack
ƒ If NIC teaming support: two -1RU
switches are required
ƒ Will two 1RU switches provide
enough port density?

Cabling remains in cabinets


ƒ Cooling requirements usually do not
permit a full rack of servers
ƒ Redundant switch power supply are
option
ƒ Redundant switch CPU
considerations: not an option
ƒ GEC or 10GE Uplink Considerations

Single Rack-2 Switches Dual Rack-2 Switches

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 59
Density and Scalability Implications
Network Topology with 1RU Switching Model

Pro: Efficient Cabling DC Core 4 Uplinks per Cabinet = 100


Pro: Improved Cooling Uplinks Total to Agg Layer
Con: Number of Devices/Mgmt
Con: Spanning Tree Load

Aggregation
GEC or 10GE Uplinks?

Access

Cabinet 1 Cabinet 2
… Cabinet 25

With ~1,000 Servers/25 Cabinets = 50 Switches

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 60
Density and Scalability Implications
Cabinet Design with Modular Access Switches
Servers Connect Directly to a Modular Switch
ƒ Cable bulk at cabinet floor entry
can be difficult to manage and
block cool air flow
ƒ Typically spaced out by placement
at ends of row or within row
ƒ Minimises cabling to Aggregation
ƒ Reduces number of
uplinks/aggregation ports
ƒ Redundant switch power and CPU
are options
ƒ GEC or 10GE Uplink
Considerations
ƒ NEBS Considerations

Cables route under raised floor or in overhead trays

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 61
Density and Scalability Implications
Network Topology with Modular Switches in the
Access

Pro: Fewer Devices/Mgmt DC Core 2 Uplinks per 6509 = 16


Con: Cabling Challenges Uplinks to Agg Layer
Con: Cooling Challenges

Aggregation

GEC or 10GE Uplinks?

Access

With ~1,000 Servers/9 Slot Access Switches= 8 Switches


~8 Access Switches to Manage
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 62
Density and Scalability Implications
1RU and Modular Comparison

Lower STP Proc

More I/O Cables

Modular Fewer Uplinks


Access

Aggregation

1RU Higher STP Proc


Access

Fewer I/O Cables

More Uplinks

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 63
Density and Scalability Implications
Density: How Many NICs to Plan For?
Front End Interface
ƒ Three to four NICs per server
OOB
are common Management
Front end or public interface Backup Network
Storage HBA or
Storage interface (GE, FC) GE NIC
Backup interface
Back end or private interface
Back End Interface
integrated Lights Out (iLO)
for OOB mgmt

Cabling remains in cabinets


ƒ May require more than two
1RU switches per rack
30 servers@ 4 ports = 120 ports
required in a single cabinet (3x48 port
1RU switches)
May need hard limits on
cabling capacity
Avoid cross cabinet and other cabling
nightmares

Single Rack-2 Switches Dual Rack-2 Switches

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 64
Density and Scalability Implications
Hybrid Example with Separate OOB
Aggregation

OOB OOB OOB OOB

ƒ 1RU End of Row Approach: ƒ Separate OOB Switch:


–Smaller failure domains –Usually 10M ethernet (iLo)
–Low power consumption (kw/hr) –Very low utilisation
–GEC or 10GE uplinks –Doesn’t require high performance
–Permits server distribution via patch –Separate low end switch ok
–Eliminates TOR port density issue –Isolated from production network
–Requires solid patch & cabling systemƒ Hybrid Modular + 1RU
–Provides flexibility in design
–Dual CPU + Power for critical apps
–Secondary port for NIC Teaming

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 65
Density and Scalability Implications
Oversubscription and Uplinks
ƒ What is the oversubscription ratio
per uplink?
Develop an oversubscription
reference model
Identify by application, tier, or other means

ƒ Considerations
Future- true server capacity
(PCI-X, PCI- Express)
Server platform upgrade cycle will increase
levels of outbound traffic
Uplink choices available
Gigabit EtherChannel 10GE
10Gig EtherChannel
Flexibility in adjusting oversubscription ratio
Can I upgrade to 10GE easily? 10G
EtherChannel?
Upgrade CPU and switch fabric? (sup1-2-720-?)

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 66
Density and Scalability Implications
Spanning Tree

ƒ 1RU switching increase chances


of larger spanning tree diameter
ƒ BladeServer switches are
logically similar to adding 1RU
switches into the access layer
ƒ A higher number of trunks will
increase STP logical port counts
in aggregation layer
ƒ Determine spanning tree logical
and virtual interfaces before
extending VLANs or adding
trunks
ƒ Use aggregation modules to
scale STP and 10GE density

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 67
67
Scaling
Bandwidth
and Density

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 68
Scaling B/W with GEC and 10GE
Optimising EtherChannel Utilisation

ƒ Ideal is graph on top right


ƒ Bottom left graph more typical
ƒ Analyze the traffic flows in and
out of the server farm:
IP addresses (how many?)
L4 port numbers (randomised?)

ƒ Default L3 hash may not be


optimal for GEC: L4 hash
may improve

agg(config)# port-channel load balance


src-dst-port
ƒ 10 GigE gives you effectively the
full bandwidth without hash
implications

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 69
Scaling B/W with GEC and 10GE
Using EtherChannel Min-Links

ƒ Min-Links feature is available


as of 12.2.18SXF
ƒ Set the minimum number of
member ports that must be in
the link-up state or declare
the link down
ƒ Permits higher bandwidth
alternate paths to be used or
lower bandwidth paths to be
avoided
ƒ Locally significant Router# configure terminal
Router(config)# interface port-channel 1
configuration Router(config-if)# port-channel min-links 2
Router(config-if)# end

ƒ Only supported on LACP


EtherChannels

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 70
Scaling B/W with GEC and 10GE
Migrating Access Layer Uplinks to 10GE

ƒ How do I scale as I migrate from GEC to 10GE uplinks?


ƒ How do I increase the 10GE port density at the agg layer?
ƒ Is there a way to regain slots used by service modules?

DC Core

Aggregation

Access Pair 1 … …
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 71
Scaling B/W with GEC and 10GE
Service Layer Switch

ƒ Move certain services out of


aggregation layer DC Core
ƒ Ideal for CSM,
SSL modules
ƒ Opens slots in agg layer for
10GE ports Service Service
Switch1 Aggregation Switch2
ƒ May need QOS or separate (Redundant)
links for FT paths
ƒ Extend only necessary L2
VLANs to service switches
via .1Q trunks (GEC/TenG)
ƒ RHI installs route in local Access
MSFC only, requiring L3
peering with aggregation

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 72
Scaling B/W with GEC and 10GE
Consolidate to ACE

DC Core

ƒ Consider consolidating multiple


service modules onto ACE Module
–SLB
–Firewall
Aggregation
–SSL
ƒ 4/8/16G Fabric Connected
ƒ Active-Active Designs
ƒ Higher CPS + Concurrent CPS
ƒ Single TCP termination, lower
latency Access
ƒ Feature gap may not permit till later
release

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 73
Spanning Tree
Scalability

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 74
Spanning Tree Scalability
Common Questions

ƒ How many VLANs can I support in


a single aggregation module?
DC Core
ƒ Can a “VLAN Anywhere” model be
supported?
ƒ How many access switches can I
support in each aggregation Aggregation
module?
ƒ What is the recommended
oversubscription rate?
ƒ What are the maximum number of
logical ports?
ƒ Are there STP hardware
restrictions?

Access Pair 1 … …
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 75
Spanning Tree Scalability
Spanning Tree Protocols Used in the DC

ƒ Rapid PVST+ (802.1w)


Most common in data centre today
Scales to large size (~10,000 logical ports)
Coupled with UDLD, Loopguard, RootGuard and BPDU Guard,
provides a strong-stable L2 design solution
Easy to implement, proven, scales
ƒ MST (802.1s)
Permits very large scale STP implementations (~30,000 logical
ports)
Not as flexible as Rapid PVST+
Service module implications (FWSM transparent mode)
More common in service providers and ASPs
This Focuses on the Use of Rapid PVST+

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 76
Spanning Tree Scalability
Spanning Tree Protocol Scaling

MST RPVST+ PVST+

Total Active STP 50,000 Total


30,000 Total with 10,000 Total 13,000 Total
Logical Interfaces
Release 12.2(17b)SXA
1,800 2 per
Switching
Total Virtual Ports 6,0002 per Module(6700) 1,800 2 per
per LineCard Switching Module Switching Module
1200 for Earlier
Modules

1 CSCed33864 Is Resolved in Release 12.2(17d)SXB and Later Releases


2 10 Mbps, 10/100 Mbps, and 100 Mbps Switching Modules Support a Maximum of 1,200
Logical Interfaces per Module
http://www.cisco.com/univercd/cc/td/doc/product/lan/cat6000/122sx/ol_4164.htm#wp26366

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 77
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Number of Total STP Active
Logical Interfaces= DC Core
ƒ Trunks on the switch * active
VLANs on the trunks + number
of non-trunking interfaces on the switch 30 VLANs
In this example, aggregation 1 will have: Te7/1

10 + 20 + 30 = 60 STP active logical interfaces Te7/2

Te7/4
AGG1#sh spann summ tot
Switch is in rapid-pvst mode Te7/3
Root bridge for: VLAN0010, VLAN0020, VLAN0030
EtherChannel misconfig guard is enabled
Extended system ID is enabled
Portfast Default is disabled
PortFast BPDU Guard Default is disabled
Portfast BPDU Filter Default is disabled
Loopguard Default is enabled
UplinkFast is disabled 10 VLANs 20 VLANs
BackboneFast is disabled
Pathcost method used is long

Name Blocking Listening Learning Forwarding STP Active


---------------------- -------- --------- -------- ---------- ----------
30 VLANs 0 0 0 60 60
AGG1#

STP Active Column = STP Total Active Logical Interfaces

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 78
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Number of Virtual Ports per
Line Card= DC Core
ƒ For line card x: sum of all trunks * VLANs *
(the number of ports in a port-channel if used)
10 + 20 + (30*2) 30 VLANs
=90 Virtual Port’s on line card 7 Te7/1
Te7/2

Te7/4
AGG1#sh vlan virtual-port slot 7
Slot 7 Te7/3
Port Virtual-ports
-------------------------
Te7/1 30 EtherChannel
Te7/2 30 10 VLANs 20 VLANs
Te7/3 10
Te7/4 20
Total virtual ports:90
AGG1#
NOTE: VPs Are Calculated per Port in Channel Groups

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 79
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Example: Calculating Total Active
Logical Ports
ƒ 120 VLANs system wide
Core
ƒ No manual pruning performed on
trunks
ƒ 1RU access layer environment Layer 3
Layer 2
ƒ 45 access switches each Aggregation 1 Aggregation 2
Primary Root Loop Secondary Root
connected with 4GEC
ƒ Dual homed, loop topology
(120 * 45 access links)+120
instances on link to
agg2=5400+120=5520 Access 1 ..… Access 45
ƒ This is under the maximum
recommendation of 10,000 when
using Rapid PVST+

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 80
Spanning Tree Scalability
Spanning Tree Protocol Scaling
Example: Calculating Virtual Ports per
Line Card
ƒ 120 VLANs system wide
Core
ƒ no manual pruning performed on
trunks
ƒ 12 access switches, each Layer 3
connected with 4GEC across Layer 2

6700 line card Aggregation 1 Aggregation 2


Primary Root loop Secondary Root

6748 Line Card

Acc1Acc2 …. Acc12
Access 1 ..… Access 12
EtherChannels to Acces Layer
(120 * 48 access links) ~10 VLANs Used on Each Access Switch
=5,760 virtual ports Maximum Number of Supported
ƒ This is above the recommended VLANs in This Design Would Be
1800/48=37.5
watermark

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 81
Spanning Tree Scalability
Why STP Watermarks Are Important
Watermarks Are Not Hard Limits But─
ƒ If exceeded, performance is unpredictable
ƒ Larger impact when interface flaps, or shut/no_shut
ƒ Small networks may not see a problem
ƒ Large networks will usually see problems
ƒ Convergence time will be affected

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 82
Spanning Tree Scalability
Design Guidelines

ƒ Add aggregation modules to


scale, dividing up the STP DC Core
domain
ƒ Maximum five hundred HSRP
instances on Sup720
(depends on other cpu driven
processes) Aggregation
Module 1
ƒ If logical/virtual ports near
upper limits perform:
–Manual pruning on trunks
–Add agg modules
–Use MST if meets requirements

Access

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 83
Increasing HA in
the Data Centre

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 84
Increasing HA in the Data Centre
Server High Availability
Common Points of Failure
1. Server network adapter
2. Port on a multi-port
server adapter
L3
L2 3. Network media
(server access)
4. Network media (uplink)
5. Access switch port
6. Access switch module
7. Access switch

These Network Failure Issues Can Be


Without Data Centre HA With Data Centre HA Addressed by Deployment of Dual
Recommendations Recommendations Attached Servers Using Network Adapter
Teaming Software
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 85
Increasing HA in the Data Centre
Common NIC Teaming Configurations
AFT—Adapter Fault Tolerance SFT—Switch Fault Tolerance ALB—Adaptive Load Balancing

Default GW Default GW Default GW


10.2.1.1 10.2.1.1 10.2.1.1
HSRP HSRP HSRP

Heartbeats
Heartbeats
Heartbeats

Eth0: Active Eth1: Standby Eth0: Active Eth1: Standby Eth0: Active Eth1-X: Active

IP=10.2.1.14 IP=10.2.1.14
IP=10.2.1.14 IP=10.2.1.14
MAC =0007.e910.ce0f MAC =0007.e910.ce0e
MAC =0007.e910.ce0f MAC =0007.e910.ce0f
One Port Receives, All Ports Transmit
On Failover, Src MAC Eth1 = Src MAC Eth0 On Failover, Src MAC Eth1 = Src MAC Eth0
Incorporates Fault Tolerance
IP Address Eth1 = IP Address Eth0 IP Address Eth1 = IP Address Eth0
One IP Address and Multiple MAC Addresses

Note: NIC manufacturer drivers are changing


and may operate differently. Also, server OS
have started integrating NIC teaming drivers
which may operate differently.

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 86
Increasing HA in the Data Centre
Server Attachment: Multiple NICs
You Can Bundle Multiple Links to Allow
Generating Higher Throughputs Between
Servers and Clients
L3
L2
EtherChannel Hash May
Not Permit Full Utilisation
for Certain Applications
(Backup Example)

Only One Link Active:


Fault Tolerant Mode

All Links Active: EtherChannels


Load Balancing

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 87
Increasing HA in the Data Centre
Failover: What Is the Time to Beat?

ƒ The overall failover time is the


combination of convergence
at L2, L3, + L4 components
Stateful devices can replicate connection
information and typically failover within 3-5sec
EtherChannels < 1sec
STP converges in ~1 sec (802.1w)
HSRP can be tuned to <1s

ƒ Where does TCP break? Microsoft,


Linux, AIX, etc.
Linux and
Microsoft XP Others Tolerate
2003 Server a Longer
Failover Time

TCP Stack Outage

L4 Convergence Tolerance
~ 5s ~ 9s
L2 Convergence L3 Convergence

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 88
Increasing HA in the Data Centre
Failover Time Comparison

ƒ STP-802.1w—One sec
ƒ OSPF-EIGRP—One sec
ƒ ACE Module with Autostate
ƒ HSRP—Three sec (using 1/3)
ƒ FWSM Module—Three sec
ƒ CSM Module—Five sec
ƒ WinXP/2003 ServerTCP Stack—Nine sec
Failover Time

TCP Stack
Tolerance
Content ~ 9s
Service
Module
HSRP FireWall ~ 5s
~ 3s Service
(may be tuned Module
Spanning Tree ACE to less) ~ 3s
OSPF/EIGRP
Sub-second ~1sec ~ 1s

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 89
Increasing HA in the Data Centre
Non-Stop Forwarding/Stateful Switch-Over

ƒ NSF/SSO is a supervisor NSF-Aware NSF-Aware


redundancy mechanism for intra-
chassis supervisor failover
ƒ SSO synchronises layer 2 protocol
state, hardware L2/L3 tables
(MAC, FIB, adjacency table), ACL
and QoS tables
ƒ SSO synchronises state for:
trunks, interfaces, EtherChannels,
port security, SPAN/RSPAN, STP,
UDLD, VTP
ƒ NSF with EIGRP, OSPF, IS-IS,
BGP makes it possible to have no
route flapping during the recovery
ƒ Aggressive RP timers may not
work in NSF/SSO environment

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 90
Increasing HA in the Data Centre
NSF/SSO in the Data Centre
ƒ SSO in the access layer:
Improves availability for single
attached servers

ƒ SSO in the aggregation layer:


Consider in primary agg layer switch
Avoids rebuild of arp, igp, stp tables
Prevents service module switchover
(can be up to ~6sec depending on which
module)
SSO switchover time less than two sec
12.2.18SXD3 or higher

ƒ Possible implications
HSRP state between Agg switches is
not tracked and will show switchover until control
plane recovers
IGP Timers cannot be aggressive (tradeoff)

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 91
Increasing HA in the Data Centre
Best Practices: STP, HSRP, Other
Rapid PVST+
UDLD Global
Spanning Tree Pathcost Method=Long

LACP+L4 Port Hash


L3+ L4 CEF Hash Dist EtherChannel for FT
and Data VLANs
Agg1: FT
Agg2:
STP Primary Root STP Secondary Root
HSRP Primary Data
HSRP Secondary
HSRP Preempt and Delay HSRP Preempt and Delay
Dual Sup with NSF+SSO Single Sup
LACP+L4 Hash
Dist EtherChannel
Min-Links

Rootguard
LoopGuard Blade Chassis
Portfast + with Integrated
Switch
BPDUguard
Rapid PVST+: Maximum Number of STP Active Logical
Ports- 8000 and Virtual Ports Per Linecard-1500

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 92
A look at the
Future…

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 93
Traffic Differentiation in Ethernet
Networks

ƒ IEEE work in progress


ƒ Possible future consolidated I/O technology

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 94
Traffic Types and Requirements

ƒ Enable Ethernet to carry LAN, SAN and IPC traffic :


I/O consolidation
Eliminates multiple backplanes (eg Blade Servers)
Should support appropriate characteristics for each traffic type

ƒ LAN
Large number of flows, not very sensitive to latency
E.g. dominant traffic type in Front End Servers

ƒ SAN
Large packet sizes, sensitive to packet drops

ƒ IPC
Mix of large & small messages, small messages latency sensitive
E.g. HPC Applications

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 95
Challenges in Traffic Differentiation

ƒ Link Sharing (Transmit)


Different traffic types may share same queues/links
Large burst from one traffic should not affect other traffic types
ƒ Resource Sharing
Different traffic types may share same resources (e.g. buffers)
Large queued traffic for one traffic type should not starve other
traffic types out of resources
ƒ Receive Handling
Different traffic types may need different Receive handling (e.g.
interrupt moderation)
Optimisation for CPU utilisation for one traffic type should not
create large latency for small message for other traffic types

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 96
IEEE Higher Speed Study Group

ƒ Objectives
Support full-duplex operation only
Preserve the 802.3/Ethernet frame format at the MAC Client
service interface
Preserve minimum and maximum FrameSize of current
802.3 Std
Support a speed of 100 Gb/s at the MAC/PLS service
interface
Support at least 10km on SMF.
Support at least 100 meters on OM3 MMF

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 97
Meet the Experts
Data Centre

ƒ Victor Moreno
Technical Leader

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 98
Recommended Reading
BRKDCT -2001

ƒ Build the Best Data


Center Facility for
Your Business
ƒ Data Center
Fundamentals

Available in the Cisco Company Store


BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 99
Q and A

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 100
Complete Your Online Session
Evaluation

BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 101
BRKDCT-2001 © 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential 102

S-ar putea să vă placă și