Documente Academic
Documente Profesional
Documente Cultură
Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)
Introduction
The broad adoption of virtualization has led to a flurry of
server consolidation projects. IT administrators are looking
to push the envelope when it comes to how many virtual
servers or Virtual Machines (VMs as they are commonly
referred to) can be packed on a single physical server. This
is a disruptive change and impacts traditional network
architectures and best practices in many ways. One direct
consequence of higher server virtualization ratios is that
as more VMs are packed on a single server, the bandwidth
demands from the server edge, all the way to the core of
the network, are growing at a rapid pace. Additionally with
more virtual machines on a single server, the redundancy
and resiliency requirements from the server edge to the
core of the network are increasing.
Traditionally, the approach to increasing bandwidth from
the server to the network edge has been to add more Network Interface Cards (NICs) and use Link Aggregation (LAG)
or NIC teaming as it is commonly called to bond links to
achieve higher bandwidth. If any of the links in the group
of aggregated links fails, the traffic load is redistributed
among the remaining links. Link aggregation provides a
simpler and easier way to both increase bandwidth and add
resiliency. Link aggregation is also commonly used between
two switches to increase bandwidth and resiliency. However, in both cases, link aggregation works only between
two individual devices, for example switch to switch, or
server to switch. If any one of the devices on either end of
the link aggregated group (or trunk as it is also called) fails,
then there is complete loss of connectivity.
In order to add device level redundancy various other
mechanisms have been deployed. Where Layer 3 routing
and segmentation is deployed in the network, various
router redundancy protocols such as VRRP, in conjunction
with interior gateway protocols such as OSPF, provide
adequate resiliency, failover and redundancy in the network. However, with virtualization driving the need for flatter Layer 2 topologies (since virtual machine movement
today is typically restricted to within a subnet boundary),
the drive towards a broader flatter Layer 2 data center network is gaining momentum. In this environment, protocols
such as the spanning tree protocol have typically provided
redundancy around both link and device failures. Spanning
tree protocol works by blocking ports on redundant paths
so that all nodes in the network are reachable through a
single path. If a device or a link failure occurs, based on
the spanning tree algorithm, a selective redundant path
or paths are opened up to allow traffic to flow, while still
reducing the topology to a tree structure which prevents
loops. Spanning tree protocol can be used in combination
with link aggregation where links between two nodes
such as switch to switch connections can be aggregated
using link aggregation to increase bandwidth and resiliency
between nodes or devices. Spanning tree would typically
treat the aggregated link as a single logical port in its calculations to come up with a loop free topology.
While spanning tree has served for many years as the de
facto network redundancy protocol, the changing requirements of data center networks today are forcing a
re-examination of the choice of redundancy mechanisms.
For example, one of the drawbacks of spanning tree protocol is that in blocking redundant ports and paths, spanning tree effectively reduces the available bandwidth
significantly, i.e. the bandwidth available on the redundant paths goes unused until a failure occurs. Additionally, in many situations the choice of which ports to block
can also lead to a suboptimal path of communication
between end nodes by forcing traffic to go up and down
the spanning tree. See Figure 1 below. Finally, the time
taken to recompute the spanning tree and propagate the
changes in the event of a failure can vary as well.
Traffic Path
LAG
LAG
LAG
LAG
STP Block
LAG
LAG
STP Block
5563-01
Figure 1
2
2011 Extreme Networks, Inc. All rights reserved.
Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)
Device 2
Device 3
ISL
M-LAG
(Device 2 & Device 3)
Device 2
Device 3
LAG
M-LAG
(Device 2)
LAG (Device 1)
LAG (Device 1)
Device 1
Device 1
5565-01
5564-01
Figure 2
Figure 3
3
2011 Extreme Networks, Inc. All rights reserved.
Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)
M-LAG
Blade Chassis
Blade
Server
LAG
VM
Pass
Through
Module
Pass
Through
Module
VM
5566-01
Figure 4
4
2011 Extreme Networks, Inc. All rights reserved.
Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)
Corporate
and North America
Extreme Networks, Inc.
3585 Monroe Street
Santa Clara, CA 95051 USA
Phone +1 408 579 2800
Asia Pacific
Phone +65 6836 5437
Japan
Phone +81 3 5842 4011
extremenetworks.com
2011 Extreme Networks, Inc. All rights reserved. Extreme Networks, the Extreme Networks logo and Direct Attach are either registered trademarks or trademarks of Extreme Networks, Inc. in the United States and/or other countries. Specifications are subject to change without notice. 1750_02 07/11