Documente Academic
Documente Profesional
Documente Cultură
Table of contents
Executive Summary .............................................................................................................................. 2 Virtual Connect FlexFabric Module Hardware Overview .......................................................................... 3 Designing an HP FlexFabric Architecture for VMware vSphere .................................................................. 4 Designing a vSphere Network Architecture with the Virtual Connect FlexFabric module ............................... 7 vNetwork Distributed Switch Design ................................................................................................... 7 Network I/O Control ........................................................................................................................ 9 Hypervisor Load Balancing Algorithms ............................................................................................... 9 Modifying Server Profiles with Online Servers .................................................................................... 10 HP NC551 and NC553 Minimum Supported Firmware and Drivers .................................................... 10 Appendix A: Virtual Connect Bill of Materials ....................................................................................... 11 Appendix B: Terminology cross-reference ............................................................................................. 11 Appendix C: Glossary of Terms ........................................................................................................... 12 For more information .......................................................................................................................... 14
Executive Summary
HP has revolutionized the way IT thinks about networking and server management. With the release of the HP ProLiant BladeSystem Generation 6 servers, along with Virtual Connect Flex-10 Ethernet modules, HP provided a great platform for VMware vSphere. Virtual Connect Flex-10 is the world's first technology to divide and fine-tune 10Gb Ethernet network bandwidth at the server edge. When combined with Virtual Connect, the BladeSystem architecture streamlines the typical change processes for provisioning in the datacenter. HP has since evolved Virtual Connect Flex-10 to the next level: Virtual Connect FlexFabric modules. By combining the power of ProLiant BladeSystem Generation 7 servers and Virtual Connect, the Virtual Connect FlexFabric module allows customers to consolidate network connections and storage fabrics into a single module. This further reduces infrastructure cost and complexity by eliminating HBA adapters and Fibre Channel modules at the edge. These servers include virtualization friendly features such as larger memory capacity, dense population, room for additional mezzanine cards and 4 - 64 (with Intel Hyper-Threading technology enabled and AMD Opteron 6200-series) processing cores. The following ProLiant BL Servers ship standard with the NC551i FlexFabric Adapter: BL465 G7 BL685 G7
The following ProLiant BL Servers ship with the NC553i FlexFabric Adapter: BL460 G7 BL490 G7 BL620/680 G7
Additionally, the NC551m and NC553m provide support for the FlexFabric Adapter in ProLiant BladeSystem G6 servers. NOTE: Please check the latest NC551m1 and NC553m2 QuickSpecs for the official server support matrix. The ProLiant Generation 8 servers do not ship with a built-in adapter. Instead, HP offers a FlexLOM option to where customers can chose which FlexFabric Adapter vendor they want. Please review the server models QuickSpecs for support FlexLOM adapters. The FlexFabric Adapter introduces a new Physical Function called the FlexHBA. The FlexHBA, along with the FlexNIC, adds the unique ability to fine-tune each connection to adapt to your virtual server channels and workloads on-the-fly. The effect of using the Virtual Connect FlexFabric module is a reduction in the number of interconnect modules required to uplink outside of the enclosure, while still maintaining full redundancy across the service console, VMkernel, virtual machine (VM) networks and
1 2
http://h18004.www1.hp.com/products/blades/components/emulex/nc551m/index.html http://h18004.www1.hp.com/products/blades/components/emulex/nc553m/index.html
storage fabrics. This translates to a lower cost infrastructure with fewer management points, switch modules, adapter cards and cables. This whitepaper will focus on designing a highly available network and vSphere Cluster architecture by providing redundant uplinks across physical modules, but also across enclosures. This design will also maximize the management features of Virtual Connect, while providing customers with the flexibility to provide any networking to any host within the Virtual Connect domain. Simply put, this design will not over-provision servers, while keeping the number of uplinks used to a minimum. This helps reduce infrastructure cost and complexity by trunking the necessary VLANs (IP Subnets) to the Virtual Connect domain, and minimizing potentially expensive 10Gb uplink ports. Finally, this document will provide key design best practices for vSphere 5 network architecture with HP FlexFabric, including: vDS design for Hypervisor networking vSwitch and dvPortGroup load balance algorithms
Ports X7 and X8 are shared with internal Stacking Link ports. If the external port is populated with a transceiver, the internal Stacking Link is disabled. At least one Stacking Link is required between modules. Please refer to the Virtual Connect User Guide or Multi-Enclosure Stacking Reference Guide3 for more information.
http://bizsupport2.austin.hp.com/bc/docs/support/SupportManual/c02102153/c02102153.pdf
Important: Even though the Virtual Connect FlexFabric module supports Stacking, stacking only applies to Ethernet traffic. FC uplinks cannot be consolidated, as it is not possible to stack the FC ports, nor provide a multi-hop DCB bridging fabric today.
Shared Uplink Sets provide administrators the ability to distribute VLANs into discrete and defined Ethernet Networks (vNet.) These vNets can then be mapped logically to a Server Profile Network Connection allowing only the required VLANs to be associated with the specific server NIC port. This also allows customers the flexibility to have various network connections for different physical Operating System instances (i.e. VMware ESX host and physical Windows host.) As of Virtual Connect Firmware 3.305 release, the following Shared Uplink Set rules apply per domain: 1000 Unique VLANs (Ethernet Networks) per Virtual Connect Domain 162 Unique Server Mapped VLANs per Server Profile Network Connection
Important: When using FlexNICs, the 162 Unique Server Mapped VLAN maximum is an aggregate of all FlexNICs that belong to the same 10Gb Physical Interface. A FlexHBA also counts as a single VNET against the 162 total. By providing two stacked Enclosures, this will allow for not only Virtual Connect FlexFabric module failure, but also Enclosure failure. The uplink ports assigned to each Shared Uplink Set (SUS) were vertically offset to allow for horizontal redundancy purposes, as shown in Figure 1-2. IP-based storage (NFS and/or iSCSI) can be dedicated and segregated by a separate vNet and assigned uplink port. This design approach provides administrators to dedicate a network (physically switched, directly connected or logical within a Shared Uplink Set) to provide access to IP-based storage arrays.
Only available with Virtual Connect Manager Firmware 2.10 or greater. Please review the Virtual Connect Manager Release Notes for more information regarding domain stacking requirements: http://h18004.www1.hp.com/products/blades/components/c-class-tech-installing.html Only when the Expanded VLAN Capacity feature is enabled in the Domain (Ethernet -> Advanced Settings)
Directly connecting an IP-based Storage array has certain limitations: Each storage array front-end port will require a unique vNet Each defined vNet will require separate server network connections You are limited to the number of IP-based arrays based on the number of unassigned uplink ports
Note: Please review the Virtual Connect with iSCSI Cookbook Third Edition or newer for more information on how to directly attach an IP-based storage device to Virtual Connect. Virtual Connect has the capability to create an internal, private network without uplink ports, by using the low latency mid-plane connections to facilitate communication. This vNet can be used for cluster heartbeat networks, or in this case VMotion and/or Fault Tolerance traffic. Traffic will not pass to the upstream switch infrastructure, which will eliminate the bandwidth otherwise consumed.
Figure 1-1: Physical VMware vSphere Cluster Design
Figure 1-2 show the physical cabling. The X5 and X6 Ethernet ports of the FlexFabric module are connecting to a redundant pair of Top of Rack (ToR) switches, using LACP (802.3AD) for link redundancy. The ToR switches can be placed End of Row to save on infrastructure cost. Ports X7 are used for vertical External Stacking Links, while X8 are used for Internal Stacking Links.
http://h20000.www2.hp.com/bc/docs/support/SupportManual/c02533991/c02533991.pdf
As noted in the previous section, Virtual Connect FlexFabric Stacking Links will only carry Ethernet traffic, and do not provide any Fibre Channel stacking options. Thus, ports X1 and X2 from each module are populated with 8Gb SFP+ transceivers, providing 16Gb net FC bandwidth for storage access. Ports X3 and X4 are available to provide additional bandwidth if FC storage traffic is necessary. If additional Ethernet bandwidth is necessary, ports Enc0:Bay2:X5, Enc0:Bay2:X6, Enc1:Bay1:X5, and Enc1:Bay1:X6 can be used for additional Ethernet Networks or Shared Uplink Sets.
Figure 1-2: Physical cabling design
Designing a vSphere Network Architecture with the Virtual Connect FlexFabric module
The vNetwork Distributed Switch7 (vDS) allows an administrator to create a centralized distributed vSwitch, which share a common networking configuration across hosts. Port Groups are still utilized in this model, but have a different association to host uplink ports. Host uplink ports are added to Uplink Groups (dvUplinkGroup), where a logical association between the dvUplinkGroup and a PortGroup (dvPortGroup) is formed. vDS can service any of the vmkernel functions; Service Console, VMotion, IP Storage, and Virtual Machine traffic. In this section, we will outline the overall vDS design.
VMkernel Function
vDS Name
dvPortGroup Name
dvPortGroup1_Mgmt dvPortGroup2_vmkernel dvPortGroup3_vmnet100
Requires vSphere 5.0 Enterprise Plus licensing; Please refer to the vSphere 5 Maximums document on supported configuration maximums http://www.vmware.com/pdf/vsphere5/r50/vsphere-50-configuration-maximums.pdf
VMkernel Function
VM NetworkingN
vDS Name
dvs3_vmnet
dvPortGroup Name
dvPortGroupN_vmnetNNN
Important: Please know that for HP CloudSystem Matrix installations, the dvPortGroup for the Service Console must be named Management Network. Mixing vSphere Standard Switches and vSphere Distributed Switches is possible. The current scalability limits to the vDS may require some designs to incorporate both. HP would recommend vmkernel functions (vMotion, Management, IP Storage, FT) be assigned to Standard vSwitches, and use vDS for Virtual Machine networking. VMware Fault Tolerance (FT) could introduce more complexity in to the overall design. VMware states that a single 1Gb NIC should be dedicated for FT logging, which would have the potential to starve any shared pNIC with that of another vmkernel function (i.e. VMotion traffic.) FT has not been taken into consideration within this document. Even though FT could be shared with another vmkernel
function, and if FT is a design requirement, then the overall impact of its inclusion should be examined. With the design example given, there are three options one could choose to incorporate FT Logging:
Table 2-2 VMware Fault Tolerance Options
FT Design Choice
Share with VMotion network
Justification
The design choice to keep VMotion traffic internally to the Domain allows the use of low latency links for inter-Domain communication. NetIOC could be used to control and provide VMotion or FT traffic higher shares, or priority, over the other traffic.
Rating
***
Non-redundant VMotion Dedicate one pNIC for VMotion traffic, and the other for FT and FT networks logging traffic. Neither network will provide pNIC redundancy. Add additional FlexFabric Adapters and Modules This option increases the overall CapEx to the solution, but will provide more bandwidth options.
** *
NetIOC can be used to control identified traffic, when multiple types of traffic are sharing the same pNIC. In our design example above, FT Logging could share the same vDS as the vmkernel, and NetIOC would be used to control the two types of traffic. NetIOC can also be used to apply an 802.1p tag for DataCenter QOS, which is not supported by Virtual Connect at this time.
Name
Originating Virtual Port ID Source MAC Address IP Hash
Algorithm
Choose an uplink based on the virtual port where the traffic entered the virtual switch. MAC Address seen on vnic port Hash of Source and Destination IPs. Requires switch assisted load balancing, 802.3ad. Virtual Connect does not support 802.3ad on server downlink ports, as 802.3ad is a Point-to-Point bonding protocol. Introduced in vSphere 4.1 and only available with a vDS, Load-Based Teaming policy monitors the flow when the mean send or receive utilization on a dvUplink exceeds 75% capacity over 30-sec intervals. Highest order uplink from the list of Active pNICs.
Works with VC
Yes Yes No
Yes
Explicit Failover
Yes
Component
Firmware Driver (Ethernet) Driver (FC/FCoE)
Version
4.0.360.15 4.0.355.1 8.2.2.105.36
When multiple NC55x devices are used on a single host, you may experience an issue where not all of the FlexNICs will appear within the ESXi kernel when using the 4.0.3551. To resolve this, you must adjust the be2net driver heap stack allocation. Please refer to the HP Customer Advisory c032201409 on the process and any future updates.
http://h20000.www2.hp.com/bizsupport/TechSupport/Document.jsp?objectID=c03220140
10
Partnumber
571956-B21 AJ716A 487649-B21 455883-B21 Or 487655-B21
Description
HP Virtual Connect FlexFabric Module HP StorageWorks 8Gb B-series SW SFP+ .5m 10Gb SFP+ DAC Stacking Cable 10Gb SR SFP+ transceiver
Qty
4 8 2 4
Customer term
Port Bonding or Virtual Port VLAN Tagging
Industry term
IEEE term
Cisco term
Etherchannel or channeling (PaGP) Trunking
Nortel term
MultiLink Trunking (MLT) 802.1Q
11
EtherChannel*
FlexNIC**
IEEE 802.3ad
LACP LOM Maximum Link Connection Speed** Multiple Networks Link Speed Settings** MZ1 or MEZZ1; LOM
12
A software that runs on a host, allowing multiple network interface ports to be combined to act as a single virtual port. The software provides greater bandwidth, load balancing, and redundancy. Physical NIC port. A FlexNIC is seen by VMware as a pNIC Combining ports to provide one or more of the following benefits: greater bandwidth, load balancing, and redundancy. A Cisco proprietary protocol aids in the automatic creation of Fast EtherChannel links. PAgP packets are sent between Fast EtherChannel-capable ports to negotiate the forming of a channel. A term typically used in the Unix/Linux world that is synonymous to NIC teaming in the Windows world. Preferred FlexNIC speed value assigned by a vNet by the network administrator. 802.1Q VLAN tagging Combining ports to provide one or more of the following benefits: greater bandwidth, load balancing, and redundancy. See also Port Aggregation. A virtual network within a physical network. Tagging/marking an Ethernet frame with an identity number representing a virtual network. A Cisco proprietary protocol used for configuring and administering VLANs on Cisco network devices. Virtual NIC port. A software-based NIC used by VMs
**The feature was added for Virtual Connect Flex-10
Port Bonding Preferred Link Connection Speed** Trunking (Cisco) Trunking (Industry)
*The feature is not supported by Virtual Connect. ***The feature was added for Virtual Connect FlexFabric modules
13
Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. The only warranties for HP products and services are set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. HP shall not be liable for technical or editorial errors or omissions contained herein. Trademark acknowledgments, if needed. c03278211, March 2012