WPDCArchitectureswM LAG 1750

Embed Size (px)

Citation preview

  • 8/18/2019 WPDCArchitectureswM LAG 1750

    1/5

    Extreme Networks White Paper

    © 2011 Extreme Networks, Inc. All rights reserved. Do not reproduce.

    Make Your Network Mobile

    Abstract

    The broad adoption of virtualization has led to a urry

    of server consolidation projects. IT administrators arelooking to push the envelope when it comes to how

    many virtual servers or Virtual Machines (VMs) can be

    packed on a single physical server. This is a disruptive

    change and impacts traditional network architectures

    and best practices in many ways. This white paper exam-

    ines the challenges and the different architectural

    approaches to meet bandwidth, redundancy and

    resiliency requirements from the server edge to the

    core of the network in a virtualized environment.

    Exploring New DataCenter NetworkArchitectures withMulti-Switch LinkAggregation (M-LAG)

  • 8/18/2019 WPDCArchitectureswM LAG 1750

    2/5

    2

    Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)

    © 2011 Extreme Networks, Inc. All rights reserved.

    Introduction

    The broad adoption of virtualization has led to a urry ofserver consolidation projects. IT administrators are lookingto push the envelope when it comes to how many virtual

    servers or Virtual Machines (VMs – as they are commonlyreferred to) can be packed on a single physical server. Thisis a disruptive change and impacts traditional networkarchitectures and best practices in many ways. One directconsequence of higher server virtualization ratios is thatas more VMs are packed on a single server, the bandwidthdemands from the server edge, all the way to the core ofthe network, are growing at a rapid pace. Additionally withmore virtual machines on a single server, the redundancyand resiliency requirements from the server edge to thecore of the network are increasing.

    Traditionally, the approach to increasing bandwidth fromthe server to the network edge has been to add more Net-work Interface Cards (NICs) and use Link Aggregation (LAG)or “NIC teaming” as it is commonly called to bond links toachieve higher bandwidth. If any of the links in the groupof aggregated links fails, the traffic load is redistributedamong the remaining links. Link aggregation provides asimpler and easier way to both increase bandwidth and addresiliency. Link aggregation is also commonly used betweentwo switches to increase bandwidth and resiliency. How-ever, in both cases, link aggregation works only between

    two individual devices, for example switch to switch, orserver to switch. If any one of the devices on either end ofthe link aggregated group (or trunk as it is also called) fails,then there is complete loss of connectivity.

    In order to add device level redundancy various othermechanisms have been deployed. Where Layer 3 routingand segmentation is deployed in the network, variousrouter redundancy protocols such as VRRP, in conjunctionwith interior gateway protocols such as OSPF, provide

    adequate resiliency, failover and redundancy in the net-work. However, with virtualization driving the need for “at-ter” Layer 2 topologies (since virtual machine movementtoday is typically restricted to within a subnet boundary),the drive towards a broader atter Layer 2 data center net-work is gaining momentum. In this environment, protocolssuch as the spanning tree protocol have typically providedredundancy around both link and device failures. Spanningtree protocol works by blocking ports on redundant pathsso that all nodes in the network are reachable through asingle path. If a device or a link failure occurs, based onthe spanning tree algorithm, a selective redundant pathor paths are opened up to allow traffic to ow, while stillreducing the topology to a tree structure which preventsloops. Spanning tree protocol can be used in combinationwith link aggregation where links between two nodes –such as switch to switch connections – can be aggregated

    using link aggregation to increase bandwidth and resiliencybetween nodes or devices. Spanning tree would typicallytreat the aggregated link as a single logical port in its calcu-lations to come up with a loop free topology.

    While spanning tree has served for many years as the defacto network redundancy protocol, the changing require-ments of data center networks today are forcing are-examination of the choice of redundancy mechanisms.For example, one of the drawbacks of spanning tree pro-tocol is that in blocking redundant ports and paths, span-ning tree effectively reduces the available bandwidth

    signicantly, i.e. the bandwidth available on the redun-dant paths goes unused until a failure occurs. Addition-ally, in many situations the choice of which ports to blockcan also lead to a suboptimal path of communicationbetween end nodes by forcing traffic to go up and downthe spanning tree. See Figure 1 below. Finally, the timetaken to recompute the spanning tree and propagate thechanges in the event of a failure can vary as well.

    Traffic Path

    X

    LAGLAGLAG

    LAGLAG

    LAG

    STP BlockX

    STP Block

    Figure 1

  • 8/18/2019 WPDCArchitectureswM LAG 1750

    3/5

    3

    Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)

    © 2011 Extreme Networks, Inc. All rights reserved.

    Multi-Switch Link Aggregation(M-LAG)

    A number of new protocols and approaches have beensuggested to address some of the shortcomings ofspanning tree protocol. One approach to addressingboth the performance as well as the resiliency require-ments of these highly virtualized data centers is toextend the link-level redundancy capabilities of linkaggregation and add support for device-level redundan-cy. This can be accomplished by allowing one end of thelink aggregated port group to be dual-homed into twodifferent devices to provide device-level redundancy. Theother end of the group is still single homed into a singledevice. See Figure 2 below.

    In Figure 2, Device 1 treats the link aggregated ports asa normal link aggregated trunk group, i.e. it does notsee anything different. Traffic from Device 1 is distrib-uted across the ports in the group using traditional linkaggregation algorithms which would typically hash thetraffic across the ports in the group using a variety ofhashing algorithms. If one of the links in the group wereto go down, traffic would get redistributed across theremaining ports in the group. However, the other endof link aggregated group is where things now functiondifferently. Device 2 and Device 3 now work togetherto create the perception of a common link aggregated

    group so that Device 1 doesn’t see anything differentfrom a link aggregation perspective, even though the

    link aggregated ports are now distributed across Device2 and Device 3, thereby leading to the term Multi-SwitchLink Aggregation (M-LAG). Device 2 and Device 3 com-municate information to each other over the Inter SwitchLink (ISL) so that forwarding, learning and bridging workconsistently without causing any loops. The ISL itself canbe a regular LAG. If either the link to Device 1 from Device2 or Device 3 were to go down, or if Device 2 or Device 3itself went down, traffic would now get forwarded acrossthe remaining link/device thus providing both link-leveland device-level redundancy. The intelligence that allowsthe ports on Device 2 and Device 3 to present itself as asingle link aggregated trunk group to Device 1 today isimplemented using mostly proprietary mechanisms, i.e.M-LAG technology is still largely proprietary. However,the proprietary nature of the technology is conned tothe layer which presents itself as a distributed link

    aggregated group, specically Device 2 and Device 3in the gure below, both of which should come fromthe same vendor. Device 1 does not participate in thisproprietary protocol, and in fact Device 1 may come froma different vendor and can in fact be a different type ofdevice. For example, Device 1 can be a server which has dualNICs teamed together, while Device 2 and Device 3 can beEthernet switches from a single vendor. M-LAG can work atdifferent layers from the access to the core of the network.M-LAG can be used in conjunction with traditional linkaggregation to increase bandwidth as well as add link-levelredundancy between devices. See Figure 3 below.

    Device 1

    Device 2 Device 3

    LAG (Device 1)

    ISL

    M-LAG(Device 2 & Device 3)

    Figure 2

    Device 1

    Device 2 Device 3

    LAG (Device 1)

    LAG(Device 2) M-LAG(Device 2 & Device 3)

    Figure 3

  • 8/18/2019 WPDCArchitectureswM LAG 1750

    4/5

    4

    Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)

    © 2011 Extreme Networks, Inc. All rights reserved.

    Combining M-LAG with theDirect Attach Architecture(M-LAG Direct Attach)

    M-LAG serves as a powerful mechanism and tool toaddress newer architectural requirements centered onbandwidth and resiliency. Since the proprietarynature of M-LAG is limited to only the switches providingthe distributed link aggregation capabilities, it can becombined with other technologies, devices and vendorequipment to build better network architectures. Forexample, M-LAG can be used in conjunction with theDirect Attach™ architecture from Extreme Networks® toeliminate tiers and simplify switching in the data center. The Direct Attach™ architecture allows virtual machinesto be directly switched in the aggregation or the core

    of the physical network, thereby eliminating multipleswitching tiers such as the virtual switch, as well as theblade switch and the access switch. M-LAG allows dualhoming links from the server into the network and usingboth links in an active-active manner. By combining thetwo, virtual machines on a single server can be dual-homeddirectly into the aggregation or core of the network, while

    using both links in an active-active manner. Creating anM-LAG Direct Attach architecture not only helps to elimi-nate multiple tiers of switches thereby eliminating multiplepoints of oversubscription, latency and power, but it alsoadds link, device and network level resiliency to the datacenter fabric. And it does this without blocking ports or linksthereby allowing full utilization of the capacity built into thedata center fabric. In effect, an M-LAG Direct Attach architec-ture provides a very scalable, high-performance, low latencynetwork fabric for highly virtualized data centers. See Figure4 below.

    For more information on Extreme Networks Direct Attacharchitecture and how it works to reduce tiers, read thewhite paper at:www.extremenetworks.com/go/DirectAttach .

    A key benet of an M-LAG and M-LAG/Direct Attach

    approach is that it can be deployed on existing data cen-ter switches using a simple software upgrade, i.e. it doesnot require an infrastructure refresh. While M-LAG itself isproprietary, it works in conjunction with standard provenlink aggregation technology commonly available acrossserver and switch vendors, as well as in conjunction withdifferent hypervisor technologies in Direct Attach mode.

    Blade Chassis

    M-LAG

    LAG

    96-por t10/100/1000BASE-T,MPU21 96-port10 /100/1000BASE-T,MPU21

    BladeServer

    PassThroughModule

    PassThroughModule

    VMVM

    Figure 4

  • 8/18/2019 WPDCArchitectureswM LAG 1750

    5/5

    Make Your Network Mobile

    Extreme Networks White Paper: Exploring New Data Center Network Architectures with Multi-Switch Link Aggregation (M-LAG)

    © 2011 Extreme Networks, Inc. All rights reserved. Extreme Networks, the Extreme Networks logo and Direct Attach are either registered trademarks or trade-marks of Extreme Networks, Inc. in the United States and/or other countries. Specications are subject to change without notice. 1750_02 07/11

    Corporateand North AmericaExtreme Networks, Inc.3585 Monroe StreetSanta Clara, CA 95051 USAPhone +1 408 579 2800

    Europe, Middle East, Africaand South AmericaPhone +31 30 800 5100

    Asia Paci cPhone +65 6836 5437

    JapanPhone +81 3 5842 4011

    extremenetworks.com

    TRILL and SPB

    M-LAG is one of several approaches to building outmodern data center network fabrics. TRILL and ShortestPath Bridging (SPB) are two other new approaches that

    are being positioned as an alternative and replacement tothe spanning tree protocol. TRILL and SPB are competingproposals being pursued in the IETF and IEEE respectively.Both TRILL and SPB use link state routing protocols tocompute optimal paths between nodes in the network.However, unlike traditional link state routing protocolsthat operate at the IP or Layer 3 level, both TRILL and SPBoperate at the Layer 2 level. While SPB leverages the IS-ISlink state protocol, TRILL uses a variant of IS-IS. Addition-ally, both TRILL and SPB use encapsulation mechanisms totransport packets across the network. TRILL uses a form of

    MAC-in-MAC encapsulation while SPB has variants for bothMAC-in-MAC as well as Q-in-Q encapsulation. Both TRILLand SPB provide for multiple active redundant paths toeffectively fully utilize available bandwidth.

    The challenge with both TRILL and SPB is that they arenew protocols that require understanding and expertiseof two new technologies in the data center, IS-IS andMAC-in-MAC encapsulation. Additionally depending onthe protocol, multicast forwarding can require comput-ing additional multicast trees which can add furthercomplexity from a troubleshooting and debugging

    perspective. Lastly, both TRILL and SPB require new infra-structure due to the encapsulation mechanisms they usefor data forwarding i.e. most existing data center networkinfrastructures will not support either TRILL or IS-IS.

    Different vendors have expressed support for either TRILL or SPB leading to some confusion as to the industrydirection for both TRILL and SPB. In the face of this uncer-tainty around TRILL and SPB, and a lack of broad supportacross product lines for either TRILL or SPB, the M-LAGDirect Attach approach provides a viable alternativetoday to deploying a scalable and resilient data center

    fabric.