24
DATA CENTER FCIP Trunking FCIP Trunking is one of the advanced features of the Brocade extension platforms, enabling bandwidth aggregation and lossless failover for increased resiliency over IP WANs. This paper details the operation and advantages of this technology in extended storage applications.

FCIP Trunking

  • Upload
    buinhi

  • View
    262

  • Download
    0

Embed Size (px)

Citation preview

Page 1: FCIP Trunking

DATA CENTER

FCIP Trunking FCIP Trunking is one of the advanced features of the Brocade extension platforms, enabling bandwidth aggregation and lossless failover for increased resiliency over IP WANs. This paper details the operation and advantages of this technology in extended storage applications.

Page 2: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 2 of 24

CONTENTS

Introduction..........................................................................................................................................................................................................................................3  FCIP Trunking Overview...................................................................................................................................................................................................................3  

Keepalives and Circuit Timeouts ....................................................................................................................................5  FCIP Batching ...................................................................................................................................................................9  Ethernet Interfaces ....................................................................................................................................................... 12  Gigabit Ethernet Interfaces .......................................................................................................................................... 12  10 Gigabit Ethernet Interfaces .................................................................................................................................... 12  Other Features and FCIP Trunking .............................................................................................................................. 13  

Design Examples ............................................................................................................................................................................................................................14  High Availability Design ................................................................................................................................................ 14  Site-to-Multisite Design ................................................................................................................................................ 15  Ring Architectures......................................................................................................................................................... 18  Three-Site Triangle with Failover Metrics .................................................................................................................... 19  FICON Dual 10 GbE Links—IOD (In-Order Delivery) .................................................................................................... 22  

Conclusion.........................................................................................................................................................................................................................................23  

Page 3: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 3 of 24

INTRODUCTION Over the last decade, extension networks for storage have become commonplace and continue to grow in size and importance. Growth is not limited to new deployments but also involves the expansion of existing deployments. Requirements for data protection will never ease, as the economies of many countries depend on successful and continued business operations and thus have passed laws mandating data protection. Modern-day dependence on Remote Data Replication (RDR) means there is little tolerance for lapses that leave data vulnerable to loss.

In mainframe and open systems environments, reliable and resilient networks—to the point of no frame loss and in-order frame delivery—is necessary for error-free operation, high performance, and operational ease. This improves availability, reduces risk, reduces operating expenses and, most of all, reduces risk of data loss. Brocade has been developing advanced Fibre Channel over IP (FCIP) extension solutions for more than 15 years and continues to be the thought leader in extension solutions for all storage applications. Brocade addresses today’s important trends and requirements with the Brocade® 7800 Extension Switch and the Brocade FX8-24 Extension Blade for the Brocade DCX® 8510 and DCX Backbone family.

FCIP TRUNKING OVERVIEW FCIP Trunking is one of the advanced extension features of the Brocade extension platforms, providing the following benefits:

• Single logical tunnel comprised of one or more individual circuits • Efficient use of VE_Ports • Aggregation of circuit bandwidth • Failover • Failover metrics • Use of disparate characteristic WAN paths • Lossless Link Loss (LLL) • In-Order Delivery (IOD) • Non-disruptive link loss • Single termination point for protocol acceleration

FCIP Trunking in essence provides a single logical tunnel comprised of multiple circuits. Terminology can be confusing here. A single-circuit FCIP tunnel is referred to as an FCIP tunnel. An FCIP tunnel with multiple circuits is referred to as an FCIP trunk, simply because multiple circuits are being trunked together. An FCIP tunnel or FCIP trunk is a single Inter-Switch Link (ISL) and should be treated as such in fabric designs. Circuits are individual FCIP connections within the trunk, each with its own unique source and destination IP address. On older Brocade extension platforms that had multiple FCIP connections, such as the Brocade 7500/FR4-18i, each connection was its own tunnel, which is no longer the case. Now, with the Brocade 7800/FX8-24, a group of circuits associated with a single VE_Port forms a single ISL (FCIP trunk).

Because an FCIP tunnel is an ISL, each end requires its own Virtual E_Port, known as a “VE_Port.” Or, if it is a router demarcation point for a “remote edge” fabric, then it is called a “VEX_Port.” Remote edge fabrics are edge fabrics connected through a WAN connection and VEX_Port. FCIP Trunking operates the same whether the virtual port is a VE_Port or a VEX_Port. If each circuit is its own FCIP tunnel, a VE_Port is required for each circuit, but with FCIP Trunking each circuit is not its own FCIP tunnel, hence the term “circuit.” Since an FCIP trunk is logically a single tunnel, only a single VE or VEX_Port is used, regardless of the fact that more than one circuit may be contained within the tunnel.

Figure 1 on the next page shows an example of two FCIP tunnels that are trunking four circuits in tunnel 1 and two circuits in tunnel 2. Each circuit has been assigned a unique IP address by way of virtual IP interfaces within the Brocade 7800/FX8-24. Those IP interfaces are, in turn, assigned to Ethernet interfaces. In this case, each IP interface has been assigned to a different Ethernet interface. This is not required, however. Ethernet interface assignment is flexible depending on the environment’s needs and assignments can be made as desired. For instance, multiple IP interfaces can be assigned to a single Ethernet interface. The circuit flows from IP interface to IP interface through the assigned Ethernet interfaces.

Page 4: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 4 of 24

Figure 1. FCIP tunnels and their IP/Ethernet interfaces and circuits

Within the architecture of the Brocade 7800 and FX8-24 there are Brocade ASICs (application-specific integrated circuits) named the GoldenEye2 (GE2) for the Brocade 7800 and the Condor2 (C2) for the Brocade FX8-24. These ASICs know only Fibre Channel (FC) protocol. The VE/VEX_Ports are logical representations of actual FC ports on those ASICs. In actuality, multiple FC ports feed a VE/VEX_Port, permitting data rates well above 8 Gbps, which is necessary for 10 Gigabit Ethernet (GbE) FCIP interfaces on the Brocade FX8-24 and for feeding the compression engine at high data rates. Think of VE and VEX_Ports as the transition point from the FC world to the TCP/IP world inside these devices.

Since a single VE/VEX_Port represents the endpoint of a trunked FCIP ISL, this affords the Brocade 7800 and FX8-24 some benefits. First, fewer VE/VEX_Ports are needed, thus the remaining virtual ports are available for other FCIP tunnels to different locations. Typically, only one VE/VEX_Port is needed to any one remote location. Second, bandwidth aggregation is achieved by merely adding circuits to an FCIP trunk. Each circuit does not have to be configured identically to be added to a FCIP trunk; however, there are limitations to the maximum differences between circuits. For example, the maximum bandwidth delta has to fall within certain limits. Circuits can vary in terms of configured committed rate, specifically Adaptive Rate Limiting (ARL) floor and ceiling bandwidth levels. Circuits do not have to take the same or similar paths. The RTT (Round Trip Time) does not have to be the same; the delta between the longest and shortest RTT is not limitless and must fall within supported guidelines. However, those guidelines accommodate nearly all practical deployments. The resulting RTT for all circuits in the trunk is that of the longest RTT circuit in the group. Bandwidth scheduling is weighted per each circuit, and clean links operate at their full available bandwidth. A common example is FCIP deployed across a ring architecture, in which one span is a relatively short distance and the opposite span is longer. It is still practical to trunk two circuits, one across each span of the ring. Please refer to the Brocade Fabric OS FCIP Administrator’s Guide for more details pertaining to a specific Brocade Fabric OS® version.

Link failover is fully automated with FCIP Trunking and, by using metrics, active and passive circuits can coexist within the same trunk. Each circuit uses keepalives to reset the keepalive expiration timer. The time interval between keepalives is a configurable setting on the Brocade 7800 and FX8-24. If the keepalive expiration timer expires, the circuit is deemed to be down and is removed from the trunk. In addition, if the Ethernet interface assigned to one or more circuits loses light, those circuits immediately are considered down. When a circuit goes offline, the egress queue for that circuit is removed from the load-balancing algorithm, and traffic continues across the remaining circuits, albeit at the reduced bandwidth, due to the removal of the offline link.

Page 5: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 5 of 24

Keepalives and Circuit Timeouts 1. How does the keepalive mechanism work on the FCIP trunk/tunnel/circuit?

A keepalive timeout value is configured for each circuit in an FCIP trunk or the circuit in an FCIP tunnel. A tunnel has one circuit, and a trunk has more than one circuit. The trunking algorithm uses that value to determine how frequently keepalive frames need to be sent over the circuit. The math for that is not completely straightforward and is beyond the scope of this document. The important thing to note is that the algorithm ensures that multiple keepalive frames are sent within the timeout period. If the receiving side algorithm does not receive any keepalives within the timeout period, the circuit is brought down due to keepalive timeout.

2. How are keepalives queued and handled through the network layers?

Keepalives are treated like normal data in Storage-Optimized TCP (SO-TCP). They are not sent on any special path through the network stack; therefore, they are intermixed with normal FCIP data that is passing through a TCP connection. However, they bypass any data queued up in scheduler waiting to go to TCP and go directly to TCP to be sent. This means that the transmission of keepalives is guaranteed through the network by SO-TCP, so it cannot be lost unless a circuit is down. The keepalive process determines true delay in getting a packet through an IP network. This mechanism takes latency, reorder, retransmits, and any other network conditions into account.

If packets are taking longer than the allotted amount of time to get through the IP network and are exceeding the keepalive timeout value, the circuit is torn down, and that data is rerouted to the remaining circuits. Configuring the keepalive timeout value based on the timeout value of the application passing through the FCIP trunk is important and recommended. Having said this, the default values for open systems and FICON (Fiber Connectivity) are most applicable in nearly all cases and should not be changed without the assistance of Brocade support. The keepalive timeout value should be less than the application level timeout when an FCIP trunk contains multiple circuits. This facilitates LLL without the application timing out, ensuring that data traversing the IP network will not time out at the application level. TCP is the preferred transport for the keepalive mechanism, because it allows tight control of the time that a segment is allowed on the WAN.

3. Can having too much data going over a connection bring down a trunk/tunnel/circuit?

Yes and no.

The answer is no if there are no issues in the IP network, and the tunnel is just not big enough to handle the workload being driven from the application. This means there is no congestion; ARL is configured properly, but the bandwidth is just not adequate for the RDR or tape application to keep up. All the configured available bandwidth is used, and flow control using Buffer-to-Buffer Credits (BBCs) apply backpressure to the incoming FC flows, preventing buffer overflow; nonetheless, the tunnel does not go down due to a keepalive timeout. The time it takes to get a keepalive through is not affected by having large amounts of data queued in the egress scheduler. It is impacted by the SO-TCP data queue. However, the amount of data that can be queued—but not sent to the network—is relatively small (in the range of only a few milliseconds as of Brocade Fabric OS v7.0). This is not enough time to cause a keepalive timeout due to circuit saturation.

Now to the “yes” part: If the circuit is congested because ARL has been misconfigured, allowing more data to enter the IP network than there is available bandwidth, and the network cannot handle the data rate, then network errors result. If these errors become severe enough, leading to buffer overruns and causing packet loss and retransmits, then a keepalive may be lost, and a timeout may result.

When a connection is lost, data is almost always lost as well. Any FCIP frames in the process of being transmitted at the time of the link outage are lost, due to partial transmission. This causes an Out-Of-Order-Frame (OOOF) problem, because some frames have already arrived, then one or more are lost, and frames continue to arrive over the remaining link(s). Now the frames are not in sequence because of the missing frame(s). This is problematic for some devices, particularly mainframes, resulting in an Interface Control Check (IFCC). For example, frames 1 and 2 are sent and received. Frame 3 is lost. Frame 4 is sent and received. The receiving side detects an OOOF, because it was expecting frame 3 but received frame 4.

Page 6: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 6 of 24

FCIP Trunking resolves this problem with LLL. When a link is lost, inevitably frames in flight are lost also, and the lost frames are retransmitted. Refer to Figure 2. Normally, when a TCP segment is lost due to a dirty link, bit error, or congestion, TCP retransmits that segment. In the case of a broken connection (a down circuit), there is no way for TCP to retransmit the lost segment, because TCP is no longer operational across the link.

Figure 2. Lossless Link Loss (LLL) process.

Using proven Brocade technology often utilized with mainframes, an elegantly simple solution is to encapsulate all the TCP sessions, each associated with an individual circuit within the trunk. This outer TCP session is referred to as the “Supervisor” TCP session, and it feeds each circuit’s TCP session through a load balancer and a sophisticated egress queuing/scheduling mechanism that compensates for different link bandwidths, latencies, and congestion events.

The Supervisor TCP is a Brocade proprietary technology, purpose-designed, and a special function algorithm. It operates at the presentation level (Level 6) of the Open Systems Interconnection (OSI) model and does not affect any LAN or WAN network device that may monitor or provide security at the TCP (Level 4) or lower levels, such as firewalls, ACL or sFlow (RFC 3176).

If a connection goes offline, and data continues to be sent over remaining connections, missing frames indicated by noncontiguous sequence numbers in the header trigger an acknowledgment by the Supervisor TCP session back to the Supervisor source to retransmit the missing segment, even though that segment was originally sent by a different link TCP session that is no longer operational. This means that segments that are held in memory for transmission by TCP have to be managed by the Supervisor in the event that the segment has to be retransmitted over a surviving link TCP session.

Page 7: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 7 of 24

Circuits can be active or passive within an FCIP trunk and configured with metrics. All circuits with the lowest metric value are active, for example, a value of 0. Circuits with a value of 1 are not active, and they only become active after all circuits with a value of 0 have gone offline. Refer to Figure 3. This permits configuration of circuits over paths that should not be used unless the normal production path has gone down, for example, a backup path. In a three-site triangular architecture, in which normal production traffic takes the primary path that is shortest in distance and latency with one hop, a metric of 0 is set. Sending traffic through the secondary path, which has two hops and typically is longer in distance and latency, is prevented unless the primary path is interrupted. This is done by setting a metric of 1 to those circuits. Nonetheless, the dormant circuits are still members of the FCIP trunk. FCIP Trunking is required to assign metrics to circuits.

Figure 3. FCIP trunk circuits with metrics.

FCIP Trunking also provides a single point of termination for multiple circuits. This is important for protocol optimization techniques like FastWrite, Open Systems Tape Pipelining (OSTP), and FICON emulation. Prior to FCIP Trunking, it was not possible to perform protocol optimization over multiple IP connections. Each connection was an individual FCIP tunnel with its own VE/VEX_Port, because traffic may take different paths outbound versus inbound. With or without being attached to an edge fabric, Dynamic Path Selection (DPS) across Virtual E_Ports (VE or VEX) on the Brocade 7800 GE2 ASIC or Brocade FX8-24 C2 ASIC prevents a bidirectional deterministic path. Using protocol optimization, it is necessary to confine traffic to a specific path bidirectionally. This can be accomplished in a few ways: 1) Use only a single physical path; 2) Configure Traffic Isolation Zones (TIZs); 3) Set Brocade FOS APTpolicy to port-based routing; or 4) Use Virtual Fabrics (VFs) with Logical Switches (LSs). All four of these methods preclude any type of load balancing and, in some cases, failover as well.

The reason why optimized traffic must be bidirectionally isolated to a specific path is because protocol optimization uses a state machine to perform the optimization. Protocol optimization needs to know, in the correct order, what happened during a particular exchange. This is so that it can properly deal with the various sequences that make up the exchange until the exchange is finished, after which the state machine is discarded. These state machines are created and removed with every exchange that passes over the FCIP connection. The Brocade 7800/FX8-24 has the capacity for tens of thousands of simultaneous state machines or the equivalent number of flows. The ingress FC frames are verified to be from a data flow that can indeed be optimized. If a data flow cannot be optimized, it is merely passed across the FCIP trunk without optimization.

These state machines reside within each FCIP tunnel endpoint. A state machine cannot exist across more than one VE/VEX_Port, and there is no communication between different state machines. As shown in Figure 4 below, if an exchange starts out traversing FCIP tunnel 1 and returns over FCIP tunnel 2, with each trunk using a different VE_Port, the exchange passes through two different state machines, each one knowing nothing about the other. This situation causes FC protocol and optimization to break, producing error conditions.

Page 8: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 8 of 24

Figure 4. Protocol optimization breaks with more than one tunnel.

An advantage to FCIP Trunking is that logically there is only a single tunnel and a single endpoint at the VE or VEX_Port on each side of the trunk, regardless of how many circuits exist within the trunk. The state machines exist at the endpoints of the Supervisor TCP and remain consistent even if an exchange uses circuit 1 outbound and circuit 2 inbound. Refer to Figure 5. FCIP Trunking permits protocol optimization to function across multiple links that are load balanced and can failover with error-free operation and without any disruption to the optimization process.

Figure 5. Protocol optimization with FCIP Trunking as a single logical tunnel.

It is important to keep in mind that if more than one VE/VEX_Port is used between two data centers, the Brocade 7800 and FX8-24 can route traffic indeterminately to either of the virtual ports, which breaks protocol optimization. In this case, one of the isolation techniques described above is required to prevent failure by confining traffic flows to the same trunk or tunnel. There is one FCIP tunnel or trunk per VE_Port or VEX_Port.

A key advantage of Brocade FCIP Trunking is that it offers a superior technology implementation for load balancing, failover, in-order delivery, and protocol optimization. FastWrite disk I/O protocol optimization, OSTP, and FICON Acceleration are all supported on Brocade extension platforms. Brocade protocol optimization requires no additional hardware or hardware licenses. The same processing complexes that perform FCIP also perform protocol optimization for a higher level of integration, greater efficiency, better utilization of resources, lower costs, and reduced number of assets.

Brocade DPS Trunking (exchange-based trunking) is provided across FCIP complexes on the same blade or across blades within the same chassis. A combination of Brocade FCIP Trunking and Brocade DPS Trunking provides resiliency for the most demanding open systems environments. DPS is not supported in a mainframe FICON environment.

Page 9: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 9 of 24

FCIP Batching The Brocade 7800 and FX8-24 use batching to improve overall efficiency, maintain full utilization of links, and reduce protocol overhead. Simply put, a batch of FC frames is formed, after which the batch is processed as a single unit. Batching forms FCIP frames at one frame per batch. A batch is comprised of up to eight FC frames that have already been compressed using the exclusive Brocade FCcomp technology. All the frames have to be from the same exchange’s data sequence. Small Computer Systems Interface (SCSI) commands, transfer readies, and responses are not batched and are expedited by immediate transmission and/or protocol optimization. If the last frame of the sequence arrives, the end-of-sequence bit is set. The batch is then known to be complete and is processed immediately, even if fewer than eight frames have been acquired. Refer to Figure 6.

Figure 6. FCIP frame created by batching.

Because batches are specific to a flow between an initiator and target, and because that flow can be prioritized using Quality of Service (QoS) high, medium, and low designations, batches are specific to a priority and are fed into the correlating TCP session associated with that priority. This is referred to as Per-Priority TCP QoS (PP-TCP-QoS). PP-TCP-QoS is the only way that QoS can function within an IP network. It is not possible for QoS to operate within an IP network if all the priorities are fed into a single FCIP TCP session; in fact, this could cause severe performance degradation. Each circuit has its own SO-TCP sessions for each priority. Refer to Figure 8.

FCIP frames are then parsed into TCP segments according to the Maximum Segment Size (MSS), which specifies only the TCP payload size without the header. MSS is not configurable and is based on the IP Maximum Transmission Unit (MTU) minus the IP and TCP headers. Depending on the size of the FCIP frame, one frame may span multiple TCP segments, or it may fit within a single TCP segment. Each TCP segment is filled to its maximum size, as long as there is data to be sent. Refer to Figure 7. This method has the lowest amount of overhead, which results in superior performance and the highest efficiency achievable.

Page 10: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 10 of 24

Figure 7. FC into FCIP into TCP into IP method.

A batch is the unit of load balancing among the circuits in a trunk. Batches are placed in the egress queues by the Supervisor TCP session, using a Weighted Round Robin (WRR) type algorithm; this is referred to as scheduling. The scheduler does take into account egress bandwidth and queuing levels for each of the circuits. When a queue becomes full, usually because of disparate circuit characteristics, the scheduler skips that queue to permit it to drain, while continuing to service other circuits. This maintains full link utilization for different bandwidth circuits and circuits with dissimilar latency. The Supervisor TCP session on the receiving side ensures in order delivery of any batches that may arrive out of order due to circuit disparity and queuing. As mentioned previously, this means circuits do not have to have the same bandwidth, ARL settings, or latency (RTT), allowing circuits to use a variety of infrastructures like Dense Wavelength-Division Multiplexing (DWDM), Virtual Private LAN Services (VPLS), Multiprotocol Label Switching (MPLS), and carrier Ethernet.

If the queues for all the circuits become full and can no longer be serviced by the scheduler, the buffers fill, and eventually buffer-to-buffer credit R_RDYs are withheld from the source to stop data from building up within the Brocade 7800/FX8-24 and to prevent overflow and lost FC frames. Brocade Fibre Channel products never drop FC frames. There is a tremendous amount of buffer memory within the Brocade 7800 and FX8-24 platforms; however, there are advantages to limiting buffering to little more than the bandwidth-delay product (the amount of data that can be in flight) of the circuit. Robust buffers on the Brocade 7800/FX8-24 can accommodate multiple long fat pipes and a very large quantity of simultaneous flows. Additionally, limiting a flow’s buffering permits the source device to better know what has or has not been sent, by having as little data as possible outstanding within the network. Once queues drain to a particular point, FC frames resume flow from the source and new batches are produced. All the while, there is no pause in transmission on any of the circuits in the FCIP trunk, as none of the queues are ever completely emptied.

Page 11: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 11 of 24

Figure 8. Load balancing batches among circuits in an FCIP trunk.

An FCIP trunk has multiple circuits, which begs the question, “How are link costs computed with multiple circuits?” Furthermore, considering that ARL has a minimum and maximum bandwidth level set for each circuit, how does that affect Fabric Shortest Path First (FSPF) costs?

An FSPF cost is determined for the FCIP trunk. Circuits are not entities that are recognized by FSPF; therefore, circuits have no FSPF cost associated with them individually. FSPF cost is calculated from the sum of the maximum bandwidth levels configured, and only from online circuits that have the lowest metric within the FCIP trunk. For example, all online metric 0 circuits have their ceiling ARL value added, and that aggregate value is used in the computation.

FSPF link cost is computed using the following function. (Refer to Figure 9.)

• If the aggregate bandwidth is greater than or equal to 2 Gbps, the cost is 500.

• If the aggregate bandwidth is less than 2 Gbps and greater than 1 Gbps, the cost is 1,000,000 divided by the aggregate bandwidth amount in Mbps (from 500 to 1000).

• If the aggregate bandwidth is less than 1 Gbps, the cost is 2000 minus the aggregate bandwidth amount in Mbps (from 1000 up to 2000).

Figure 9. FSPF costs for FCIP trunks based on maximum aggregate bandwidth.

Page 12: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 12 of 24

If multiple circuits are assigned to the same Ethernet interface, the aggregate of the minimum bandwidth rates cannot exceed the interface speed. This prevents oversubscribing the interface when guaranteed minimum values have been configured. It is possible, however, to configure the aggregate of the maximum bandwidth values beyond the capacity of the physical Ethernet interface. This permits a circuit to use bandwidth that another circuit is not using at that moment. For example, two circuits have their maximum bandwidth level set to full line rate. If both are demanding bandwidth, then they equalize at 500 Mbps each. If one uses bandwidth only during the day, and one only at night, then each gets the available bandwidth of that interface at that time.

Ethernet Interfaces Ethernet interfaces are assigned by associating them with IP interfaces. IP interfaces on the Brocade 7800/FX8-24 are virtual entities within the platform. The IP interfaces, via their IP addresses, are assigned to circuits by designating the source IP address. The circuit is grouped into an FCIP trunk by designating the VE/VEX_Port to which it belongs. Tunnels/trunks are identified by their VE/VEX_Port. If an FCIP tunnel/trunk already exists, the circuit is added. If it is the first circuit, it forms a new FCIP tunnel.

Only one FCIP processing complex can control the circuits within a trunk. The Brocade 7800 has only one complex; however, the Brocade FX8-24 has two complexes.

Gigabit Ethernet Interfaces On the Brocade 7800/FX8-24 extension platforms, the GbE interfaces have been abstracted from the VE/VEX_Ports, IP interfaces, and FCIP tunnels/trunks. This means that the GbE interfaces are not fixed to the VE/VEX_Ports, IP interfaces, or FCIP tunnels/trunks. On the older Brocade 7500/FR4-18i platforms, the IP interface, tunnels, and Ethernet interfaces were all fixed in relation to each other; however, that is no longer the case.

There are six (6) Gigabit Ethernet (GbE) interfaces on the Brocade 7800 and ten (10) on the Brocade FX8-24. On these platforms, circuits within a trunk can be configured to use any combination of interfaces, because they are all controlled by a single FCIP complex. An individual circuit cannot be split across more than one Ethernet interface.

There are no specific requirements for data link connectivity to LAN and DWDM devices, other than that those devices should view the Brocade 7800/FX8-24 Ethernet ports as if a server Network Interface Card (NIC) is connected, and not another switch. No Ethernet bridging, Spanning Tree, or IP routing is occurring on the Brocade 7800/FX8-24 platforms, and the Ethernet interfaces are the origination and termination point of TCP flows—the same as most servers.

10 Gigabit Ethernet Interfaces The Brocade FX8-24 blade has two (2) 10 GbE interfaces, which are enabled by an optional license. The 10 GbE license also enables the second FCIP complex, doubling the capacity of the Brocade FX8-24 blade from 10 Gbps (1,000 MB/s) to 20 Gbps (2,000 MB/s).

There are two (2) separate FCIP complexes on the Brocade FX8-24 blade. One complex controls the ten 1 GbE interfaces or up to 10 Gbps of FCIP data for the two 10 GbE interfaces, and the other complex controls an additional 10 Gbps for the two 10 GbE interfaces. Either the 10GbE-1 interface or the ten 1 GbE interfaces may be enabled at any one time, but not both.

Each FCIP complex can generate a single 10 Gbps circuit, up to ten 1 Gbps circuits, or anything in between, in increments of 1 Gbps. Multigigabit circuits can be configured from 1 to 10 Gbps in 1 Gbps increments. The aggregate bandwidth of the circuits from an FCIP complex cannot be more than 10 Gbps, because the FCIP complex cannot process data faster than that. If less than an entire gigabit increment is used by ARL, an entire increment must still be consumed. For example, if your WAN is an OC-48 (2.5 Gbps), then a 3 Gbps circuit must be configured, and an ARL maximum (ceiling) value is set at 2.5 Gbps.

FCIP Trunking can be used on the 10 GbE interfaces to combine circuits into one logical ISL connection. For example, two 5 Gbps circuits are created with a metric of 0, and both stem from VE_Port 12. Circuit 1 is assigned to 10 GbE interface 0, and Circuit 2 is assigned to 10 GbE interface 1. The circuits take different OC-192 paths across

Page 13: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 13 of 24

different carrier networks. The two circuits join up again on the remote side. Logically, this is a single 10 Gbps ISL between the two data centers.

For FCIP Trunking, an FCIP complex must control all the member circuits. There is no distributed processing, load sharing, or lossless failover (LLL) across FCIP complexes. Failover between FCIP complexes is done at the FC level by the Brocade ASICs, provided the configuration permits it. The only shared components are the two 10 GbE interfaces. Not all the circuits have to be members of the same trunk (VE or VEX_Port) on a 10 GbE interface. There can be multiple FCIP trunks on a 10 GbE interface. The extreme is to create ten separate 1 Gbps circuits, each with their own VE/VEX_Port, which would create ten FCIP tunnels and be used to connect ten different locations. Typically, only 1 VE/VEX_Port is used to connect two sites with scaling, redundancy, and failover being facilitated by the member circuits. The location where the other end of the circuit connects is arbitrary. The IP network routes the traffic accordingly, based on the destination IP addresses. 10 GbE interfaces provide a single convenient connection to the data center LAN for one to ten circuits.

Other Features and FCIP Trunking Features such as compression, IPsec, VLAN tagging, Fibre Channel Routing (FCR), Virtual Fabrics (VF), and QoS all function with FCIP Trunking and without any adverse effects.

Compression is configured on a per-circuit basis. Some circuits may use one mode of compression, while others use a different mode. This is beneficial, because a slower speed link can use one of the higher compression ratio algorithms (like mode 2), while a higher speed link uses a faster rate algorithm (like mode 1). This provides ultimate flexibility and optimization of bandwidth resources.

IP Security (IPsec) is provided by Brocade at no additional cost. It is included from the entry level Brocade 7800 4/2 all the way up to the Brocade FX8-24 10 GbE platform. IPsec processes in hardware, adds a negligible amount of latency at approximately 5 µs, and runs at full line rate; therefore, there is no performance degradation when using IPsec, even for synchronous applications. FCIP Trunking is fully compatible with IPsec. Brocade IPsec costs only the time and effort to configure the feature, making it prudent to enable IPsec in most cases.

VLAN tagging (IEEE 802.1Q) inserts VLAN tag headers into Ethernet frames and is fully supported with FCIP Trunking. There is no requirement for circuits to participate in VLAN tagging. If tagging is needed, circuits can be individually configured into the same or different VLANs.

The Brocade 7800 and FX8-24 are the newest generation of extension products, with a focused design for high-performance extension. These products were not developed specifically as FC Routers (FCRs); however, FCR functionality was specifically designed into the Brocade 8G FC switching ASICs, found in the Brocade 7800/FX8-24. There is no longer a need for specialized hardware to perform FCR, and it is easily enabled with Integrated Routing (IR).

By enabling FCR on the Brocade 7800 or the Brocade DCX-8510/DCX/DCX-4S Backbone chassis in which Brocade FX8-24 blades have been placed, it is easy to build a best-practice Edge-Backbone-Edge architecture. IR also enables VEX Ports to be configured for Backbone-Remote Edge or Edge-Backbone-Remote Edge architectures.

Brocade Virtual Fabrics are useful when implementing Edge-Backbone-Edge with the Brocade FX8-24 blade, as shown in Figure 10. So that EX_Port connections can be made to an edge from the backbone, it is not necessary to purchase a separate Brocade DCX 8510 or DCX-4S to install the Brocade FX8-24 blades into. Instead, it is more cost-effective to put the Brocade FX8-24 blade directly into a Brocade 8510 or DCX that is part of the edge fabric and to create a backbone from the Base Logical Switch. The VE_Ports on the Brocade FX8-24 blade and EX_Ports are part of the Base switch. All other device connection ports and E_Ports that participate in the edge fabric belong to the Fab-A Logical Switch.

Page 14: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 14 of 24

Figure 10. Using VF to implement Edge-Backbone-Edge.

The Brocade 7800 and FX8-24 have various QoS functions, including Differentiated Services Code Point (DSCP), 802.1P (L2 CoS), and PP-TCP-QoS. DSCP and L2 CoS perform marking of IP packets and VLAN tags, respectively, and are fully compatible with FCIP Trunking.

PP-TCP-QoS is a special QoS function exclusive to Brocade extension and developed especially for high-performance FCIP. For QoS to function properly across an IP network, it is essential that each priority have its own flow. This is not possible using only a single TCP session, because there would only be a single merged flow. PP-TCP-QoS provides a TCP session for each priority flow, so that the IP network can perform according to the QoS settings. FCIP Trunking is fully compatible with PP-TCP-QoS.

DESIGN EXAMPLES

High Availability Design A popular design is the four Brocade 7800 High Availability architecture, as shown in Figure 11. This design can easily accommodate two WAN connections; however, in practice many enterprises suffice with a single WAN connection. This requirement is determined on a case-by-case basis and is predicated on cost versus risk.

Figure 11. Four Brocade 7800 High Availability Architecture.

The Brocade 7800 4/2 is an entry-level version of the Brocade 7800, with 4 active 8 Gbps FC ports and 2 active full-speed GbE interfaces. The Brocade 7800 4/2 has two enabled VE_Ports and is capable of 2 FCIP trunks. If dictated by future growth or tape applications, the Brocade 7800 4/2 can be upgraded by way of a software license to a Brocade 7800 16/6.

The four Brocade 7800 4/2 High Availability (HA) architecture uses a single FCIP trunk (a single VE_Port) to take advantage of a two-path (dual core) IP network. The FCIP trunks are equivalent to a single FC ISL between 1A and 2A, and another FC ISL between 1B and 2B. As shown in the diagram, there is a local site (Site 1) and remote site (Site

Page 15: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 15 of 24

2) and each site has two Brocade 7800s, one attached to controller A and one attached to controller “B.” There is a total of two FCIP trunks; one on each Brocade 7800, and each trunk has two circuits. Each circuit uses its own dedicated Ethernet interface. The A controllers use the blue and purple circuits within their trunk, and the B controllers use the green and red circuits, as shown in the figure. The circuits are consolidated by the data center LAN prior to being forwarded to the WAN.

Adaptive Rate Limiting (ARL) most likely is necessary in this architecture, since a total of four FCIP Ethernet interfaces from two different channel extenders are vying for the same bandwidth across the WAN. Please refer to the white paper on ARL for a detailed explanation of ARL uses and operation.

This design has a capacity—from the point of view of the storage array—of twice the bandwidth of the WAN connection, if 2:1 compression is achievable for the data using Lempel-Ziv (LZ) (mode 1). Often in practice the Brocade Deflate (mode 3) compression achieves greater than 4:1 compression. The compression ratio is specific to the actual data that is being compressed. The applicable compression mode depends on the available WAN bandwidth, and mode 3 is not appropriate in all cases. Brocade makes no promises, guarantees, or claims as to the compression ratio that your specific data will achieve.

Site-to-Multisite Design Figure 12 illustrates a Site to Multisite design example. (Note that this example only shows the A side of the Storage Area Network (SAN); there is a mirror of this for the B side. The bandwidths across the WAN must be doubled to account for both A and B fabrics, as shown in Table 1.) This design uses FCIP trunks that originate in the main data center and terminate in one of three remote sites. Both disk RDR and tape flows are traversing the network, and a significant amount of bandwidth is required for this purpose. FCIP Trunking is used to aggregate bandwidth and provides failover between multiple physical connections to the LAN.

Figure 12. Site to multisite design using FCIP Trunking, fabric A only.

Page 16: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 16 of 24

Table 1. Failure impact on bandwidth availability; Site to multisite dual fabric design using FCIP Trunking and ARL.

Bandwidth

Site A

Site B

Site C

ARL (min) Gbps

ARL (max) Gbps

Site A 1 Gbps

Incr.

Site B & C

1 Gbps Incr.

Site A Optic

Failure

Site B & C

Optic Failure

Site A Blade Failure

Site B & C

Blade Failure

Fabric A

Complex 0 10 GbE-0 10 GbE-1

2.5 2.5

0.833 0.833

1 1

6 3 3

3 0 3

0 0 0

Complex 1 10 GbE-0

10 GbE-1

0.625 0.625

0.625 0.625

0.625 0.625

1 1

4 1 & 1 1 & 1

0.83 0

0.83

0 0 0

Fabric B

Complex 0 10 GbE-0 10 GbE-1

2.5 2.5

0.833 0.833

1 1

6 3 3

6 3 3

6 3 3

Complex 1 10 GbE-0

10 GbE-1

0.625 0.625

0.625 0.625

0.625 0.625

1 1

4 1 & 1 1 & 1

1.67 0.83 0.83

2 1 1

Total (Gbps) 10.0 2.5 2.5 12.0 8.0 9.0 2.5 6.0 2.0

WAN Link BW 10 2.5 2.5 10.0 2.5 10 2.5

BW Delta (Failure) (1.00) (1.00) (4.00) (0.50)

% BW Available 90% 100% 60% 80%

For storage applications, the WAN from the main data center can accommodate a high bandwidth of 15 Gbps dedicated to just storage, and moderately high bandwidth at each remote site (OC-192 10 Gbps at site A, and OC-48 2.488 Gbps at sites B and C). There is only a single IP connection designated for storage at the remote sites, due to a cost versus risk determination. This is common at most enterprises.

Each of the 10 GbE interfaces contains FCIP trunks with member circuits destined for one of the remote sites. Brocade FX8-24 10 GbE interfaces can be configured with multigigabit circuits, however, that does not work when the receiving side is a Brocade 7800, because the interfaces accept only 1 Gbps. Therefore, multiple 1 Gbps circuits must be used with Brocade FCIP Trunking, to logically combine them.

Site A uses trunk 0 on FCIP complex 0 and has six 1 Gbps circuits. Remember that this refers to only fabric “A,” and there is a mirror of this on fabric “B,” both of which feed a single WAN connection. There is no mirror of the WAN, only on the SAN. To site A, 10 Gbps is the maximum that can be passed over the WAN, because it is an OC-192. Three of the trunk’s circuits are assigned to interface 10GbE-0 and three to interface 10GbE-1 for redundancy on the Brocade FX8-24 blade. Since optics are the most common failure in this type of system, it is prudent to have interface level redundancy. ARL is implemented in this design; if there is a failure or maintenance of a 10 GbE interface, an optic, a cable, a remote Brocade 7800, or an entire Brocade FX8-24 blade, the other circuits passing over the WAN increase their output to utilize as much bandwidth as possible. If there is no failure on fabric “B,” and only one of the two 10 GbE interfaces on fabric A are offline, then ARL adjusts the bandwidth up to 9 Gbps out of the 10 Gbps of available bandwidth. The failure condition is near perfect, providing 90 percent of the bandwidth to sustained operations. If an entire Brocade FX8-24 blade fails, or the A site Brocade 7800 fails, 60 percent of the bandwidth is maintained. During normal operation with no failures, ARL keeps each circuit at an aggregate 5 Gbps

Page 17: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 17 of 24

on each fabric (A and B), consuming all the bandwidth. This is a rate limit of about 833 Mbps, which is automatically set by ARL.

Site B uses trunk 1 on FCIP complex 1, has two 1 Gbps circuits, and uses 0.625 Gbps from each, for a total of 1.25 Gbps (½ OC-48 for fabric A to site B). One circuit is assigned to interface 10GbE-0, and one is assigned to interface 10GbE-1. Since there are only two circuits needed, the remote site could deploy a Brocade 7800 4/2 if there was no need for OSTP. In this example, however, there is tape, and OSTP is required, driving the need for a Brocade 7800 16/6. ARL is implemented in this design; if there is a failure or maintenance of a 10 GbE interface, an optic, a cable, a Brocade FX8-24 blade, or a remote Brocade 7800, rate limiting automatically adjusts to provide as much bandwidth as is available. If one 10 GbE interface or optic were to go offline, 3 Gbps of interface bandwidth would remain—2 Gbps on one fabric and 1 Gbps on the other—providing enough bandwidth for full utilization of an OC-48. If a Brocade FX8-24 blade or Brocade 7800 were to go offline on either fabric A or B, 2 Gbps would remain—reducing the operating environment down by only 0.5 Gbps and maintaining 80 percent of operational bandwidth.

Site C uses trunk 2 on FCIP complex 1 and is the same scenario as trunk 1 above.

It should be noted that it was not absolutely necessary to use the 10 GbE interfaces in this example. The ten 1 GbE interfaces on the Brocade FX8-24 blade could have been used instead. Also, it was not required to use both FCIP complexes, as a single complex can process enough FCIP data alone to accommodate the A side of this architecture, which is 7.5 Gbps. With failure scenarios, however, the use of two FCIP complexes provides a more robust solution. The 10 GbE interfaces do provide for consolidated connectivity within the data center. If the remote sites used Brocade FX8-24 blades instead of the Brocade 7800 switch, the use of 10 GbE interfaces with multi-gigabit circuits becomes very compelling.

The way the circuits are routed in the IP network is key to the redundancy and resiliency of this architecture. There are two core switches/routers, each connected to a 10 GbE interface on the Brocade FX8-24 blade. Please keep in mind that there is a mirror B side that connects to these same core switches.

The IP network routes the circuits to their destination using traditional methods. Nearly all types of WAN architectures are supported. The only architecture not supported is one that uses Per-Packet Load Balancing (PPLB) within the IP network. PPLB tends to work against TCP in most cases by causing excessive and chronic Out-Of-Order-Packets (OOOPs), which leads to delay of data to ULPs (Upper-Layer Protocols), higher response times, excessive CPU utilization, increased overall bandwidth utilization, and lower throughput. Flow-based load balancing in the IP network is fully supported and does not cause these issues.

Although not applicable in this specific example, the Condor2 ASIC normally performs Dynamic Path Selection (DPS) between the two FCIP complexes on the Brocade FX8-24 blade. These FCIP complexes are the engines that process FC into FCIP and perform all transport functions. DPS operates automatically; the APTpolicy is set accordingly when it sees equal-cost paths from the viewpoint of FSPF based on the FCIP trunks. DPS is an exchange-based FC routing technique and it load shares and fails over traffic between complexes based on hashing the source ID, destination ID, and exchange ID of flows. If one of the 10 GbE interfaces, optics, cable, or anywhere along the IP path were to fail; DPS would fail the traffic over to the other equal-cost path. DPS may be disabled in FICON applications via the APTpolicy, in which case, Port Based routing would be used instead, plus some method of traffic isolation if protocol optimization is being used.

In this specific case, DPS would not be used, because there are not multiple equal-cost paths within fabric A at site A, or within fabric A to site B, and so on. There is just one path. Also, there is just one VE_Port and one logical ISL. Redundancy is built into the overall infrastructure of A and B fabrics; multiple paths are not created within the same fabric using DPS. The storage applications have the ability to load balance their data over the multiple paths seen by the application. This means that the storage applications divide the workload evenly across both A and B fabrics.

Page 18: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 18 of 24

Ring Architectures Rings are another common architecture used for RDR and tape, due to the proliferation of DWDM infrastructure in many countries. There are many different kinds of offerings from service providers involving rings. Customers can buy service across only one span of the ring, permitting the service provider to sell the other span to a different customer. Or both spans can be commissioned at a higher cost for the purpose of aggregating bandwidth, or driving higher availability. Optionally, at a reduced cost only one span may be active, leaving the other span passive until it is needed for failover. This example discusses a ring in which both spans are active/active.

As shown in Figure 13, there is a single FCIP trunk with two circuits. Each circuit has been assigned to its own dedicated Ethernet interface, and the trunk is capable of 2 Gbps of aggregated bandwidth. The Brocade 7800 4/2 with Trunking and ARL is an appropriate solution for this example. Traffic is load balanced across the two circuits on a per batch basis. Of course, using ARL, the bandwidth can be rate limited to a smaller amount—as low as 10 Mbps per circuit. The amount of bandwidth on each circuit does not have to be the same. The ARL configuration on both ends of a circuit does have to be the same.

The Ethernet interfaces connect directly into the DWDM equipment provided by the service provider. This piece of equipment is often referred to as CPE (Customer Premise Equipment). The service provider must provision a GbE circuit across the blue span and a GbE circuit across the red span. Sometimes a full GbE circuit may not be practical, and an OC-12 must be considered; this operates the same way, but has less bandwidth (622 Mbps versus 1,000 Mbps). In the event of a fiber cut, which would bring down one of the ring’s spans, FCIP Trunking fails all traffic over to the remaining circuit, recovers any frames lost in transit, and—from the perspective of the storage array or mainframe—all frames arrive in order. When the ring is repaired, and the span comes online again, FCIP Trunking automatically adds the circuit back into the trunk and resumes transmitting and load balancing data. The keepalive mechanism is persistent, and it detects when the span comes online again.

Figure 13. Ring architecture utilizing FCIP Trunking over dissimilar circuits.

Page 19: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 19 of 24

Three-Site Triangle with Failover Metrics Many large enterprises, especially financial institutions involved with frequent and sizable transactions, have requirements for both synchronous RDR (RDR/S) and asynchronous RDR (RDR/A). Why use both? It is because RDR/S is required to confidently capture all writes to disk, except for the write that is in progress as the disaster takes out the server performing the current write. Other than that last write in progress, every write to disk is safely replicated to a remote location—often referred to as a “bunker site.” There are many things to consider in this type of architecture.

Due to the limitations of the speed of light, RDR/S has a limited distance before it causes poor response times for user applications. 100 km (62 mi.) is about the average maximum distance; however, the distance actually depends on a number of factors not discussed in this paper. RDR/S provides a significantly better RPO (Recovery Point Objective), which may represent a significant amount of transaction revenue.

Considering the possibilities and types of disasters that can occur in the world today, it is prudent for these large enterprises to replicate data to a third data center that is assured of being outside the perimeter of the event. To extend data beyond the bunker site, it is necessary to use RDR/A. Most of the major storage suppliers offer software that manages both RDR/S and RDR/A, so that if any one of the connections or sites fails, replication will continue without the need to reinitialize any of the volume pair relationships (which can take a considerable amount of time).

As shown in Figure 14, this example uses the Brocade FX8-24 blade in the Brocade DCX for the RDR/S connection and the RDR/A connections. Addressing the RDR/S connection, a primary design criterion for the Brocade FX8-24 was for it to be an extremely fast and low-latency technology that properly lends itself to RDR/S. In this example, one of the 10 GbE interfaces has been placed in its own VF LS (Logical Switch) and dedicated to RDR/S, which is best practice. RDR/A should run on a different LS, in which the interface for the other FCIP complex has been added. Again, this is best practice considering how critical the environment is. These mission-critical architectures also deploy A and B RDR fabrics for redundancy by way of completely physically isolated networks.

Figure 14. Three data center RDR architecture.

Page 20: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 20 of 24

Test results have shown faster response times relative to native FC implementations, as shown in Figure 15. Brocade hardware compression (mode 1) and IPsec add only a negligible amount of latency, allowing these features to be utilized in a synchronous environment. FastWrite protocol optimization should be enabled if needed, or left disabled if the RDR application does not make use of it. Additionally, RDR/S requires enough bandwidth for the demand peaks, and with 10 Gbps of bandwidth, plus hardware based compression achieving typical ratios of 2:1, the Brocade FX8-24 can supply much more bandwidth for peak loads compared to native FC. If there is not enough bandwidth for peak demands, data must be buffered or delayed, causing poor response times for synchronous applications. This is a benefit of using the Brocade FX8-24 blade in such an environment.

The IPsec AES encryption implementation on the Brocade FX8-24 blade is all done in hardware for each FCIP complex and can easily accommodate full line rate. In addition, it introduces only 5 µs of added latency.

Figure 15. Brocade FX8-24 versus native FC response times and throughput.

Another advantage of using the Brocade FX8-24 blade instead of native FC ports is that service providers charge significantly more money for FC DWDM lambdas compared to Ethernet, and FCIP uses Ethernet.

The infrastructure used to transport RDR/S and RDR/A also differs. RDR/S connections are short, and many service providers can offer DWDM for such distances at a reasonable price. RDR/A can operate with much less bandwidth relative to RDR/S, average versus peak. RDR/A needs only the high average over a finite period of time. The period of time cannot be so long that the average is artificially low—due to respites during the evening, for example. On the other hand, it cannot be so short that the average approximates the actual peak, like in RDR/S.

Another consideration when determining the adequate bandwidth for RDR/A is the capabilities of the storage array. Some storage arrays can journal data that is waiting to be sent across the link; however, this should be the exception, not the rule. More importantly, elongation of cycle times should be considered. The amount of data that has to be sent during each cycle can be considered a gauge of the average. If the available WAN bandwidth is below this average, cycle times tend to elongate. If the cycle times tend to always finish on time, and the link utilization is low, the average may have been overestimated, leaving ample room for growth. These issues must be evaluated on a case-by-case basis.

Page 21: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 21 of 24

Ping-ponging happens when a WAN connection to a site goes down, and traffic ping-pongs to other remote sites before finding its way to the destination. In Figure 16, one of the links (blue) goes down; however, the FC routing tables in the Brocade 7800/FX8-24 can still reach the final destination by ping-ponging the traffic back and forth across the WAN. Refer to Figure 16. The FSPF cost is greater, but it is still viable as it is the only path. Depending on the WAN links, this may introduce considerable latency, since a single RTT now becomes 3xRTT. Additionally, it causes undesired WAN utilization effects and breaks protocol optimization by performing optimization on exchanges that are already being optimized. Ultimately, this is not a best-practice architecture. The best-practice architecture is to create parallel A and B paths and not to cross-connect FCIP ISLs across the WAN infrastructure. This also reduces overall availability by not keeping A and B completely separated from array to array. Alternatively, this can be fixed using Traffic Isolation Zones (TIZs) with failover disabled or VF LS (Brocade DCX only) to prevent an alternate path in the event a link goes down. However, that would still not be best practice.

Figure 16. Ping-ponging.

Of course, rerouting over a less desirable path can happen in the three data center architecture, because each site is connected by two links in a triangle configuration. This is not desirable. For example, the RDR/S DWDM connection drops, and traffic starts to find its way via the two links with 30 ms of RTT (Round Trip Time) each, for a grand total of 60 ms RTT. That does not work well for synchronous traffic, of course.

To prevent ping-ponging in this design, either VF LS or TIZs have to be used. Failover has to be disabled for the synchronous connection, because if the DWDM link fails, there is no viable alternative path. This is clear. Alternatively, the synchronous 10 GbE interface from the Brocade FX8-24 blade should be connected directly to the DWDM CPE equipment, to prevent FCIP flows from being routed onto the IP network, which is the long way around the triangle.

As for the RDR/A traffic, whether you should reroute it across the DWDM link and then to the remote site, or whether you should prevent rerouting, and let the bunker site continue the replication process, depends on a few things: First, is there enough bandwidth available on the DWDM network to failover the RDR/A traffic onto it and maintain peak RDR/S demand? Often, there is not. Second, what is the RDR application software’s best practice? Best practice is often to prevent the RDR/A flows from being rerouted to the bunker site and then to the remote site, unless separate bandwidth is available that is specific to this purpose. This is because all too often, encroachment onto the RDR/S bandwidth is not workable.

Preventing RDR/A traffic from taking the DWDM path works the same way as for the RDR/S traffic. TIZs or VF LS (Brocade DCX only) for the RDR/A traffic prevents it from taking the DWDM link. As stated above, the RDR/A 10 GbE interface should not have any direct or indirect connectivity to the DWDM equipment, making DWDM not an option. In this case, the DWDM network should be its own isolated network. Alternatively, you can configure the IP network to prevent specific traffic from taking a particular route.

Page 22: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 22 of 24

FICON Dual 10 GbE Links—IOD (In-Order Delivery) Mainframe environments are often high-bandwidth environments that can take advantage of the Brocade 10 GbE FCIP interfaces on the Brocade FX8-24 blade. In this design example, a local and remote site are connected together using two 10 Gbps FCIP trunks within the two 10 GbE links, shown in blue and green (refer to Figure 17). The blue trunk is dedicated to IBM XRC Global Mirror, and the green trunk is dedicated to FICON tape. One Brocade FX8-24 blade has a total of 20 Gbps of capacity using two FCIP complexes. 10 Gbps of bandwidth is associated with complex-0 and 10 Gbps is associated with complex-1.

Two 5 Gbps multigigabit circuits are configured for each FCIP trunk. Data flows must remain isolated from end to end to prevent Interface Control Checks (IFCCs) on the mainframe. This includes the FCIP complex, because lossless DPS or DLS from the Condor2 ASIC to the internal FCIP complexes is not supported. Isolation is best done using VF LS, putting each VE_Port into its own respective LS. Data flows cannot pass between LS boundaries. 10 GbE interfaces must also be assigned to a LS; however, foreign logical switches can share Ethernet interfaces that belong to a different LS. This allows Ethernet interfaces to accommodate circuits coming from trunks that are not in its own LS. This is a new capability as of Brocade FOS 7.0. Alternatively, TIZ with failover disabled can be used as well.

Each 5 Gbps multigigabit circuit member of the trunk is assigned to each of the two 10 GbE interfaces on the Brocade FX8-24 blade. There are two 5 Gbps circuits on each 10 GbE interface, one from each FCIP trunk (blue and green). If one of the WAN connections or a segment of the data center LAN goes offline, the circuits taking the other path maintain connectivity to both FICON tape and XRC, even though each FCIP trunk receives half of the bandwidth at 5 Gbps each.

Once the circuits enter the core L2/L3 switch, because each circuit has its own source and destination IP address pair, the individual circuits can be routed across the different WAN connections as necessary, possibly using VLAN tagging (802.1Q). The Brocade FX8-24 blade can tag the Ethernet frames of a specific circuit such that the circuit’s data uses the particular path of that VLAN. QoS (802.1P and/or DSCP) can also be marked on these frames. The VLAN’s path is determined and configured by the IP network administrators. Policy-Based Routing (PBR) is another alternative to directing data flows over a specific path.

The FICON frames that are trunked across multiple similar or dissimilar WAN connections, as in this example, are always delivered to the ULP in the proper order in which it was sent. This is a function of the Supervisor TCP session that manages the data across all the circuits in a trunk. This is a requirement in mainframe environments and a unique Brocade technology.

FICON applications require lossless failover to prevent any frame loss. When one or more frames are lost, and traffic resumes immediately after the failover or rerouting of data, this causes what appears to be Out Of Order Frames (OOOFs). OOOF results in an IFCC on mainframes, which is a significant event. To prevent IFCCs, it is better to completely halt the traffic, versus rerouting and keeping flows moving, which is not entirely intuitive. Because of this, TIZs are set to not failover when a path is no longer valid or, as in the case of VF LS, when there is no other path to fail to within the same LS. In this example, both links have to go down before traffic completely stops. Moreover, traffic is not stopped due to TIZ not failing over; rather, it is because literally there are no more physical paths to the remote site.

It is appropriate to implement Virtual Fabric Logical Switches (VF LSs) or TIZs to constrain traffic to their respective paths in a FICON environment. This is needed only when more than one VE_Port is connected to the opposing data center. If all data flows are using one FCIP trunk (denoted by using only a single VE_Port at each data center within the same fabric, for example, the A fabric), then no traffic isolation or VF LS is required. This is because path ambiguity is eliminated, and failover to a different VE_Port by the rerouting of FICON is not possible. There is only one set of VE_Ports connected by a single logical ISL.

Brocade FICON Acceleration supports both XRC and tape across the same circuits simultaneously. There is no requirement to separate the data flows; however, it is the practice of many FICON shops to keep the data flows separate. In fact, if desired, simultaneous FICON and Open Systems data flows across the same FCIP trunk are fully supported by Brocade and IBM. Special consideration must be given in such blended architectures, which is beyond the scope of this paper.

Page 23: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 23 of 24

Figure 17. FCIP Trunking used with mainframe applications.

CONCLUSION Brocade extension solutions offer many advanced features that have been developed for a wide variety of critical environments found in small/medium businesses as well as the largest, most demanding enterprises. As a thought leader and innovator in extension technology, Brocade was the first to develop technologies such as FCIP Trunking, FC Routing, FICON Emulation, 10 GbE FCIP, ARL, FastWrite, OSTP, FCcomp, FCIP IPsec, Per-Priority TCP QoS, and much more. These critical capabilities deliver unmatched performance, efficiency, and flexibility, and drive down capital and operating costs.

In addition to industry-leading extension technology, Brocade has deep technical expertise and offers assessment, design, and implementation services to help organizations optimize an existing SAN, or architect a new SAN that meets the requirements of your specific environment. When it comes to distance extension solutions, no one can match the FCIP experience and best practice deployment knowledge of Brocade or can offer as thorough a service for all storage and tape platforms.

Page 24: FCIP Trunking

DATA CENTER TECHNICAL BRIEF

Brocade FCIP Trunking 24 of 24

© 2012 Brocade Communications Systems, Inc. All Rights Reserved. 01/12 GA-TB-418-00

Brocade, Brocade Assurance, the B-wing symbol, DCX, Fabric OS, MLX, SAN Health, VCS, and VDX are registered trademarks, and AnyIO, Brocade One, CloudPlex, Effortless Networking, ICX, NET Health, OpenScript, and The Effortless Network are trademarks of Brocade Communications Systems, Inc., in the United States and/or in other countries. Other brands, products, or service names mentioned may be trademarks of their respective owners.

Notice: This document is for informational purposes only and does not set forth any warranty, expressed or implied, concerning any equipment, equipment feature, or service offered or to be offered by Brocade. Brocade reserves the right to make changes to this document at any time, without notice, and assumes no responsibility for its use. This informational document describes features that may not be currently available. Contact a Brocade sales office for information on feature and product availability. Export of technical data contained in this document may require an export license from the United States government.