Data Center Top-of-Rack Architecture Design
Data Center Top-of-Rack Architecture Design
Data Center Top-of-Rack Architecture Design
Introduction
The data center landscape is changing rapidly. IT departments building new data centers, expanding existing data center footprints, or updating racks of equipment all have to design a cabling and switching architecture that supports rapid change and mobility and accommodate transitions to 10, 40, and 100 Gigabit Ethernet over time. The main factors that IT departments must address include the following:
Modularity and flexibility is of paramount importance: The need to rapidly deploy new applications and easily scale existing ones has caused server-at-a-time deployment to give way to a rack-at-a-time model. Many IT departments are ordering preconfigured racks of equipment with integrated cabling and switching and as many as 96 servers per rack. The time required to commission new racks and decommission old ones is now a matter of hours rather than days or weeks. Because different racks have different I/O requirements,
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 1 of 13
White Paper
data center switching and cabling strategies must support a wide variety of connectivity requirements at any rack position.
Bandwidth requirements are increasing: Todays powerful multisocket, multicore servers, blade systems, and integrated server and rack systems, often running virtualization software, are running at higher utilization levels and impose higher bandwidth demands. Some server racks are populated with servers requiring between five and seven Gigabit Ethernet connections and two Fibre Channel SAN connections each.
I/O connectivity options are evolving: I/O connectivity options are evolving to accommodate the need for increasing bandwidth, and good data center switching and cabling strategies need to accommodate all connectivity requirements at any rack position. Racks today can be equipped with Gigabit Ethernet or 10 Gigabit Ethernet or a unified network fabric with Fibre Channel over Ethernet (FCoE).
Virtualization is being added at every layer of the data center: Server virtualization is promoting server consolidation and increasing the need for bandwidth and access to network-attached storage (NAS). Virtualization is one of the main areas of focus for IT decision makers. Estimates suggest that the server virtualization market will grow by 44 percent over the next 4 years. The change to virtualization can be disruptive and necessitate a redesign of the networking infrastructure to gain all the benefits of the virtualized computer platform.
The challenge facing data centers today is how to support the modularity and flexibility that is needed to promote business agility and maintain a companys competitive edge. The same strategy that allows the intermixing of different rack types and I/O requirements must also support a varied set of connectivity options including Gigabit Ethernet and 10 Gigabit Ethernet as well as a unified network fabric.
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 2 of 13
White Paper
The TIA/EIA-942 specification provides a simple reference for data center cabling that supports different cabling schemes, EoR or ToR, to meet differing needs from a physical and operational perspective. The ToR model defines an architecture in which servers are connected to switches that are located within the same or adjacent racks, and in which these switches are connected to aggregation switches typically using horizontal fiber-optic cabling. ToR switching allows oversubscription to be handled at the rack level, with a small number of fiber cables providing uniform connectivity to each rack. The advantage of this solution is that horizontal fiber can support different I/O connectivity options, including Gigabit Ethernet and 10 Gigabit Ethernet as well as Fibre Channel. The use of fiber from each rack also helps protect infrastructure investments as evolving standards, including 40 and 100 Gigabit Ethernet, are more likely to be implemented using fiber before any other transmission mechanism. By limiting the use of copper to within racks, the ToR model isolates the cabling that changes most often to the parts of the data center that change most frequently: the racks themselves. The use of fiber runs from racks provides a flexible data center cabling infrastructure that supports the transition from Gigabit Ethernet to 10 Gigabit Ethernet now, while enabling transition to 40 and 100 Gigabit Ethernet in the future.
Support for storage devices (Fibre Channel, Small Computer System Interface [SCSI] or NAS, FCoE, etc.)
Support for convergence and unified fabric with growth factors incorporated Reliability, scalability, and redundancy High-capacity and high-density server access requirements Flexibility and expandability with easy access for moves, additions, and changes Migration from Gigabit Ethernet to 10 Gigabit Ethernet server connectivity with future support for 40 and 100 Gigabit Ethernet
Cabling architecture balance with power, cooling, structural loading, management, and operations needs
In the context of the TIA/EIA-942 simplified logical data center layout, the ToR architecture maps directly to the EDA and the horizontal distribution area (HDA). Figure 1 shows a mapping of the logical network architecture to the physical infrastructure.
Figure 1. TIA/EIA-942 Logical Layout
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 3 of 13
White Paper
The EDA in the TIA/EIA-942 logical layout corresponds to the area where server racks are placed. Traditional structured copper cabling with a mix of fiber if needed for SAN or high-speed server connectivity is used to connect the EDA to the HDA. The environment requires careful planning to help ensure that the structured cabling meets the initial design requirements with enough room for growth. In cases where server racks are not yet in place or without the physical infrastructure needed to support rack-level flexibility, a zone distribution area (ZDA) cabling model is used. The ZDA allows for structured cabling to be placed under the floor or above the rack in anticipation of future server racks requiring connectivity to the network equipment that may be housed in the HDA. The ZDA follows a structured cabling model to the HDA. The primary difference in the horizontal cabling model between the ZDA and EDA is that the cables are terminated in the EDA racks, whereas the ZDA uses zone distribution blocks located outside the server racks. The ToR cabling and network architecture in Figure 1 optimizes the requirement for horizontal cabling from the server rack by placing the ToR aggregation device at the top of the server rack. Actual placement of the ToR device may vary based on customer requirements (for example, in or above the server rack or ToR aggregation per two or three server racks) and seeks to optimize customer requirements for density, cabling, and design methodology. The aggregation of the ToR device will be in the HDA based on the specific TIA/EIA reference cabling model deployed. The ToR cabling design model follows a logical network layer construct in which the server network connectivity is aggregated at the rack level in the network access layer. The access layer is in turn connected to the network aggregation layer. The ToR model uses fiber as the backbone cabling infrastructure that connects the EDA with the HDA and main distribution area (MDA). The ToR model augments the TIA/EIA-942 logical approach shown in Figure 1 by extending fiber as the backbone cabling of choice to the EDA or server rack. The amount of fiber required will vary based on design requirements. For example, in a migration-type design in which a ToR model is used for Ethernet aggregation and Fibre Channel connectivity is not unified I/O (UIO), additional fiber will be required for those servers requiring
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information. Page 4 of 13
White Paper
Fibre Channel connectivity. In models in which UIO for Fibre Channel and Ethernet is assumed, the fiber requirement to the rack will be reduced. Other design augmentations, such as ToR for few racks with inter-rack cabling, will modify the fiber requirements. These are a few examples of some deployment models that have been adopted based on particular environment and requirements. The Pod Approach The easiest way to scale data center capacity is to use a highly modular architecture that enables the rapid deployment of infrastructure. One approach is to use a rack of servers as the base building block, with ToR switches, and with all copper cabling contained within the rack. The TIA/EIA model provides a logical cabling layout that can be modularized for a data center implementation. When deploying large numbers of servers in the data center, it is extremely important that the design footprint be scalable. One way to simplify the design and simultaneously incorporate a scalable layout is to divide the data center floor space into modular, easily duplicated subareas. This modular building block is used to design scalability into the network architecture at both OSI Layers 1 and 2. This design assumes that all computer resources incorporate resilient network, power, and storage resources, which translates to multiple LAN, SAN, and power connections within the physical layer infrastructure. The logical architecture is divided into three discrete layers (core, aggregation, and access), and the physical infrastructure is divided into manageable subareas called pods. From a network architecture perspective, a pod is defined by a pair of Ethernet aggregation switches. Pods support access-layer data link connectivity for low-latency interprocessor communications (IPC), unified fabric and unified IO, and Ethernet networks. A modular data center design coupled with a ToR access-layer architecture provides facilities and network architecture scalability, flexibility, and mobility to the rack level. Figure 2 shows a modular layout for a Greenfield data center. The pod-based modular concept incorporates the various cabling architectures that may be required within the data center. More important, it enables a modular building block approach for server rack-and-roll deployment. In the reference architecture shown, the data center floor plan is divided into zones that are further subdivided into pods. A pod is used to represent a building block that defines the number of servers connected to a network aggregation block. Within the pod, server connectivity is handled at the rack level by the ToR device, which is in turn connected to the network aggregation layer.
Figure 2. Modular Data Center Layout
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 5 of 13
White Paper
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 6 of 13
White Paper
Top-of-Rack Access Media Considerations When designing for 10-Gigabit Ethernet cabling infrastructure, the physics of the cabling plant, including signal attenuation, latency, and distance, as well as cabling installation and termination best practices, require careful consideration. Investment in the optimal cabling media for 10, 40, and 100 Gigabit Ethernet connectivity involves striking a balance among bandwidth, flexibility, and scalability. Table 1 shows the media and transceiver options for 10 Gigabit Ethernet connectivity available today.
Table 1. 10 Gigabit Ethernet Server Connectivity Options: UTP/F-UTP, MMF, SMF, Twinax, and CX4
Cable Twinax Twinax MM OM2 MM OM3 SFP+ SR MMF short reach MM OM2 MM OM3 RJ45 10GBASE-T copper Cat6 Cat6a/7 Cat6a/7 Distance <10m <15m 10m 100m 82m 300m 55m 100m 30m ~ 6W*** ~ 6W*** ~ 4W*** 2.5 s 2.5 s 1.5 s IEEE 802.3an 1W ~0 IEEE 802.3ae Power (Each Side) ~ 1.5W 4W 1W Transceiver Latency (Link) ~ .1 s ~ 0.1 s ~0 Standard SFF 8431** IEEE 802.3ak None
Connector (Media) SFP+ CU* copper X2 CX4 copper SFP+ USR MMF ultra-short reach
* Terminated cable ** Draft 3.0, not final *** As of 2008; expected to decrease over time
Several media and transceiver options are available today for 10 Gigabit Ethernet server connectivity. Considerations for use cases depend on variables such as cost, latency, distance, power consumption, and technology availability. For 10 Gigabit Ethernet server connectivity today, Small Form-Factor Pluggable (SFP) technology provides cost-effective and flexible options for server network interface cards (NICs) and ToR switches. SFP+ copper Twinax (SFP+cu) is currently the best choice for server connectivity. For EDA-to-HDA horizontal connectivity, SFP+ ultra-short-reach (USR) and short-reach (SR) cabling is better suited for longer distances between the server (EDA) rack and network equipment racks located in the HDA. SFP+ USR provides cost-effective fiber connectivity for fiber-based server connectivity options. SFP+ copper may also interconnect a number of EDA racks to a central EDA ToR device with horizontal backbone connectivity to the HDA or MDA. 10GBASE-T is currently power inefficient when compared to CX-1 and optical technologies. However, as more power-efficient third- and fourth-generation silicon architectures coupled with IEEE 802.3az Energy-Efficient Ethernet become available over the next few years, 10GBASE-T LAN on motherboard (LOM) and dense network switching products will become technically and economically viable server connectivity options. Fiber and Copper Cabling Characteristics A number of optics and media options are available for data center designs. Choices depend on the physical facilities requirements and media characteristics (physical medium dependent [PMD] factors) of the data center (Figure 3).
Figure 3. Media Options Based on Physical Requirements
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 7 of 13
White Paper
Fiber When considering a data center buildout, evaluation of the near- and mid-term future is required to maximize the cable plant investment. Because 10, 40, and 100 Gigabit Ethernet will be standardized in the next 3 to 5 years, organizations should assess fiber-optic cabling. Several grades of high-bandwidth laser-optimized fiber are available for use in high-speed network installations, each with a different reach and data rate:
62.5/125-micrometer (OM1) fiber, designed to achieve 10- and 100-Mbps data rates; now largely an obsolete fiber.
50/125-micrometer (OM2) fiber, used to achieve 1-Gbps data rates 50/125-micrometer (OM2+, OM3, and OM3+) fiber, used to achieve 10-Gbps data rates; OM2+ and OM3+ fiber grades offer nearly double the bandwidth of their parent fibers (+ indicates extended-reach OM2 and OM3 fiber)
Standard single-mode fiber (SMF) (ITU GR.652, TIA/EIA-493CAAA) is designed to support highcapacity, low-cost transmission components developed for the 1310-nanometer (nm) space. SMF features low dispersion and is optimized for use in the 1310-nm wavelength region. SMF is also used effectively with wavelength-division multiplexing (WDM) systems operating in the 15,550-nm wavelength region. SMF can be used for cross-aisle and interdata center applications. SMF will support 40 and 100 Gigabit Ethernet with serial implementations. The most cost-effective fiber cabling is multimode fiber (MMF). It is found in nearly all data centers today and can support 10 Gigabit Ethernet. However, as 40 and 100 Gigabit Ethernet become standardized, either MMF with parallel optics or SMF will be required. SMF, although simpler to manage and more amenable to higher bit-rate transmission than MMF, is more expensive to terminate and also more expensive in terms of the optical transceivers that are required at the switch and host devices.
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information. Page 8 of 13
White Paper
Copper Copper cabling can be used for 10 Gigabit Ethernet server connectivity. Category 6A (Cat6a) copper cabling was developed in conjunction with the 10GBASE-T standard to achieve reliable 10-Gbps operation over 100-meter copper twisted-pair channels. Cat6a shielded and unshielded products are designed to extend usable bandwidth up to 500 MHz and to drastically reduce alien crosstalk interference. In the context of ToR architectures, Cat6a can be used within the rack. 1X-based Twinax copper* is an alternative copper solution that uses SFP+ direct-attach Twinax cabling. Although this solution has a limited cable distance of up to 10 meters, it provides a robust, power-efficient, and cost-effective solution for 10 Gigabit Ethernet transmission. The SFP+ direct-attach solution is a fully integrated SFP+ cable that is available in multiple lengths up to 10 meters. As the cabling distance is limited, each server is directly connected to a ToR switch with no intermediate patch panels to manage. This approach dramatically simplifies cabling and termination as the cabling is contained within a single rack, and it works well with the concept of a modular data center. The SFP+ direct-attach solution draws 1.5 watts (W) of power per port, has a latency of 0.1 microsecond, and is available today. Twinax cabling has been optimized for differential pair applications to support 10-Gbps signaling.
1
In rack:
Cabling within the rack depends on connectivity requirements, which can be optimized for interface speed, latency, and cost of optics and transceiver
Copper options: SFP+ CX-1 Twinax (UTP, F/UTP, and S/FTP) Fiber options: Low-cost SFP+ USR or SX, SR, and LM for short reach over MMF
It employs a unique homogeneous construction with 100 percent shielding that enables new levels of data rates over a single line with essentially no crosstalk. Both delay and amplitude skew are reduced because the one integral dielectric eliminates material variations and forces a nearly identical distance between conductors to be maintained. Twinax cabling is limited today to 10 meters for passive implementation and approximately 30 meters for active cabling solutions.
Page 9 of 13
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
White Paper
Fiber (OM3 and OM4) if available; MMF recommended for lower cost on the fiber and the optics required for termination
Figure 4 shows a facilities rack view of a raised-floor server rack and network rack. The fiber-optic cabling is located over the server racks and terminates in fiber patch panels. Depending on requirements, the deployment may vary slightly, with the ToR device supporting a single rack or multiple racks.
Figure 4. Server and Network Racks
Top-of-Rack Architecture with the Cisco Nexus 5000 Series Switches and 2000 Series Fabric Extenders
The ToR architecture enables data center managers to implement a single cabling model that can support Gigabit Ethernet and 10 Gigabit Ethernet and unified network fabric today, while supporting future 40 and 100 Gigabit Ethernet standards as they come to market. Using a single overhead cable tray for fiber-optic cable management, data center managers have the flexibility to deploy preconfigured racks with different connectivity requirements in any rack position. For example, a rack of servers running multiple Gigabit Ethernet connections can be placed next to a rack of servers with 10 Gigabit Ethernet and FCoE connections to each server. As described here, Cisco Nexus products can be used to facilitate the ToR switching and cabling model. Top-of-Rack Solution in Gigabit Ethernet Environments Cisco offers a compelling ToR solution that is supported by Cisco Nexus products. Using the Cisco Nexus 2148T Fabric Extender and Cisco Nexus 5000 Series Switches at the access layer, data centers can build self-contained racks of servers with Gigabit Ethernet connectivity using a small number of 10 Gigabit Ethernet fiber or CX-1 connections to an EoR or middle-of-row (MoR) switch. The Cisco Nexus 2148T is an innovative server aggregation mechanism that extends Cisco Nexus 5000 Series Switches into Gigabit Ethernet environments. Acting as a line card on a Cisco Nexus 5000 Series switch, the Cisco Nexus 2148T Fabric Extender aggregates up to 48 Gigabit Ethernet (fixed speed ) connections at the rack and passes the network traffic up to the access-layer switch at the middle or end of the row. Because the Cisco Nexus fabric extender is an extension of the
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 10 of 13
White Paper
switch itself, it offers massive scale with no increase in management complexity. Physically, it distributes the access layer across data center racks. Logically, the access layer remains at the end of the row and is managed by the Cisco Nexus 5000 Series switch. The Cisco Nexus fabric extender provides up to four SFP+ 10 Gigabit Ethernet uplinks supporting either 10 Gigabit Ethernet SFP+ optical transceivers or CX-1 direct-attach cable assemblies. Figure 5 shows a row of dual-homed Gigabit Ethernet attached servers with two Cisco Nexus fabric extender devices in each rack that are connected to Cisco Nexus 5010 or 5020 Switches installed at the end of each row. Two Cisco Nexus fabric extenders in each rack support 40 servers per rack with no oversubscription if four 10 Gigabit Ethernet uplinks are configured. Gigabit Ethernet attached servers are connected to the ToR Cisco Nexus fabric extender using Cat5e RJ-45 patch cables. Fiber uplinks provide connectivity between the fabric extender and upstream Cisco Nexus 5010 and 5020 Switches. If all four fabric extender uplinks are used to reduce oversubscription in the rack, then a total of 8 fiber strands are used from each rack to the end of the row.
Figure 5. Distributed Access Fabric with Nexus 5010 at End of Row
Short cable runs for 10 Gigabit Ethernet can be supported with SFP+ direct-attach copper (CX1) cables, which provide a low-cost, low-latency, and power-efficient server connection. SFP+ directattach 10 Gigabit Ethernet copper cables can connect the ToR Cisco Nexus fabric extender devices to the Cisco Nexus 5010 and 5020 Switches installed in the middle of the row. MoR configuration reduces the distance for the horizontal cable run, bringing it within the 7-meter range of CX1 cable. This layout can accommodate five or six 24-inch racks on either side of the Cisco Nexus 5020 and 5010 Switches installed in the middle of the row (Figure 6).
Figure 6. Distributed Access Fabric with Nexus 5010 at Middle of Row
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 11 of 13
White Paper
The Cisco Nexus 5010 is ideal for supporting 16 to 20 servers per rack in a ToR configuration. With 20 fixed SFP+ ports that support both fiber and SFP+ direct-attach 10 Gigabit Ethernet copper, each server can connect to the switch using CX-1 copper. If more than 20 ports are required, then a single expansion module can be installed to support up to 6 additional 10 Gigabit Ethernet connections. This architecture facilitates rack-and-roll deployment by allowing redundant networks to be constructed with two switches per rack and occupying a total of two rack units (2RU). The Cisco Nexus 5000 Series Switches ease the transition to a unified network fabric by optionally carrying FCoE traffic from each server to the switch, and then native Fibre Channel from the switch to the SAN. Native Fibre Channel is supported through a choice of two expansion modules, one of which has eight Fibre Channel ports and one that has four 10 Gigabit Ethernet and four Fibre Channel ports. Figure 7 shows that the same model of fiber in cable trays above racks and copper cabling within racks can support a mixed Gigabit Ethernet and 10 Gigabit Ethernet attached server environment within the same row. 10 Gigabit Ethernet attached servers are connected to the Cisco Nexus 5010 and 5020 ToR switches using direct-attach CX1 copper cables. Cisco Nexus 5010 and 5020 Switches are connected to the upstream aggregation-layer switch using fiber cables in the overhead cable tray. For the Gigabit Ethernet attached server racks, servers are connected to the Cisco Nexus ToR fabric extender using regular patch cords while fabric extenders are wired to the Cisco Nexus 5010 and 5020 EoR switches using fiber cables.
Figure 7. Distributed Access Fabric for 10GE and 1GE attached Servers
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 12 of 13
White Paper
Conclusion
Data centers are undergoing a fundamental shift. Application developments and server virtualization are causing IT to seek a more flexible, efficient, and available infrastructure that can dynamically adapt to the needs of the business. To achieve this vision, data center architectures are changing to become more modular, so that data center managers can easily add racks of servers by adding racks that are precabled to a ToR switch and then connected to aggregationlayer switches using fiber-optic cables. To deliver the necessary bandwidth, servers will transition to 10 Gigabit Ethernet in the short term for server and interswitch connectivity. The transition to 10 Gigabit Ethernet will also enable the adoption of a unified fabric using FCoE when organizations are ready to take advantage of the technology to achieve further efficiencies within the data center. In the context of this data center evolution, cabling infrastructure must be considered to help ensure delivery of reliable high-bandwidth performance and application availability. Careful consideration also is needed to help increase power and cooling efficiency, as power and cooling have emerged as the top concerns for many organizations. Another consideration for Greenfield and Brownfield data center designs is the modular approach to data cabling and network architecture (this document refers to modules as pods). The modular approach improves data center design, deployment, and operations flexibility and increases efficiency and responsiveness, reducing overall capital expenditures (CapEx) and operating expenses (OpEx). The ToR architecture suits the modular approach and enables rack-and-roll deployment for further modularity and flexibility at the rack level. The ToR network and cabling architecture model enables a graceful transition from Gigabit Ethernet to 10 Gigabit Ethernet attached servers and provides infrastructure readiness for future adoption of 40 and 100 Gigabit Ethernet server and switch-to-switch connectivity without the need to change the cabling plant.
Printed in USA
C11-522337-00 02/09
2009 Cisco Systems, Inc. All rights reserved. This document is Cisco Public Information.
Page 13 of 13