Integrating Servers and Networking using an XOR-based Flat Routing Mechanism in 3-cube Server-centric Data Centers
|
|
|
- Garey Hines
- 10 years ago
- Views:
Transcription
1 Integrating Servers and Networking using an XOR-based Flat Routing Mechanism in 3-cube Server-centric Data Centers Rafael Pasquini 1, Fábio L. Verdi 2 and Maurício F. Magalhães 1 1 Department of Computer Engineering and Industrial Automation (DCA) School of Electrical and Computer Engineering (FEEC) State University of Campinas (Unicamp) C. P Campinas SP Brazil 2 Federal University of São Carlos (UFSCar) Sorocaba SP Brazil {pasquini,mauricio}@dca.fee.unicamp.br [email protected] Abstract. This paper introduces the use of an XOR-based flat routing mechanism for server-centric data centers (DCs) organized in 3-cube topologies. Essentially, the networking DC architectures available in the literature create a black box network unaware about the servers, increasing the complexity to deploy services inside DCs. The main reason for this black box is the lack of scalability when the servers are inserted in the forwarding tables of the switches. The server-centric 3-cube topology used in this work directly connects servers, removing the need for switches and/or routers in order to forward traffic inside the DC, approximating the servers to the networking infrastructure. The proposed XOR-based flat routing mechanism introduces a new semantic for routing, where a small information about the entire network allows traffic forwarding across the DC, providing the required scalability for inserting the servers in the routing tables. This paper presents the proposed 3-cube server-centric DC networking architecture, the XOR-based flat routing mechanism, and evaluations demonstrating the achieved scalability in terms of signaling, routing table entries, stretch and load distribution. 1. Introduction Nowadays, innumerable applications demanding massive storage, memory and processing support are executed in large-scale DCs deployed by companies such as Microsoft, Amazon, Google and Yahoo. To cope with this scale and lower total ownership costs, these DCs use custom software such as BigTable [Chang et al. 26], Google File System (GFS) [Ghemawat et al. 23], Map-Reduce [Dean and Ghemawat 28], and customized hardware encompassing servers [Cisco 21, Google 21], racks and power suppliers. However, the only component that has not really changed yet is the network, which usually leverages the current available Internet-related routing technologies. As detailed in [Esteve et al. 21], there is a set of requirements for the development of new DC architectures. For example, it is needed to consider how the DC communicates with external networks, such as the Internet, accepting external requests, homogeneously spreading these requests among the servers (normally using the Valiant
2 Load Balance (VLB) and/or Equal Cost Multi Path (ECMP)), and adopting hash-based mechanisms in order to preserve the communication sessions. In this way, there have been a number of proposals for redesigning DC networking architectures, many of them focusing on fat-tree topologies for scaling IP and/or Ethernet forwarding to large-scale DCs [Greenberg et al. 28, Greenberg et al. 29]. Nevertheless, the routing solutions used in these proposals are basically mapping services, in which the IPs and/or MAC addresses of servers are associated to the MAC addresses of ingress, intermediary and egress switches through the fat-tree topology. Consequently, this scenario achieves scalability by creating a black box totally unaware about the servers present in the DC, delegating the maintenance costs related to changes in the servers structure to the mapping service. Basically, servers and network infrastructure maintain an oblivious relationship, increasing the complexity to deploy services inside the DCs. This paper presents a server-centric DC architecture in which an XOR-based flat routing mechanism provides direct server-to-server communication at the layer 2, as an alternative to the scalability problems faced in other DC proposals. In the proposed scenario, a 3-cube topology is used to connect servers due to its intrinsic higher path redundancy, simplicity for wiring and fault resilience. The proposed approach, using the XOR-based flat routing mechanism in conjunction with the 3-cube topology, removes all the dependence on traditional switches and/or routers for server-to-server traffic forwarding inside the DC, closing the gap between servers and network infrastructure and, in this way, creating a server-aware DC for the deployment of services. Essentially, the proposed XOR-based flat routing mechanism, already instantiated in other scenarios [Pasquini et al. 29, Pasquini et al. 21a, Pasquini et al. 21b], uses a metric based on the bitwise XOR operation to organize the routing tables in columns (called buckets). The main contribution of the proposed XOR-based mechanism is the capability of performing routing with a small knowledge about the entire network, providing the required scalability for integrating servers with the DC networking infrastructure. Afterwards, the proposed XOR-based mechanism only requires uniqueness while assigning flat identifiers (IDs) to servers, being a totally random distribution of IDs ideal for its operation, i.e., no topological or semantical IDs organization is required. In this way, the IDs are assigned to serves inside the DC by simply selecting the smallest MAC address among the NICs present on each server, creating a plug-and-play environment for the configuration of new 3-cube-based DCs. Moreover, since it operates at layer 2, it transparently supports all the applications running on upper layers or inside virtual machines. According to it, this paper is focused on describing the proposed server-centric DC architecture, detailing the internal server-to-server communication using the XORbased flat routing mechanism at the layer 2. Consequently, other architectural details, such as external communication with the Internet, are kept out of the scope of this paper. The evaluations present in this work consider 3-cube topologies with 64, 128, 256, 512, 124 and 248 servers. All of them were evaluated using a developed emulation tool, where each server is executed as an independent thread, being able to exchange signaling messages to build the routing tables and forward traffic. The results detail the number of signaling messages, the number of entries in each routing table, the route stretch and the load distribution among servers while forwarding traffic inside the DC. Essentially, the results reveal a flexible and scalable routing mechanism where the number of signaling
3 messages and routing tables do not grow linearly with the size of the DC network, indicating the feasibility of using this proposal in large-scale DCs (ten of thousands of servers). Furthermore, it offers small stretch and adequate load distribution. The remainder of this paper is organized as follows. Section 2 focuses on presenting the main server-centric related work which uses cube-based topologies. Section 3 details the proposed 3-cube server-centric DC networking architecture. Section 4 formalizes the proposed XOR-based flat routing mechanism. Section 5 presents the evaluations focusing on the performance levels achieved by the XOR-based flat routing mechanism in the proposed DC architecture. Section 6 brings the next steps and concludes the paper. 2. Related Work The work presented in [Costa et al. 29] proposes the concept of server-to-server communication, eliminating the use of intermediary switches and routers. The rationale is that since the DC infrastructure is owned and controlled by a single entity, it is easy to create topologies optimized to obtain the best performance while forwarding traffic through the network. In such work, routing is done on the Network Interface Card (NIC) itself, with packets forwarded directly from server NIC to server NIC by using a multi-port network interface card. In fact, there is a service responsible for processing and forwarding the packets between servers. This routing on the NIC approach is possible through commercially available NICs, capable of running computing capabilities [NIC 21]. Another related work is presented in [Guo et al. 29]. Such proposal also uses a server-centric approach, in which servers are equipped with multiple network ports that connect to few commodity and low cost mini-switches. The idea is to provide multiple parallel paths in order to increase the bandwidth and improve fault-tolerance. Both works use the cube-based topology, which naturally offers a better path diversity in the network. By having such path diversity, the bisection bandwidth is increased, fault-tolerance is obtained, wiring complexity and the total cost are reduced, and load balance can be better exploited, potentially providing energy savings. The cube topology has been seen as an excellent alternative for attending the metrics mentioned above [Costa et al. 29, Guo et al. 29]. Although fat-trees topologies are commonly used in DC networks, they suffer from low resilience to failures and high wiring complexity [Costa et al. 29]. Therefore, our work applies the XOR-based flat routing in 3-cube topologies focused on the server-centric approach. 3. Proposed 3-cube server-centric DC networking architecture According to the networking architecture proposed in this paper, the DC is composed of a set of servers organized in a 3-cube topology, where all the servers are topologically distributed in three axis (x, y and z). In such server-centric scenario, servers are the fundamental elements in the DC, not requiring the use of other network elements, such as routers and/or switches to provide traffic forwarding during server-to-server communication. To this aim, each server comprises a general purpose multi-core processor, with memory, persistent storage and six NICs named from eth to eth5. In order to be settled in the 3-cube topology, each server uses the NICs eth and eth1 in the x axis, the NICs eth2 and eth3 in the y axis, and the NICs eth4 and eth5 in the z axis, as illustrated in Figure 1.
4 Figure 1. How to settle the NICs of server a in the 3-cube topology. Considering the use of the NICs in each one of the three axis, wiring the proposed DC network become a simple task as exemplified in Figure 2. In this figure, there are three servers (a, b and c) located in the same row of the DC along the x axis. Basically, the required connections in a given axis are established using the two interfaces assigned to it, also considering servers located in the edges as neighbors. According to these procedures, it is possible to check the desirable organization in Figure 2, where eth of server a is connected to eth1 of server b, eth of server b is connected to eth1 of server c, and eth of server c is connected to eth1 of server a, wrapping the x axis to establish this last link. The same pattern is used for wiring the y and z axis, where eth2 is connected to eth3 and eth4 is connected to eth5, respectively. Figure 2. Wiring servers a, b and c in the x axis. When compared to other topologies, the 3-cube topology not only offers simpler wiring and higher path redundancy, but also reduces to half the maximum distance (the radius) between servers, which is obtained by wrapping the links between the edge servers. In order to simplify the perception of the benefits of the links established between edge servers, Figure 3 presents a 3-cube topology composed of 27 servers. Note in this figure the existence of six NICs in all servers, the linkage between all the servers in the three axis, and the wrapping connections between the edge servers. Such wrapping connections raise the edge servers to a special condition in the DC, since they are able to invert traffic in the extremities of the axis. Essentially, in the server-centric 3-cube topology, there are three different levels of edge servers as seen in Figure 3. The levels are relative to the ability that each edge server has to invert traffic along the three axis used in the topology. In the first level are the edge servers located in the inner parts of the 3-cube surface, labeled as S servers in the figure. These servers are able to perform traffic inversion in a single axis. After, in the second level are the edge servers located in the borders (B servers) of the 3-cube. These servers are able to perform traffic inversion in two axis. Finally, in the third level are the eight edge servers located in the corners (C servers) of the 3-cube. These eight servers are able to perform traffic inversion in all the three axis. In this way, when compared to other topologies, the 3-cube topology also offers higher resilience. For example, the
5 Figure 3. Example of a 3-cube topology where the axis are x = 3, y = 3 and z = 3. DCell [Guo et al. 24] network can be partitioned even if less than 5% of the servers or links fail. The server-centric 3-cube topology is able to operate even if nearly to 5% of the servers or links fail, its symmetrical and redundant structure allows failures in some regions of the topology, without impacting the performance of the remaining servers. Regarding the DC configuration, there are available in the literature several techniques to bootstrap the configuration of all elements comprising the DC, such as servers, routers and switches. For example, in Portland [Mysore et al. 29] it is used a hierarchical addressing scheme to assign Pseudo MACs to all servers according to their position in the DC. Afterwards, a Location Discovery Protocol (LDP) is used to help switches composing the fat tree topology to realize their role in the network, whether they are edge, aggregation or core switches. In this work, due to the proposed XOR-based flat routing mechanism, uniqueness is the only requirement to assign flat identifiers to servers. Despite uniqueness, none topological adherence or semantic organization is required for the IDs. Actually, a total random distribution of IDs creates the ideal scenario for routing using the XORbased mechanism. In this way, each server selects the smallest MAC address (among its six NICs) to be its 48-bits flat ID. Although such mechanism is simplistic, it suffices to achieve the routing mechanism requirements, and avoids the use of human intervention and/or bootstrap mechanism in order to configure the servers, i.e., all the required configuration is plug-and-play. Furthermore, switches and routers are not required to provide server-to-server communication, avoiding the use of special protocols to configure these equipments. However, if such equipments are used, they are intended to just offer a network bus connecting servers. In order to perform server-to-server traffic forwarding, this proposal adopts MAC-
6 in-mac encapsulation, due to the fact that the proposed flat routing mechanism is designed to operate using MACs as flat IDs of all servers inside the DC. Besides the legacy support to all applications running on upper layers of the network stack or inside virtual machines, the option for performing routing at the layer 2 in conjunction with the MACin-MAC technology has the advantage of fixing in two the number of packet headers from source to destination. Such characteristic simplifies the definition of the maximum payload size and avoids segmentation of packets on the path traversed inside the DC. Figure 4 exemplifies the traffic forwarding from server a towards server e. In this figure, it is possible to verify that the inner header carries the server-to-server information (SRC Flat ID and DST Flat ID). This header is preserved during the entire packet transmission and, essentially, it is used by all the servers on the path to perform the XOR-based routing (detailed in Section 4). On the other hand, the outer header has only local link scope, being changed in all the links from source to destination. This header carries the MAC address of the server currently forwarding the packet, and the MAC address of the next server (next hop) present on the path towards the destination, i.e., it is used to transfer traffic between two adjacent servers. Figure 4. Traffic forwarding from server a to server e using MAC-in-MAC. Finally, although it is kept out of the scope of this work, the proposed DC networking architecture predicts that among all the servers present in the DC, a fraction of the servers has one extra NIC to forward traffic in and out of the DC. For example, one possible solution for placing these nodes considers the edge servers, once they are able to invert traffic along the axis of the 3-cube topology. In this case, the designer of the DC can start by using servers located in the corners (C servers - 3 axis inversion), after servers located in the borders (B servers - 2 axis inversion) and, finally, servers located in the surface (S servers - 1 axis inversion) of the 3-cube topology. 4. XOR-based Flat Routing Mechanism The proposed XOR-based routing discards the use of mapping services required in other proposals found in the literature [Greenberg et al. 28, Greenberg et al. 29, Mysore et al. 29]. Usually, these mapping services provide source routing and/or tunneling information required to forward packets inside the DC network, characterizing an oblivious relation between servers and network structure. The routing mechanism proposed uses n-bit flat identifiers to organize the routing tables in n columns and route packets through the network. Its routing principle uses the distance between two flat identifiers a and b as their bitwise exclusive or (XOR), which is represented by d(a, b) = a b, being d(a, a) = and d(a, b) >, a, b. Given a packet originated by server x and destined to server z, and denoting Y as the set of identifiers
7 contained on x s routing table, the XOR-based routing mechanism applied at server x selects the server y Y that minimizes the distance towards z, which is expressed by the following routing policy R = argmin{d(y, z)}. (1) y Y Each server maintains a routing table in which its knowledge about neighbor servers is spread in n columns denominated buckets and represented by β i, i n 1. Table 1 presents an example of a routing table for server 1 in an identity space where n = 4. Each time a server a knows a novel neighbor b, it stores that information in the bucket β n 1 i given the highest i that satisfies the following condition 1 d(a, b) div 2 i = 1, a b, i n 1. (2) For example, consider a = 1 and b = 1. The distance d(a, b) = 11 and the highest i that satisfies condition (2) is i = 1, concluding that the identifier b = 1 must be stored in the bucket β n 1 i = β 2. Basically, condition (2) denotes that server a stores server b in the bucket β n 1 i, where n 1 i is the length of the longest common prefix (lcp) existent between both identifiers of servers a and b. This can be observed in Table 1, where the buckets β, β 1, β 2, β 3 store the identifiers having lcp of length, 1, 2, 3 with server 1. Such routing tables approach is one of the main advantages of the XOR-based mechanism, since a server only needs to know n (1 neighbor per bucket) of the possible 2 n servers available in the network to successfully route packets. Table 1. Hypothetic routing table for server 1 in an identity space where n = 4. β β 1 β 2 β The proposed mechanism considers a K factor which defines the minimum amount of information required per bucket. Since a bucket β i has at most 2 n 1 i entries, if K > 2 n 1 i we limit K for that bucket to K = 2 n 1 i. By varying K, servers have wider or narrower view of the DC network. In this way, during the process for building the routing tables, servers are aimed at finding K neighbors for each bucket. Essentially, a server a can know a server b from two distinct ways in the proposed process for building the routing tables: 1) a discovery process in which servers actively search for neighbors to fill the buckets and 2) a learning process, where servers use the information contained in the signaling messages that cross them to add more information to their buckets in a costless passive fashion Discovery Process In the discovery process a HELLO message is used to insert servers directly (physically) connected in the routing tables, such servers are denominated as physical neighbors. Each 1 div denotes the integer division operation on integers.
8 discovered physical neighbor is then stored in the bucket having the greatest i that solves condition (2). After introducing all the physical neighbors in the routing table, a server that still has buckets with less information than the defined K factor, actively searches for neighbors that can fill such buckets. Since such neighbors are not physically connected, they are called virtual neighbors. In this way, servers start to send QUERY messages to the neighbors already stored in the buckets (in the initial step the buckets only contain physical neighbors), describing the buckets which still require information. The neighbor servers that receive the QUERY message and have at least one neighbor that fills in one of the requested buckets, answer with a RESPONSE message. After receiving the RESPONSE message, servers store the discovered virtual neighbors in the respective buckets. New queries are sent to the discovered neighbors if at least one of the buckets is still requiring information. The following information are associated to each entry of the routing table: 1) the physical neighbor (next hop) towards the discovered neighbor, 2) the NIC used to establish the link with the physical neighbor (next hop) and 3) the physical distance in number of hops to the discovered neighbor. Afterwards, if a server a discovers a server b through different NICs, it can store all the information for path diversity purposes Learning Process Figure 5 is used to explain the learning process. In this figure, server 1 sends a QUERY message to server 2, and server 2 sends back a RESPONSE message informing server 1 about the existence of server 3. Assuming that server 1 still needs information to satisfy the K factor in some of its buckets, it sends a QUERY message to server 3 just discovered, and server 3 sends back a RESPONSE informing server 1 about the existence of servers 4 and 5 (considering that both servers 4 and 5 satisfy what server 1 is looking for). Figure 5. Exemplification scenario for the Learning Process. The learning process takes place in two distinct moments. Firstly, it occurs when server 3 receives the QUERY message from server 1. At that instant, server 3 knows server 1 in a passive way, using the source ID present in the QUERY message. Secondly, the learning process occurs when server 3 sends the RESPONSE message to server 1 containing information about servers 4 and 5. In this case, the RESPONSE message crosses server 2 which, also passively, learns about both servers based on the information contained in the RESPONSE message destined to server 1. Essentially, the discovery and learning processes were designed to assure symmetry between the routing tables of servers in the DC network, as defined in Property 1. Afterwards, the process for building the routing tables is designed to prioritize the discovery of neighbors physically near (in
9 number of hops), contributing in this way to the convergence of the routing tables, and reducing the required number of signaling messages. Property 1 If a server a has a server b in its routing table, then server b has server a in its own routing table Routing Process Each time a server receives/generates a packet to forward, it tries to route the packet by applying the routing policy R defined in (1). The routing process starts by identifying the bucket to be used. Considering that a server a is routing a packet destined to b, as exemplified in Figure 6, it defines the bucket β i, where the index i is the highest one that solves condition (2). Then server a routes the packet to the server available in its routing table that is closest to server b. In other words, the packet is routed to n R, which is selected by server a computing the following solution n R = argmin id β i {d(id, b)}, (3) where id represents each one of the identifiers contained in the bucket β i. Since n R is the identifier of a server that can be a virtual neighbor j hops away (j > 1), server a uses the information associated to n R in its routing table entry to forward the packet. Essentially, the information associated is the physical next hop g and the NIC linking both servers (previously detailed in Section 4.1), which are resultant of the discovery and learning processes for building the routing tables. Figure 6. Example of the XOR-based routing process. When the packet arrives at g, it also computes (3) and, again, due to the process for building the routing tables, it finds n R in its bucket β i, delivering the packet to n R. Finally, n R is the server which represents progress towards the destination server b in the flat identity space, i.e., the server pointed by the previous routing decision, firstly taken at source server a, which reduces the distance towards server b according to the routing police R defined in (1). In this way, based on the example of Figure 6, server n R computes (3) and finds the destination server b in its routing table, towards its NIC connected to the physical neighbor h. Consequently, the remaining actions are responsible for delivering the packet to server b passing through server h. 5. Evaluations This section presents the evaluations of the proposed XOR-based flat routing mechanism in the proposed 3-cube server-centric DC networking architecture. The main objective of this section is to demonstrate, through evaluations using a developed emulation tool, the level of scalability achieved by the XOR-based routing mechanism in terms of number of entries in the routing tables. Such information is essential to demonstrate the feasibility
10 of integrating servers and networking, creating a server-aware DC architecture for the deployment of services, opposing the oblivious approach found in the literature. This section also details the required signaling to converge the flat routing mechanism, presents the route stretch incurred in such scenario of reduced routing tables, and depicts the load distribution among all the servers during traffic forwarding inside the proposed 3-cube DC network. These evaluations were performed using six networks of different sizes 64, 128, 256, 512, 124 and 248 servers using the K factor set to 1, 3, 5 and 1, resulting in more than 22 million computed paths, i.e., it was computed the full combination of source/destination paths. The developed emulation tool has an important contribution in the evaluations, since it offers the possibility of emulating individual servers provided with a full implementation of the XOR-based flat routing mechanism. Basically, each server inside the 3-cube DC is instantiated as an individual thread, being able to perform all the required interactions between the servers, exchanging signaling messages to build the routing tables according to the protocol specification. Afterwards, the threads (the emulated servers) are able to forward traffic inside the 3-cube DC, contributing with the analyzes related to route stretch and load distribution. The tool also includes a topology generator capable of creating 3-cube topologies, considering all the required links and assuring the occurrence of uniqueness and randomness while assigning the flat IDs. The first results are presented in Figure 7, where it is possible to find information regarding the signaling complexity required to converge the XOR-based routing tables. Figure 7(a) describes the average number of signaling messages generated per server in all the six evaluated topologies. This number represents the exchange of QUERY and RESPONSE messages from server-to-server during the discovery process, and shows that as the size of the topologies and the value of the K factor increase, the number of signaling messages also increases. Messages (#) K=1 K=5 K=3 K= Number of servers (a) Number of messages exchanged. Interaction (%) K=1 K=5 K=3 K= Number of servers (b) Percentage of interaction. Figure 7. Results regarding the average signaling exchanged per server. However, analyzing the percentage of interaction between the servers present in the six topologies, it is possible to verify that the percentage of interaction between the servers does not increase linearly with the size of the topology, as shown in Figure 7(b). This behavior has origin in the proposed discovery and learning processes designed to extract the maximum information from the signaling, prioritizing the exchange of messages with servers located physically near to the requesting server.
11 The next result detailed in Figure 8 presents the average number of entries required in the routing tables of all servers. Similar to the results obtained for the signaling messages, the bigger the network and the K factor used, the higher the number of entries present in the routing tables, as seen in Figure 8(a). However, the percentage of routing information required in the routing tables also decreases as the network topologies increase, as presented in Figure 8(b). The main reason for this behavior is the proposed XOR-based routing tables which require only a small fraction of the entire routing information available in the network, in order to assure traffic delivery. Entries (#) K=1 K=5 K=3 K=1 Entries (%) K=1 K=5 K=3 K= Number of servers (a) Number of entries in the routing tables Number of servers (b) Percentage of network knowledge. Figure 8. Results regarding the average routing tables per server. The results presented in Figure 8 not only demonstrate that the proposed flat routing mechanism is able to operate with a fraction of the overall routing information, offering better control for the rate at which the routing tables grow, but also prove its ability for integrating servers and network structure in a scalable manner, offering the required infrastructure for inserting information regarding the servers composing the DC in the routing tables. Such condition is ideal for the development of new services totally integrated with the available DC network structure. Normally, based on the small number of entries present in the routing tables, it is expected that the proposed XOR-based flat routing mechanism does not find the shortest paths for all the server-to-server communication cases. Basically, route stretch is a tradeoff of the number of entries present in the routing tables. Nonetheless, Figure 9 details the obtained stretch values, achieving values near to optimal (stretch 1). For example, considering the results obtained for the topology with 248 servers using K = 1, the servers exchanged approximately 4 signaling messages (see Figure 7(a)), corresponding to the interaction with 2% of the servers, in order to create routing tables with approximately 1 entries (see Figure 8(a)). Such number of entries represents a knowledge of 5% of the entire network, and even with this reduced number of entries the protocol is able to deliver traffic with an average route stretch around 1.85, not even doubling the length of the shortest paths available. In the sequence, according to the basic trade-off mentioned, in the results obtained for the topology with 248 servers using K = 1, it is possible to check the bigger routing tables, containing approximately 25% of the entire routing information, and the reduced stretch at the level of Finally, the next results demonstrate the load distribution among all the servers. Figures 1(a), 1(b), 1(c) and 1(d) detail the load for all the evaluated topologies using
12 Stretch (#) K=1 K=5 K=3 K= Number of servers Figure 9. Results for route stretch. the K factor set to 1, 3, 5 and 1, respectively. Essentially, the results indicate that approximately 8% of all servers, independent of the number of servers in the DC and the K factor used, operate at a load level ranging from.8 to 1.2, i.e., most of the servers deviate only 2% from the average traffic forwarding load. Firstly, such numbers indicate the robustness of the 3-cube topology, a topology capable of providing an elevated path redundancy ideal not only for load distribution, but also for fault resilience. Afterwards, such numbers prove the perfect coupling between the 3-cube topology and the proposed XOR-based flat routing mechanism, indicating that the plug-and-play scenario of totally random distributed flat IDs is ideal to spread the forwarding load among all the servers, not even requiring the usage of other load balancing mechanism, such as VLB or ECMP. 6. Conclusion and Future Work This paper presented a new server-centric DC architecture where servers are organized in a 3-cube topology. In this work, servers are the main elements of the DC, eliminating the need for traditional network equipments such as switches and routers in order to provide server-to-server communication. The main novelty is the integration between servers and networking, which is done by the scalable XOR-based flat routing mechanism. Among the benefits of the 3-cube topology are the higher path redundancy and the simpler wiring. The establishment of wrapped links between the edge servers significantly reduce the maximum distance between two servers, and increases the resilience of the DC. Essentially, the edge servers are able to invert traffic in the three axis, assuring that the DC operates even if approximately 5% of servers and/or links fail. The 3-cube topology in conjunction with the XOR-based flat routing mechanism creates a plug-and-play scenario, efficient for load distribution among all the servers in the DC, independent of the size of the topology and of the K factor used. Once the proposed routing mechanism requires only a fraction of the entire routing information, it provides the total integration between servers and network structure, eliminating the concept of a black box network. As presented in the evaluations in this paper, as the DC network grows in number of servers, the required signaling and the number of entries in the routing tables do not linearly follow such growth. At the same time, although the routing tables have just a small percentage of information, the paths from source to destination present a small stretch penalization. As future work, one objective is to investigate the connectivity between 3-cubes.
13 Percentile rank Percentile rank Normalized server utilization (a) Load distribution for K = Normalized server utilization (c) Load distribution for K = 5. Percentile rank Percentile rank Normalized server utilization (b) Load distribution for K = Normalized server utilization (d) Load distribution for K = 1. Figure 1. Results for load distribution among all servers. Nowadays, it is expected that enormous DCs (tens of thousands of servers) are modularized, simplifying their deployment and maintenance. For example, it is possible to ship these DCs in containers, simplifying their transport and reducing cots regarding energy and cooling systems, as proposed in [Guo et al. 29, Wu et al. 29]. The idea is to use some specific servers for inter-3-cube communication through fiber channels. Acknowledgment The authors would like to thank Ericsson Brazil and CNPq for supporting this work. References Chang, F., Dean, J., Ghemawat, S., Hsieh, W. C., Wallach, D. A., Burrows, M., Chandra, T., Fikes, A., and Gruber, R. E. (26). BigTable: A Distributed Storage System for Structured Data. In Proceedings of OSDI 26, Seattle, WA, USA. Cisco (21). Cisco Unified Computing System. Available at com/go/unifiedcomputing. Costa, P., Zahn, T., Rowstron, A., O Shea, G., and Schubert, S. (29). Why Should we Integrate Services, Servers, and Networking in a Data Center? In Proceedings of the 1st ACM Workshop on Research on Enterprise Networking (WREN 9), pages , New York, NY, USA. ACM.
14 Dean, J. and Ghemawat, S. (28). DCell: A Scalable and Fault-Tolerant Network Structure for Data Centers. In Proceedings of ACM SIGCOMM 28, Seattle, WA, USA. Esteve, C., Pasquini, R., Verdi, F. L., and Magalhães, M. F. (21). Novas Arquiteturas de Data Center para Cloud Computing. Book Chapter published as a Mini Course of the 28th Brazilian Symposium on Computer Networks and Distributed Systems (SBRC 21), Organized by: C. A. Kamienski; L. P. Gaspary; M. P. Barcellos. ISBN: Idiom: Portuguese. Gramado, RS, Brazil. Ghemawat, S., Gobioff, H., and Leung, S. T. (23). The Google File System. In Proceedings of SOSP 23, Bolton Landing (Lake George), New York, USA. Google (21). Google s Custem Web Server, Revealed. Available at tinyurl.com/google-custom-server. Greenberg, A., Hamilton, J. R., Jain, N., Kandula, S., Kim, C., Lahiri, P., Maltz, D. A., Patel, P., and Sengupta, S. (29). VL2: A Scalable and Flexible Data Center Network. In Proceedings of the ACM SIGCOMM 29 - Conference on Data Communication, Barcelona, Spain. Greenberg, A., Lahiri, P., Maltz, D. A., Patel, P., and Sengupta, S. (28). Towards a Next Generation Data Center Architecture: Scalability and Commoditization. In Proceedings of the ACM Workshop on Programmable Routers For Extensible Services of Tomorrow, Seattle, WA, USA. Guo, C., Lu, G., Li, D., Wu, H., Zhang, X., Shi, Y., Tian, C., Zhang, Y., and Lu, S. (29). BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers. In Proceedings of the ACM SIGCOMM 29 - Conference on Data Communication, Barcelona, Spain. Guo, C., Wu, H., Tan, K., Shiy, L., Zhang, Y., and Luz, S. (24). MapReduce: Simplified Data Processing on Large Clusters. In Proc. of OSDI 24, San Francisco, CA, USA. Mysore, R. N., Pamboris, A., Farrington, N., Huang, N., Miri, P., Radhakrishnan, S., Subramanya, V., and Vahdat, A. (29). PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. In Proceedings of the ACM SIGCOMM 29 - Conference on Data Communication, Barcelona, Spain. NIC, K. (21). Killer NIC. Available at Pasquini, R., Verdi, F. L., and Magalhães, M. F. (29). Towards a Landmark-based Flat Routing. 27th Brazilian Symposium on Computer Networks and Distributed Systems (SBRC 29), Recife, PE, Brazil. Pasquini, R., Verdi, F. L., Magalhães, M. F., and Welin, A. (21a). Bloom filters in a Landmark-based Flat Routing. In Proc. of IEEE ICC 21, Cape Town, South Africa. Pasquini, R., Verdi, F. L., Oliveira, R., Magalhães, M. F., and Welin, A. (21b). A Proposal for an XOR-based Flat Routing Mechanism in Internet-like Topologies. In Proceedings of IEEE GLOBECOM 21, Miami, FL, USA. Wu, H., Lu, G., Li, D., Guo, C., and Zhang, Y. (29). MDCube: A High Performance Network Structure for Modular Data Center Interconnection. In Proceedings of the ACM CONEXT 29, Rome, Italy.
Load Balancing Mechanisms in Data Center Networks
Load Balancing Mechanisms in Data Center Networks Santosh Mahapatra Xin Yuan Department of Computer Science, Florida State University, Tallahassee, FL 33 {mahapatr,xyuan}@cs.fsu.edu Abstract We consider
Data Center Network Topologies: FatTree
Data Center Network Topologies: FatTree Hakim Weatherspoon Assistant Professor, Dept of Computer Science CS 5413: High Performance Systems and Networking September 22, 2014 Slides used and adapted judiciously
Scafida: A Scale-Free Network Inspired Data Center Architecture
Scafida: A Scale-Free Network Inspired Data Center Architecture László Gyarmati, Tuan Anh Trinh Network Economics Group Department of Telecommunications and Media Informatics Budapest University of Technology
A Reliability Analysis of Datacenter Topologies
A Reliability Analysis of Datacenter Topologies Rodrigo S. Couto, Miguel Elias M. Campista, and Luís Henrique M. K. Costa Universidade Federal do Rio de Janeiro - PEE/COPPE/GTA - DEL/POLI Email:{souza,miguel,luish}@gta.ufrj.br
Data Center Network Architectures
Servers Servers Servers Data Center Network Architectures Juha Salo Aalto University School of Science and Technology [email protected] Abstract Data centers have become increasingly essential part of
Advanced Computer Networks. Datacenter Network Fabric
Advanced Computer Networks 263 3501 00 Datacenter Network Fabric Patrick Stuedi Spring Semester 2014 Oriana Riva, Department of Computer Science ETH Zürich 1 Outline Last week Today Supercomputer networking
Evaluating the Impact of Data Center Network Architectures on Application Performance in Virtualized Environments
Evaluating the Impact of Data Center Network Architectures on Application Performance in Virtualized Environments Yueping Zhang NEC Labs America, Inc. Princeton, NJ 854, USA Email: [email protected]
PortLand:! A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric
PortLand:! A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya,
TRILL for Data Center Networks
24.05.13 TRILL for Data Center Networks www.huawei.com enterprise.huawei.com Davis Wu Deputy Director of Switzerland Enterprise Group E-mail: [email protected] Tel: 0041-798658759 Agenda 1 TRILL Overview
AIN: A Blueprint for an All-IP Data Center Network
AIN: A Blueprint for an All-IP Data Center Network Vasileios Pappas Hani Jamjoom Dan Williams IBM T. J. Watson Research Center, Yorktown Heights, NY Abstract With both Ethernet and IP powering Data Center
Enabling Flow-based Routing Control in Data Center Networks using Probe and ECMP
IEEE INFOCOM 2011 Workshop on Cloud Computing Enabling Flow-based Routing Control in Data Center Networks using Probe and ECMP Kang Xi, Yulei Liu and H. Jonathan Chao Polytechnic Institute of New York
International Journal of Emerging Technology in Computer Science & Electronics (IJETCSE) ISSN: 0976-1353 Volume 8 Issue 1 APRIL 2014.
IMPROVING LINK UTILIZATION IN DATA CENTER NETWORK USING NEAR OPTIMAL TRAFFIC ENGINEERING TECHNIQUES L. Priyadharshini 1, S. Rajanarayanan, M.E (Ph.D) 2 1 Final Year M.E-CSE, 2 Assistant Professor 1&2 Selvam
A Comparative Study of Data Center Network Architectures
A Comparative Study of Data Center Network Architectures Kashif Bilal Fargo, ND 58108, USA [email protected] Limin Zhang Fargo, ND 58108, USA [email protected] Nasro Min-Allah COMSATS Institute
OVERLAYING VIRTUALIZED LAYER 2 NETWORKS OVER LAYER 3 NETWORKS
OVERLAYING VIRTUALIZED LAYER 2 NETWORKS OVER LAYER 3 NETWORKS Matt Eclavea ([email protected]) Senior Solutions Architect, Brocade Communications Inc. Jim Allen ([email protected]) Senior Architect, Limelight
OpenFlow based Load Balancing for Fat-Tree Networks with Multipath Support
OpenFlow based Load Balancing for Fat-Tree Networks with Multipath Support Yu Li and Deng Pan Florida International University Miami, FL Abstract Data center networks are designed for satisfying the data
Scaling 10Gb/s Clustering at Wire-Speed
Scaling 10Gb/s Clustering at Wire-Speed InfiniBand offers cost-effective wire-speed scaling with deterministic performance Mellanox Technologies Inc. 2900 Stender Way, Santa Clara, CA 95054 Tel: 408-970-3400
Chapter 6. Paper Study: Data Center Networking
Chapter 6 Paper Study: Data Center Networking 1 Data Center Networks Major theme: What are new networking issues posed by large-scale data centers? Network Architecture? Topology design? Addressing? Routing?
BURSTING DATA BETWEEN DATA CENTERS CASE FOR TRANSPORT SDN
BURSTING DATA BETWEEN DATA CENTERS CASE FOR TRANSPORT SDN Abhinava Sadasivarao, Sharfuddin Syed, Ping Pan, Chris Liou (Infinera) Inder Monga, Andrew Lake, Chin Guok Energy Sciences Network (ESnet) IEEE
Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya and Amin Vahdat
Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya and Amin Vahdat 1 PortLand In A Nutshell PortLand is a single logical
Generalized DCell Structure for Load-Balanced Data Center Networks
Generalized DCell Structure for Load-Balanced Data Center Networks Markus Kliegl, Jason Lee,JunLi, Xinchao Zhang, Chuanxiong Guo,DavidRincón Swarthmore College, Duke University, Fudan University, Shanghai
Portland: how to use the topology feature of the datacenter network to scale routing and forwarding
LECTURE 15: DATACENTER NETWORK: TOPOLOGY AND ROUTING Xiaowei Yang 1 OVERVIEW Portland: how to use the topology feature of the datacenter network to scale routing and forwarding ElasticTree: topology control
Data Center Network Topologies: VL2 (Virtual Layer 2)
Data Center Network Topologies: VL2 (Virtual Layer 2) Hakim Weatherspoon Assistant Professor, Dept of Computer cience C 5413: High Performance ystems and Networking eptember 26, 2014 lides used and adapted
A Hybrid Electrical and Optical Networking Topology of Data Center for Big Data Network
ASEE 2014 Zone I Conference, April 3-5, 2014, University of Bridgeport, Bridgpeort, CT, USA A Hybrid Electrical and Optical Networking Topology of Data Center for Big Data Network Mohammad Naimur Rahman
Experimental Framework for Mobile Cloud Computing System
Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 00 (2015) 000 000 www.elsevier.com/locate/procedia First International Workshop on Mobile Cloud Computing Systems, Management,
System Interconnect Architectures. Goals and Analysis. Network Properties and Routing. Terminology - 2. Terminology - 1
System Interconnect Architectures CSCI 8150 Advanced Computer Architecture Hwang, Chapter 2 Program and Network Properties 2.4 System Interconnect Architectures Direct networks for static connections Indirect
What is Analytic Infrastructure and Why Should You Care?
What is Analytic Infrastructure and Why Should You Care? Robert L Grossman University of Illinois at Chicago and Open Data Group [email protected] ABSTRACT We define analytic infrastructure to be the services,
Wireless Link Scheduling for Data Center Networks
Wireless Link Scheduling for Data Center Networks Yong Cui Tsinghua University Beijing, 10084, P.R.China [email protected] Hongyi Wang Tsinghua University Beijing, 10084, P.R.China wanghongyi09@mails.
Data Center Networking Designing Today s Data Center
Data Center Networking Designing Today s Data Center There is nothing more important than our customers. Data Center Networking Designing Today s Data Center Executive Summary Demand for application availability
Extending the Internet of Things to IPv6 with Software Defined Networking
Extending the Internet of Things to IPv6 with Software Defined Networking Abstract [WHITE PAPER] Pedro Martinez-Julia, Antonio F. Skarmeta {pedromj,skarmeta}@um.es The flexibility and general programmability
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family
Intel Ethernet Switch Load Balancing System Design Using Advanced Features in Intel Ethernet Switch Family White Paper June, 2008 Legal INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL
TRILL Large Layer 2 Network Solution
TRILL Large Layer 2 Network Solution Contents 1 Network Architecture Requirements of Data Centers in the Cloud Computing Era... 3 2 TRILL Characteristics... 5 3 Huawei TRILL-based Large Layer 2 Network
Data Center Network Structure using Hybrid Optoelectronic Routers
Data Center Network Structure using Hybrid Optoelectronic Routers Yuichi Ohsita, and Masayuki Murata Graduate School of Information Science and Technology, Osaka University Osaka, Japan {y-ohsita, murata}@ist.osaka-u.ac.jp
Resolving Packet Loss in a Computer Centre Applications
International Journal of Computer Applications (975 8887) olume 74 No., July 3 Resolving Packet Loss in a Computer Centre Applications M. Rajalakshmi C.Angel K. M. Brindha Shree ABSTRACT The modern data
Interconnection Networks. Interconnection Networks. Interconnection networks are used everywhere!
Interconnection Networks Interconnection Networks Interconnection networks are used everywhere! Supercomputers connecting the processors Routers connecting the ports can consider a router as a parallel
MPLS-TP. Future Ready. Today. Introduction. Connection Oriented Transport
MPLS-TP Future Ready. Today Introduction As data traffic started dominating telecom networks, there was a need for transport data networks, as opposed to transport TDM networks. Traditional transport technologies
Broadband Networks. Prof. Karandikar. Department of Electrical Engineering. Indian Institute of Technology, Bombay. Lecture - 26
Broadband Networks Prof. Karandikar Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 26 Optical Network &MPLS So, as you were discussing in the previous lectures, next
Ethernet Fabrics: An Architecture for Cloud Networking
WHITE PAPER www.brocade.com Data Center Ethernet Fabrics: An Architecture for Cloud Networking As data centers evolve to a world where information and applications can move anywhere in the cloud, classic
Secure Cloud Computing with a Virtualized Network Infrastructure
Secure Cloud Computing with a Virtualized Network Infrastructure Fang Hao, T.V. Lakshman, Sarit Mukherjee, Haoyu Song Bell Labs, Alcatel-Lucent {firstname.lastname}@alcatel-lucent.com ABSTRACT Despite
VXLAN: Scaling Data Center Capacity. White Paper
VXLAN: Scaling Data Center Capacity White Paper Virtual Extensible LAN (VXLAN) Overview This document provides an overview of how VXLAN works. It also provides criteria to help determine when and where
Avoiding Network Polarization and Increasing Visibility in Cloud Networks Using Broadcom Smart- Hash Technology
Avoiding Network Polarization and Increasing Visibility in Cloud Networks Using Broadcom Smart- Hash Technology Sujal Das Product Marketing Director Network Switching Karthik Mandakolathur Sr Product Line
Layer-3 Multipathing in Commodity-based Data Center Networks
Layer-3 Multipathing in Commodity-based Data Center Networks Ryo Nakamura University of Tokyo Email: [email protected] Yuji Sekiya University of Tokyo Email: [email protected] Hiroshi Esaki University of
SummitStack in the Data Center
SummitStack in the Data Center Abstract: This white paper describes the challenges in the virtualized server environment and the solution Extreme Networks offers a highly virtualized, centrally manageable
TeachCloud: A Cloud Computing Educational Toolkit
TeachCloud: A Cloud Computing Educational Toolkit Y. Jararweh* and Z. Alshara Department of Computer Science, Jordan University of Science and Technology, Jordan E-mail:[email protected] * Corresponding
Paolo Costa [email protected]
joint work with Ant Rowstron, Austin Donnelly, and Greg O Shea (MSR Cambridge) Hussam Abu-Libdeh, Simon Schubert (Interns) Paolo Costa [email protected] Paolo Costa CamCube - Rethinking the Data Center
Non-blocking Switching in the Cloud Computing Era
Non-blocking Switching in the Cloud Computing Era Contents 1 Foreword... 3 2 Networks Must Go With the Flow in the Cloud Computing Era... 3 3 Fat-tree Architecture Achieves a Non-blocking Data Center Network...
Virtual PortChannels: Building Networks without Spanning Tree Protocol
. White Paper Virtual PortChannels: Building Networks without Spanning Tree Protocol What You Will Learn This document provides an in-depth look at Cisco's virtual PortChannel (vpc) technology, as developed
MPLS Layer 2 VPNs Functional and Performance Testing Sample Test Plans
MPLS Layer 2 VPNs Functional and Performance Testing Sample Test Plans Contents Overview 1 1. L2 VPN Padding Verification Test 1 1.1 Objective 1 1.2 Setup 1 1.3 Input Parameters 2 1.4 Methodology 2 1.5
Cloud Networking Disruption with Software Defined Network Virtualization. Ali Khayam
Cloud Networking Disruption with Software Defined Network Virtualization Ali Khayam In the next one hour Let s discuss two disruptive new paradigms in the world of networking: Network Virtualization Software
M.Sc. IT Semester III VIRTUALIZATION QUESTION BANK 2014 2015 Unit 1 1. What is virtualization? Explain the five stage virtualization process. 2.
M.Sc. IT Semester III VIRTUALIZATION QUESTION BANK 2014 2015 Unit 1 1. What is virtualization? Explain the five stage virtualization process. 2. What are the different types of virtualization? Explain
On Tackling Virtual Data Center Embedding Problem
On Tackling Virtual Data Center Embedding Problem Md Golam Rabbani, Rafael Esteves, Maxim Podlesny, Gwendal Simon Lisandro Zambenedetti Granville, Raouf Boutaba D.R. Cheriton School of Computer Science,
Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs
Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs As a head of the campus network department in the Deanship of Information Technology at King Abdulaziz University for more
Large-Scale Distributed Systems. Datacenter Networks. COMP6511A Spring 2014 HKUST. Lin Gu [email protected]
Large-Scale Distributed Systems Datacenter Networks COMP6511A Spring 2014 HKUST Lin Gu [email protected] Datacenter Networking Major Components of a Datacenter Computing hardware (equipment racks) Power supply
SummitStack in the Data Center
SummitStack in the Data Center Abstract: This white paper describes the challenges in the virtualized server environment and the solution that Extreme Networks offers a highly virtualized, centrally manageable
Juniper Networks QFabric: Scaling for the Modern Data Center
Juniper Networks QFabric: Scaling for the Modern Data Center Executive Summary The modern data center has undergone a series of changes that have significantly impacted business operations. Applications
SAN Conceptual and Design Basics
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer
APPLICATION NOTE 210 PROVIDER BACKBONE BRIDGE WITH TRAFFIC ENGINEERING: A CARRIER ETHERNET TECHNOLOGY OVERVIEW
PROVIDER BACKBONE BRIDGE WITH TRAFFIC ENGINEERING: A CARRIER ETHERNET TECHNOLOGY OVERVIEW By Thierno Diallo, Product Specialist Originally designed as a local-area network (LAN) communication protocol,
VXLAN Bridging & Routing
VXLAN Bridging & Routing Darrin Machay [email protected] CHI-NOG 05 May 2015 1 VXLAN VM-1 10.10.10.1/24 Subnet A ESX host Subnet B ESX host VM-2 VM-3 VM-4 20.20.20.1/24 10.10.10.2/24 20.20.20.2/24 Load
Rohde & Schwarz R&S SITLine ETH VLAN Encryption Device Functionality & Performance Tests
Rohde & Schwarz R&S Encryption Device Functionality & Performance Tests Introduction Following to our test of the Rohde & Schwarz ETH encryption device in April 28 the European Advanced Networking Test
Fibre Channel over Ethernet in the Data Center: An Introduction
Fibre Channel over Ethernet in the Data Center: An Introduction Introduction Fibre Channel over Ethernet (FCoE) is a newly proposed standard that is being developed by INCITS T11. The FCoE protocol specification
Deterministic Ethernet and the Intrinsic Value of Synchronization
Deterministic Ethernet and the Intrinsic Value of Synchronization Norman Finn Cisco Fellow WSTS June 2014 NIST-ATIS WSTS San Jose CA June 2014 1 NIST-ATIS WSTS San Jose CA June 2014 2 IEEE 1588 technology
Chapter 1 Reading Organizer
Chapter 1 Reading Organizer After completion of this chapter, you should be able to: Describe convergence of data, voice and video in the context of switched networks Describe a switched network in a small
Lecture 7: Data Center Networks"
Lecture 7: Data Center Networks" CSE 222A: Computer Communication Networks Alex C. Snoeren Thanks: Nick Feamster Lecture 7 Overview" Project discussion Data Centers overview Fat Tree paper discussion CSE
PCube: Improving Power Efficiency in Data Center Networks
PCube: Improving Power Efficiency in Data Center Networks Lei Huang, Qin Jia, Xin Wang Fudan University Shanghai, China [email protected] [email protected] [email protected] Shuang Yang Stanford
Master Course Computer Networks IN2097
Chair for Network Architectures and Services Prof. Carle Department for Computer Science TU München Master Course Computer Networks IN2097 Prof. Dr.-Ing. Georg Carle Christian Grothoff, Ph.D. Chair for
Green Routing in Data Center Network: Modeling and Algorithm Design
Green Routing in Data Center Network: Modeling and Algorithm Design Yunfei Shang, Dan Li, Mingwei Xu Tsinghua University Beijing China, {shangyunfei, lidan, xmw}@csnet1.cs.tsinghua.edu.cn ABSTRACT The
Network Virtualization and Data Center Networks 263-3825-00 Data Center Virtualization - Basics. Qin Yin Fall Semester 2013
Network Virtualization and Data Center Networks 263-3825-00 Data Center Virtualization - Basics Qin Yin Fall Semester 2013 1 Walmart s Data Center 2 Amadeus Data Center 3 Google s Data Center 4 Data Center
TRILL for Service Provider Data Center and IXP. Francois Tallet, Cisco Systems
for Service Provider Data Center and IXP Francois Tallet, Cisco Systems 1 : Transparent Interconnection of Lots of Links overview How works designs Conclusion 2 IETF standard for Layer 2 multipathing Driven
Core and Pod Data Center Design
Overview The Core and Pod data center design used by most hyperscale data centers is a dramatically more modern approach than traditional data center network design, and is starting to be understood by
Applying NOX to the Datacenter
Applying NOX to the Datacenter Arsalan Tavakoli UC Berkeley Martin Casado and Teemu Koponen Nicira Networks Scott Shenker UC Berkeley, ICSI 1 Introduction Internet datacenters offer unprecedented computing
APPLICATION NOTE 211 MPLS BASICS AND TESTING NEEDS. Label Switching vs. Traditional Routing
MPLS BASICS AND TESTING NEEDS By Thierno Diallo, Product Specialist Protocol Business Unit The continuing expansion and popularity of the Internet is forcing routers in the core network to support the
VMDC 3.0 Design Overview
CHAPTER 2 The Virtual Multiservice Data Center architecture is based on foundation principles of design in modularity, high availability, differentiated service support, secure multi-tenancy, and automated
IP Networking. Overview. Networks Impact Daily Life. IP Networking - Part 1. How Networks Impact Daily Life. How Networks Impact Daily Life
Overview Dipl.-Ing. Peter Schrotter Institute of Communication Networks and Satellite Communications Graz University of Technology, Austria Fundamentals of Communicating over the Network Application Layer
Cisco s Massively Scalable Data Center
Cisco s Massively Scalable Data Center Network Fabric for Warehouse Scale Computer At-A-Glance Datacenter is the Computer MSDC is the Network Cisco s Massively Scalable Data Center (MSDC) is a framework
Scale and Efficiency in Data Center Networks
Scale and Efficiency in Data Center Networks Amin Vahdat Computer Science and Engineering Center for Networked Systems UC San Diego [email protected] UC San Diego Center for Networked Systems Member Companies
Trends and impacts of new generation data center networking First CPqD International Workshop on New Architectures for Future Internet
Trends and impacts of new generation data center networking First CPqD International Workshop on New Architectures for Future Internet Christian Esteve Rothenberg, 24/09/2006 University of Campinas Agenda
Optimizing Data Center Networks for Cloud Computing
PRAMAK 1 Optimizing Data Center Networks for Cloud Computing Data Center networks have evolved over time as the nature of computing changed. They evolved to handle the computing models based on main-frames,
International Journal of Advance Research in Computer Science and Management Studies
Volume 2, Issue 8, August 2014 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online
A Coordinated. Enterprise Networks Software Defined. and Application Fluent Programmable Networks
A Coordinated Virtual Infrastructure for SDN in Enterprise Networks Software Defined Networking (SDN), OpenFlow and Application Fluent Programmable Networks Strategic White Paper Increasing agility and
TV INSIGHTS APPLICATION OF BIG DATA TO TELEVISION
TV INSIGHTS APPLICATION OF BIG DATA TO TELEVISION AN ARRIS WHITE PAPER BY: BHAVAN GANDHI, ALFONSO MARTINEZ- SMITH, & DOUG KUHLMAN TABLE OF CONTENTS ABSTRACT... 3 INTRODUCTION INTERSECTION OF TV & BIG DATA...
Analysis and Research of Cloud Computing System to Comparison of Several Cloud Computing Platforms
Volume 1, Issue 1 ISSN: 2320-5288 International Journal of Engineering Technology & Management Research Journal homepage: www.ijetmr.org Analysis and Research of Cloud Computing System to Comparison of
Chapter 3. Enterprise Campus Network Design
Chapter 3 Enterprise Campus Network Design 1 Overview The network foundation hosting these technologies for an emerging enterprise should be efficient, highly available, scalable, and manageable. This
IMPACT OF DISTRIBUTED SYSTEMS IN MANAGING CLOUD APPLICATION
INTERNATIONAL JOURNAL OF ADVANCED RESEARCH IN ENGINEERING AND SCIENCE IMPACT OF DISTRIBUTED SYSTEMS IN MANAGING CLOUD APPLICATION N.Vijaya Sunder Sagar 1, M.Dileep Kumar 2, M.Nagesh 3, Lunavath Gandhi
Technical Bulletin. Enabling Arista Advanced Monitoring. Overview
Technical Bulletin Enabling Arista Advanced Monitoring Overview Highlights: Independent observation networks are costly and can t keep pace with the production network speed increase EOS eapi allows programmatic
STATE OF THE ART OF DATA CENTRE NETWORK TECHNOLOGIES CASE: COMPARISON BETWEEN ETHERNET FABRIC SOLUTIONS
STATE OF THE ART OF DATA CENTRE NETWORK TECHNOLOGIES CASE: COMPARISON BETWEEN ETHERNET FABRIC SOLUTIONS Supervisor: Prof. Jukka Manner Instructor: Lic.Sc. (Tech) Markus Peuhkuri Francesco Maestrelli 17
ConnectX -3 Pro: Solving the NVGRE Performance Challenge
WHITE PAPER October 2013 ConnectX -3 Pro: Solving the NVGRE Performance Challenge Objective...1 Background: The Need for Virtualized Overlay Networks...1 NVGRE Technology...2 NVGRE s Hidden Challenge...3
Software Defined Networking and OpenFlow: a Concise Review
Software Defined Networking and OpenFlow: a Concise Review Stefano Forti [email protected] MSc in Computer Science and Networking Scuola Superiore Sant'Anna - University of Pisa 1. Introduction
Virtualizing the SAN with Software Defined Storage Networks
Software Defined Storage Networks Virtualizing the SAN with Software Defined Storage Networks Introduction Data Center architects continue to face many challenges as they respond to increasing demands
