Front.Comput.Sci. DOI REVIEW ARTICLE Rethinking the architecture design of data center networks Kaishun WU 1,2, Jiang XIAO 2, Lionel M. NI 2 1 National Engineering Research Center of Digital Life, State-Province Joint Laboratory of Digital Home Interactive Applications, School of Physics and Engineering, Sun Yat-sen University, Guangzhou 510006, China 2 Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Hong Kong, China c Higher Education Press and Springer-Verlag Berlin Heidelberg 2012 Abstract In the rising tide of the Internet of things, more and more things in the world are connected to the Internet. Recently, data has kept growing at a rate more than four times of that expected in Moore s law. This explosion of data comes from various sources such as mobile phones, video cameras and sensor networks, which often present multidimensional characteristics. The huge amount of data brings many challenges on the management, transportation, and processing IT infrastructures. To address these challenges, the state-of-art large scale data center networks have begun to provide cloud services that are increasingly prevalent. However, how to build a good data center remains an open challenge. Concurrently, the architecture design, which significantly affects the total performance, is of great research interest. This paper surveys advances in data center network design. In this paper we first introduce the upcoming trends in the data center industry. Then we review some popular design principles for today s data center network architectures. In the third part, we present some up-to-date data center frameworks and make a comprehensive comparison of them. During the comparison, we observe that there is no so-called optimal data center and the design should be different referring to the data placement, replication, processing, and query processing. After that, several existing challenges and limitations are discussed. According to these observations, we point out some possible future research directions. Keywords data center networks, switch-based networks, Received August 23, 2011; accepted January 15, 2012 E-mail: kwinson@ust.hk, jxiao@cse.ust.hk, ni@cse.ust.hk direct networks, hybrid networks 1 Introduction Because of rapid data explosion, many companies are outgrowing their current server space and more data centers are required. These data-intensive systems may have hundreds of thousands of computers and an overwhelming requirement of aggregate network bandwidth. Different from traditional hosting facilities, in these systems the computations continue to move into the cloud and the computing platforms are becoming warehouses full of computers. These new datacenters should not be simply considered as a collection of servers because plenty of the hardware and software resources are working together by taking a good Internet service provider (ISP) into account. To support such services, industry companies have invested greatly the data center constructions such as ebay, Facebook, Microsoft, and Yahoo. For example, Google had 10 million servers and Microsoft also had 50 000+ servers in their data centers in 2009 as shown in Fig. 1 [1]. During the last few years, research on data centers is growing fast. In these developments, networking is the only component that has not been changed dramatically [2]. Though this component does not have the largest cost among others, it is considered as one of the key sources to reduce cost and improve performance. Being aware of this, the architecture design of the data centers has become a hot research topic in the last decade. Nowadays, typical network architectures are http://www.datacenterknowledge.com/archives/2009/10/20/googleenvisions-10-million-servers
Kaishun WU et al. Rethinking the architecture design in data center networks 2 Fig. 1 Microsoft data center in Chicago a hierarchy of routers and switches. When the network size scales up and the hierarchy becomes deeper, more powerful (and much more expensive) routers and switches are needed. As the data centers further develop and expand, the gap between the desired bandwidth and provisioning increases, though the hardware develops quickly as well. Thus, one of the major challenges in architecture design is how to achieve higher performance while keeping the cost low. This article focuses on this question and will present some current works in this area. In the remainder of this article, we will first introduce some basic design principles. Subsequently, we give details of the interconnection techniques using commodity switches, such as fat-tree [3], DCell [4], and BCube [5]. We then make a comparison of the different data center architectures. Based on what we learned, we highlight some open challenges and limitations of the existing works and suggest some possible future directions. 2 Design principles/issues In this section, we present some design criteria and considerations of modern data center designs. Scalability: as the amount of data grows, we need more storage capacity to keep the data in the data center. One typical way to increase storage is to add more components instead of replacing old ones. As more hardware is integrated in data centers, the scalability of the data center network is crucial. Incremental scalability: in practice, instead of adding a huge number of servers at a time, we usually add a small number of storage hosts at a time. We expect minimal impact during the add-on on both the system operator and the system itself [6]. Cabling complexity: in traditional networks (e.g., homes and offices), the cabling complexity is simple. In data center environments, however, cabling is a critical issue when tens of thousands of nodes are hosted. The massive number of cables introduces many practical problems such as the connecting efforts, maintenance, and cooling. Bisection bandwidth: bisection bandwidth is defined as the bandwidth between two equal parts segmented from the original network using an arbitrary partition manner. This metric is widely used in performance evaluation of data center networks [3 5, 7, 8]. Aggregated throughput: the aggregate throughput measures the sum of the aggregated data rates when a networkwide broadcast is conducted. It is also known as the system throughput. Oversubscription: given a particular communication topology in a data center network, its oversubscription is defined as the ratio of the maximal aggregate bandwidth among the end hosts in the worst case to the total bisection bandwidth [3]. Fault Tolerance: hardware failures are common in largescale data centers, which make data center networks suffer from poor reliability and low utilization [9]. When hardware failures occur, alternative means are needed to ensure the availability of the data center networks. Energy consumption: recently, the power efficiency of data center networks has become increasingly important. To save the energy consumption, we can use low-power CPUs and GPUs, equip more efficient power supplies, and apply water-cooling mechanisms. Software means, such as visualization and smart cooling, can also help. Besides these, the architecture design is also critical for controlling the energy cost. Costs: the cost greatly affects the design decisions for building a large-scale data center networks. We hope to leverage economic off-the-shelf hardware for large-scale network interconnection [2]. Fairness: for applications that farm work out to many workers and finish when the last worker finishes,such fairness can greatly improve overall performance in data center networks. Reliability: high reliability is one of the most essential criteria for designing data center networks. In fact, a great waste of computing resources can be induced by an unreliable data center networks that cause the operation of applications and services to fail. Security: security is also critical for the success of the data
Front. Comput. Sci. China 3 center network services. Data exchanged between different nodes should be isolated from other unintended services to guarantee security. Latency: in data center networks, the delay incurred in the end systems or transmission between network nodes is called latency. Low latency interconnection in data center networks will benefit international data traffic. For example, by reducing the international data transmission latency, the colocation cost can be reduced. These criteria interplay with each other to influence the performance of data center networks. Such interaction includes checks and balances. For example, a data center network will induce high latency when some link fails (because there are too many hops to transmit a packet); a data center network with high reliability should also be fault tolerant. 3 Existing DCN architectures The state-of-art data center networks can be classified into three main classes according to the different network interconnection principles, namely switch-based networks, direct networks, and hybrid networks. In the next part we will elaborate and exemplify each of them. 3.1 Switch-based network A switch-based network, called an indirect network, typically consists of a multi-level tree of switches to connect the end servers (typically two or three levels). Switch-based networks are widely adopted and implemented in today s terascale data centers. They are able to support communications between tens of thousands of servers. Take a conventional three-level switch-based network as an example. The leaf switches (also known as the top of rack (ToR) switches) have a set of 1 Gbps Ethernet ports and are responsible for transferring packets within the rack. The layer two aggregation switches have 10 Gbps links to interconnect ToR switches, and these layer-2 switches will be connected by a more powerful switch when more hierarchy structure is applied In switch-based network architectures, the bottleneck is at the top level of the tree. Such bandwidth bottleneck is often alleviated by employing more powerful hardware at the expense of high-end switches. These solutions may increase the oversubscription problem and cause scalability issues. To address these issues, Fat-tree architecture [3] has been proposed. http://www.datacenterknowledge.com/archives/2009/10/20/googleenvisions-10-million-servers Instead of the skinny links that are used in a traditional tree, Fat-tree allows fatter links from the leaves towards the root. A typical Fat-tree can be split into three layers: core, aggregation, and edge (see Fig. 2). Suppose there are k pods (a pod is a small bunch of servers with certain connections in between) in the aggregation layer and each pod supports non-blocking operation among ( k 2 )2 core switches in a data center network (in the example of Fig. 2, k=4). The aggregate switches are divided into two groups: one is directly connected to k 2 servers and the other is connected to the remaining group. Consecutively, the total number of servers supported by a fat tree is k3 4. Each core switch can connect to other pod switches and ultimately connect to the servers. Fat-tree architecture performs as well as the traditional tree architecture but uses commodity switches only. It avoids the high-end expensive devices. Recently, the proliferation of cloud services incentivizes the construction of data center monsters. To supply a myriad of distinct services in such scale data centers, server utilizations should be improved as well. Towards this end, the agility of assigning any server to any service is an essential property for a data center network design. The architecture of higher agility can achieve a high utilization and the capability of dynamically allocating resources. For instance, Greenberg et al. [9] introduce the virtual layer 2 (VL2) architecture based on the basic fat tree topology. VL2 presents the attractive agility of connecting all the servers to a separate VL2 Ethernet switch with servers ranging from one to 100 000. VL2 deploys valiant load balancing (VLB) among multipath to ensure non-interfering network. To better host online applications running on substantial servers within a common multi-rooted tree data center, PortLand [10] is proposed. PortLand adopts a plug-and-play layer-2 design to enhance the fault tolerance and scalability. By employing an OperFlow fabric manger with the local switches in the edge of data center networks, PortLand can delicately make the appropriate forwarding decision. 3.2 Direct network Another option to make connections between the servers is by a direct network (also termed as router-based network ). Direct networks directly connect servers to other servers in the network without any switch, routers, or network devices. Servers will serve both as a server and a network forwarder. Direct networks are often used to provide better scalability, fault tolerance, and high network capacities. Some practical implementations of direct networks will be presented
4 Kaishun WU et al. Rethinking the architecture design in data center networks DCell1 Core DCell0[0] DCell0[1] DCell0[2] DCell0[3] DCell0[4] Aggregation Edge [0.0][0.1][0.2][0.3] [1.0][1.1][1.2][1.3] [2.0][2.1][2.2][2.3] [3.0][3.1][3.2][3.3] [4.0] 4.1] [4.2][4.3] Fig. 2 Switch-base architecture Fig. 3 DCell architecture Ethernet switch BCube k BCube k-1[0] BCube k-1[1] BCube k-1[n-1]..................... [0] [1]... [n k-1 ] [0] [1]... [n k-1 ] [0] [1]... [n k-1 ] Optical switch Fig. 4 BCube architecture Fig. 5 Hybrid architecture here. DCell [4] is one of the first direct data center networks. In DCell, servers are connected to several other servers via mini-switches with bidirectional communication links. A high-level DCell is constructed in a top-to-down, recursive manner. More specifically, denote a level k DCell as DCell k where k 0. Primarily, n servers and a mini-switch form a DCell 0 in which all servers are connected to the mini-switch. DCell 1 is therefore made of n + 1 DCell 0. All pairs of these n DCell 0 will be connected by a single link. Therefore in DCell 1, each sever has two links, one to connect to its miniswitch and the other to connect to a server in a DCell 0 (see Fig. 3 as an exmaple). Similarly we can construct DCell 2 with n + 1 DCell 1 and so on for DCell k using n + 1 DCell k 1. High network capacity and good fault tolerance are desirable traits in DCell. Though DCell is advantageous to scale out, its incremental scalability is incredibly poor. Specifically, as long as a DCell architecture is accomplished, it is very hard to add a small number of new servers to the architecture without ruining the original architecture structure. Moreover, the imbalanced traffic load also makes DCell perform poorly. To support unevenly distributed traffic load in data centers, generalized DCell framework [7] is proposed which has a smaller diameter and higher symmetry structure. With the data-intensive services spread out all over the world, a high degree mobility modular data center (MDC) is urged to emerge. Shipping-container based MDC is ideal for eliminating hardware administration tasks (e.g., installation, trouble-shooting, and maintenance). It achieves cost effectiveness and environmental robustness by deploying a severcentric approach. For instance, BCube [5] structure is specially devised for MDC consisting of multi-port servers and switches. Based on DCell, BCube is recursively constructed from n BCube 0 and n n-port switches. Such BCube k (k 1, denotes the level) is built from n BCube k 1 in which each server has k +1 ports. It is easy to see that a BCube k comprises n k+1 servers and k+1 levels of switches. Figure 4 illustrates the basic procedure of constructing BCubek. The employed strategy of BCube guarantees that switches merely connect to servers rather than other switches. Recently, Microsoft research proposed a project named CamCube [11] that applies a 3D torus topology. The 3D torus topology shares a similar idea of BCube. In 3D torus, each server directly connects to six other servers bypassing the usage of switches or routers. As the communication links between servers in data center are directed, higher bisection bandwidth is expected. 3.3 Hybrid network A novel approach to interconnect servers and switches appears in the rising tide of employing optical circuit switches. Compared to packet switching, the optical circuit switching
Front. Comput. Sci. China 5 is superior in terms of its ultra-high bandwidth, low transmission loss, and low power consumption. More importantly, optical switches are becoming commodity off-the-shelf (COT- S) and require shorter reconfiguration time thanks to the recent advances in micro-electro-mechanical systems (MEM- S). With these improvements, a number of data center networks deploy both optical circuit switching and electrical packet switching to make the connections. We call these hybrid networks as shown in Fig. 5. For instance, Helios [12] explores a hybrid 2-level multi-rooted tree architecture. By simply programming the packet switches and circuit switches, Helios creates an opportunity to provide preferable ultrahigh network bandwidth and a reduced number of wiring cables. Besides Helios, another hybrid data center network architecture is called c-through [13]. c-through makes a better use of the transient high capacity optical circuits by integrating optical circuit switches and packet switching servers. The optical circuit switches will buffer traffic to collect sufficient volumes for high-speed transmission. A key difference between Helios and c-through is that Helios implements its traffic load on switches while c-through alternatively utilizes the hosts to buffer data. Helios is advantageous for its transparency of traffic control at the end hosts, but requires the modification of every employed switch. In contrast, c- Through buffers data in the hosts, which allows c-through to amortize the workload over a longer period of time and utilize the optical link more effectively. Helios and c-through are t- wo typical hybrid schemes that attempt to optimize the data center networks by taking the advantages from both kinds of switches to make the data center networks most beneficial by optimizing both kind of switches. 4 A sea of architectures: which to choose? In the previous section we give insights of the state-of-art data center network architectures. These proposals exhibit promising features according to their measurements and performance evaluations and etc. It is, however, not clear how they perform when mutual comparison is conducted. We therefore make a comprehensive comparison between them. In this section we construct a typical data center network context and compare the performances of different proposals using the metrics in Section 2. In our comparisons, we compare the following alternatives and summarize them in Table 1. The traditional hierarchical tree structure presents the advantages of ease-of-wire but is limited by poor scalability. It is well known that tree-based architectures are vulnerable to link failures between switches and routers and therefore fault-tolerance is poor. Fat-tree solves this problem to some extent by increasing the number of aggregation switches but the wiring become much more complex. Multipath routing is effective in maximizing the network capacity such as the TwoLevelTable, hot-spot-routing used by VL2, and locationdiscovery-protocol (LDP) by PortLand. To cope with the tremendous workload volatility in data centers, fat-tree adopts VLB to guarantee the balance among different traffic patterns. In terms of the fault-tolerance, fat tree provides gracefully degraded performance, making it greatly outperform the tree structure. It develops a failure broadcast protocol to handle two groups of link failure between: (a) the lower- and upper-layer switches, and (b) the upper layer and core switches. Fat tree is also much more cost effective than the tree structure as it requires no expensive high-end switches and routers. DCell is an alternative proposal that adopts direct recursively defined interconnection topology. In DCell, servers in identical layers are fully connected, which makes it more s- calable than fat tree. However, incremental development is a strenuous mission for DCell due to the significant cabling complexity. In addition, traffic imbalance could be a severe obstacle to considering DCell as a primary choice. In most commercial companies today, a shippingcontainer-based modular data center meets the need of high degree mobility. BCube is the first representative Modular data center. It packs sets of servers and switches into a s- tandard 20- or 40- feet shipping-container and then connects different containers through external links. Based on DCell, BCube is designed to support various traffic loads and provide high bisection bandwidth. Load balancing is an appealing advantage of BCube compared to DCell. MDCube [14] s- cales the BCube structure to a mega level while ensuring high capacity at a reasonable cost. The server centric MDCube deploys a virtual generalized hypercube at the container level. This approach directly interconnects multiple BCube blocks using 10 Gbps optical switch links. Each switch functions as a virtual interface and each BCube block is treated as a virtual node. In a way that one node can have multiple interfaces; MDCube can interconnect a huge number of BCube blocks with high network capacity. Also, it delicately provides load balancing and fault tolerant routing to immensely improve the performance. For hybrid structures, electrical switches provide lowlatency immediate configuration and optical switches are advanced at the ultra-high speed data transmission, low loss, ultra-high bandwidth and low power consumption. To com-
6 Kaishun WU et al. Rethinking the architecture design in data center networks Table 1 General comparison of state-of-art data center architecture Tree Fat tree DCell BCube Hybrid Scalability Poor Good Excellent Good Good (scale up) (scale out) (scale out) (scale out) Incremental Good Good Poor Not necessary Good scalability Wiring Easy Easy Very difficult Difficult Easy Multipath No Switch and router End-host protocol upgrade Switch and router routing protocol upgrade protocol upgrade Fault Poor Against switch Against switch, router and Against switch, router tolerance and router failures end-host port failures and end-host port failures Cost High-end switches Low-end customized switches and routers Low-end Ethernet and routers (cheap but many) and optical switches Traffic balance No Yes No Yes Yes Graceful Poor Good Excellent Excellent Good degradation bine the best of both worlds, hybrid networks develop traffic demand estimation and traffic demultiplexing mechanisms to dynamically allocate traffic onto the circuit or packet switched network. In this section, we observe that the existing topologies of data center networks are similar with HPC. The difference between them is in the low layer design methods. For example, latency is a key issue in the both data center network and H- PC. But data transfer from memory in HPC is different from the data transfer between the servers in data center network. Existing data center architectures are all fixed. They may not provide adequate support for dynamic traffic patterns and different traffic loads. It is still an open question as to which architecture will perform best and whether the adaptive dynamic data center networks are feasible. 5 Challenges With the existing data center network designs, we identify some key limitations and point out some open challenges that can be the subject of future research. First, existing interconnection designs are all symmetrical. These symmetric architectures are difficult to extend when we need to add a small number of servers, or we have to lose the original network structure. In other words, these architectures have poor incremental salability. For example, in order to expand a data center of the hypercube architecture, the number of servers has to be doubled every time. In practice, however, most companies cannot afford the cost of adding such a large number of servers at one time. In BCube, D- Cell, and Fat-tree, when the present configuration is full, the network performance will be lower due to imbalance when only a small number of new servers are added. Besides these interconnection problems, heterogeneity is also a major issue in the network design. In 10 years time new technologies will be accessible, we will face a practical dilemma: either we have to integrate the old and new technologies into a single system, or we have to obsolete the old ones. It is yet an open question which will be the better choice. To deal with this problem, is it a good idea to reserve some place for such potential applications at the present time? Second, not only should we consider the connections within the data center but also the connections to the external world. In a switch based network such as Fat-tree, it is easy to connect to external world. In the direct networks we merely focus on the interconnection of the data center. Instead, take no account for connection to the external world into account. Clearly, the latter problem is also crucial. For example, HTTP serves the external traffic while MapReduce serves the internal traffics. In that case, we should not treat them as the same and take uniform actions. Different flows may have a varied impact on the network performance. We might consider the problem as a quality of service (QoS) problem and find an optimal design to better schedule the external traffic flow. Third, as MEMS further develops energy issues become increasingly important. The demand of different applications, may present various traffic patterns with certain unique characteristics. For example, Amazon EC2 is a cloud service that provides platform as a service (PAAS). In EC2, many users and applications run concurrently within a datacenter. Workloads are affected by user activities which are difficult if not impossible to predict. Thus, the traffic pattern con-
Front. Comput. Sci. China 7 stantly changes over time. In such cases, the networking design should not assume any fixed traffic pattern. Additionally, the design should also emphasize the network connections to the outside world (the Internet). Another example is that for a datacenter that runs data-intensive applications such as Hadoop, the network design may be optimized for bisectionbandwidth. It is less important to consider how to connect to the outside world. We can observe that some data may be used very frequently at a given time (called hot data ). Without careful design, the disk and servers may consume a lot of energy in the transfer between sleep and wake-up states. Some servers, in contrast, store data for backup purpose only (called cold data ); such servers and disks can safely stay in a sleep state to save energy. In practice, optimizations can be achieved by appropriate scheduling of servers between sleep and wake-up cycles. With this in mind, we can observe that data placement is also important for green data centers. Notice that there may not be a single optimal design that is suitable for all applications and thus the choice is likely to be application-dependent. But it is not yet clear what the implications are for such the applications. Today, the data may come from various sources such as mobile phones, camera videos, and sensor networks that present multidimensional characteristics. Different users may have different requirements according to their data. User requirements may have different workloads and thus have different traffic patterns. In that case, suppose we are designing a data center for the traffic data from monitor cameras or sensors. The optimal data center architecture is not straightforward. For example, if the trajectory of a taxi is distributed across several servers. How do we place this data so that we can search the trajectory of this taxi quickly? Should we replicate this data? When new trajectory data arrives at the data center, how can we migrate the original data? If we design a data center for such data, the data placement, replication and migration will also become challenges. All these questions are challenging to answer in practical environments. In addition, the application relies on the queries to be used. Each query also implies some communication patterns between nodes. 6 Conclusion and future directions This paper discusses the architecture design issues in current data center networks. Motivated by a better support for data intensive applications and supplying higher bandwidth performance, how to optimize the interconnection of data centers becomes a fundamental issue. We begin by introducing the development atmosphere of current data center architectures. Then, we comprehensively elaborate on the prevalent data center frameworks deployed in current enterprises and research associations. We compare several well-known architectures and remark on the existing limitations. After reviewing several representative architecture designs, we list some possible research directions. Thanks to the maturity of optical techniques, some aforementioned hybrid architectures have begun to use both optical and electrical components, e.g., Helios and c-through [12, 13]. As optical devices become more and more inexpensive, all optical architectures may become another direction for future data center architectures. In [6], the authors show us an all optical architecture and compare the cost with Fattree. Though their work is in the initial stages without extensive experimental evaluation, we believe that the all optical architecture will become a good choice for its high capacity nature. However, these pure wire data center have static characteristics which cannot solve dynamic cases. This means that once a data center is built, it is hard to change its topology. There are two possible solutions for future data center network design. The first applicable to some fixed applications, we design the architecture for their special traffic patterns using the metrics we mentioned in Section 2. With the spread of container based data centers, the architecture design is fixed in some cases. In that case, the data center should enable some dynamic design applications to achieve higher performance. Moreover, the cabling complexity is a big issue in data centers as it will waste much space, can be difficult to connect, hard to maintain and needs adequate cooling. On demand of these, the hybrid data center which combines wired and wireless networks may be a good choice for such requirements. Recently, the authors in [15] have proposed to leverage multi-gigabit 60 GHz wireless links in data centers that reduce interference and enhance reliability. In [16], the feasibility of wireless data centers has been explored by applying a 3D beamforming technique, and thus has introduced improvements in link range and concurrent transmissions. Similarly, by taking advantages of line-of-sight (LOS) on top-ofrack servers, steered-beam mmwave links have been applied in wireless data center networks [8]. As multi-gigabit-speed wireless communication is being developed and specified by the Wireless Gigabit Alliance (WiGig), wireless data centers (WDC) are likely to arrive in the near future. Instead of using wired connections, wireless technologies will bring many http://wirelessgigabitalliance.org
8 Kaishun WU et al. Rethinking the architecture design in data center networks advantages. For example, it is easy to expand WDC and its topology can be easily changed. Maintenance of the WDC is much easier since wireless nodes can be replaced easily. Also by using wireless connections, we can simply transmit packets from one node to any other node we wish. Building such data centers will take much less time for the connection of the servers. However, compared with the wired connections, wireless is less reliable and has lower channel capacity. How to design a good hybrid data center which combines wired and wireless to illustrate its performance will be a future research topic. Acknowledgements This research was supported in part by Pearl River New Star Technology Training Project, Hong Kong RGC Grant (HKUST617710, HKUST617811), the National High Technology Research and Development Program of China (2011AA010500), the NSFC-Guangdong Joint Fund of China (U0835004, U0935004, and U1135003), and the National Key Technology Research and Development Program of China (2011BAH27B01). References 1. Yang M, Ni L M. Incremental design of scalable interconnection networks using basic building blocks. IEEE Transactions on Parallel and Distributed Systems, 2000, 11(11): 1126 1140 2. Greenberg A, Hamilton J, Maltz D A, Patel P. The cost of a cloud: research problems in datacenter networks. ACM SIGCOMM Computer Communication Review, 2009, 39(1): 68 73 3. Al-Fares M, Loukissas A, Vahdat A. A scalable commodity data center network architecture. In: Proceedings of the ACM SIGCOMM 2008 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2008, 63 74 4. Guo C, Wu H, Tan K, Shi L, Zhang Y, Lu S. DCell: a scalable and fault-tolerant network structure for data centers. In: Proceedings of the ACM SIGCOMM 2008 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2008, 75 86 5. Guo C, Lu G, Li D, Wu H, Zhang X, Shi Y, Tian C, Zhang Y, Lu S. BCube: a high performance, servercentric network architecture for modular data centers. In: Proceedings of the ACM SIGCOMM 2009 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2009, 63 74 6. Singla A, Singh A, Ramachandran K, Xu L, Zhang Y. Proteus: a topology malleable data center network. In: Proceedings of 9th ACM workshop on Hot Topics in Networks. 2010 7. Kliegl M, Lee J, Li J, Zhang X, Guo C, Rincon D. Generalized DCell structure for load-balanced data center Networks. In: 2010 INFOCOM IEEE Conference on Computer Communications Workshops. 2010 8. Katayama Y, Takano K, Kohda Y, Ohba N, Nakano D. Wireless data center networking with steered-beam mmwave links. In: Proceeding of 2011 IEEE Wireless Communications and Networking Conference. 2011, 2179 2184 9. Greenberg A G, Hamilton J R, Jain N, Kandula S, Kim C, Lahiri P, Maltz D A, Patel P, Sengupta S.VL2: a scalable and flexible data center network. In: Proceedings of the ACM SIGCOMM 2009 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2009, 51 62 10. Mysore R N, Pamboris A, Farrington N, Huang N, Miri P, Radhakrishnan S, Subramanya V, Vahdat A. Port- Land: a scalable fault-tolerant layer 2 data center network fabric. In: Proceedings of the ACM SIGCOMM 2009 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2009, 39 50 11. Costa P, Donnelly A, O Shea G, Rowstron A. CamCube: a key-based data center. Technical Report, Microsoft Research, 2010 12. Farrington N, Porter G, Radhakrishnan S, Bazzaz H H, Subramanya V, Fainman Y, Papen G, Vahdat A. Helios: a hybrid electrical/optical switch architecture for modular data centers. In: Proceedings of the ACM SIGCOM- M 2010 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2010, 339 350 13. Wang G, Andersen D G, Kaminsky M, Papagiannaki K, Ng T S E, Kozuch M, Ryan M P. c-through: part-time optics in data centers. In: Proceedings of the ACM SIG- COMM 2010 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2010, 327 338 14. Wu H, Lu G, Li D, Guo C, Zhang Y. Mdcube: a high performance network structure for modular datacenter interconnection. In: Proceedings of the 2009 ACM Conference on Emerging Networking Experiments and Technology. 2009, 25 36 15. Halperin D, Kandula S, Padhye J, Bahl P, Wetherall D. Augmenting data center networks with multi-gigabit wireless links. In: Proceedings of the ACM SIGCOMM 2011 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications. 2011, 38 49 16. Zhang W, Zhou X, Yang L, Zhang Z, Zhao B Y, Zheng H. 3D beamforming for wireless data centers. In: Proceedings of the 10th ACM Workshop on Hot Topics in Networks. 2011
Front. Comput. Sci. China Kaishun Wu is currently a research assistant professor at the Hong Kong University of Science and Technology (HKUST). He received PhD degree in Computer Science and Engineering from HKUST in 2011. He received his BEng degree from Sun Yat-sen University in 2007. His research interests include wireless communications, mobile computing, wireless sensor networks, and data center networks. Jiang Xiao is a first year PhD student in Hong Kong University of Science and Technology. Her research interests focus on wireless indoor localization systems, wireless sensor networks, and data center networks. Lionel M. Ni is chair professor in the Department of Computer Science and Engineering at the Hong Kong University of Science and Technology (HKUST). He also serves as the special assistant to the president of HKUST, he is the dean of HKUST Fok Ying Tung Graduate School and visiting chair professor of Shanghai Key Lab of Scalable Computing and Systems at Shanghai Jiao Tong University. A fellow of IEEE, Prof. Ni has chaired over 30 professional conferences and has received six awards for authoring outstanding papers. 9