An MPI Tool for Automatically Discovering the Switch Level Topologies of Ethernet Clusters
|
|
|
- Pearl Gilbert
- 9 years ago
- Views:
Transcription
1 An MPI Tool for Automatically Discovering the Switch Level Topologies of Ethernet Clusters Joshua Lawrence Xin Yuan Department of Computer Science, Florida State University, Tallahassee, FL {lawrence, Abstract We present an MPI topology discovery tool for homogeneous Ethernet switched clusters. Unlike existing Ethernet topology discovery methods that rely on Simple Network Management Protocol (SNMP) queries to obtain the topology information, our tool infers the topology from end-toend measurements. The tool works on clusters connected by managed and/or unmanaged Ethernet switches, and does not require any special privilege. We discuss the theoretical foundation of the tool, present the algorithms used, and report our evaluation of the tool. 1 Introduction Ethernet is the most popular local area networking technology due to its low cost and high performance cost ratio. It provides the connectivity for many low-end high performance computing (HPC) clusters. We will use the term Ethernet switched clusters to refer to clusters with commodity workstations/personal computers connected by Ethernet switches. Such clusters include both dedicated HPC clusters and casual clusters that are formed by a set of random computers working for the same application. Many HPC applications running on Ethernet switched clusters are message passing programs that use Message Passing Interface (MPI) [9] routines to perform communications. MPI defines the API for message passing and is the industrial standard for developing portable and scalable message passing programs. Two types of communication routines are specified in MPI: point-to-point routines that perform communications between two processes and collective communication routines that carry out communications among a group of processes. For an Ethernet switched cluster to efficiently support MPI applications, it is essential that the MPI communication performance is maximized. The physical switch level topology of an Ethernet switched cluster strongly affects the performance of MPI collective operations. In particular, in an Ethernet switched cluster with multiple switches, using topology specific communication algorithms can drastically improve the performance over the topology unaware algorithms used in common MPI libraries such as MPICH [10] and OPEN MPI [11]. To achieve high MPI performance, automatic routine generators that take the topology information as input and automatically produce topology specific MPI collective routines have been developed [4, 5, 12, 13]. Unfortunately, for many clusters, especially the casual clusters, the switch level topologies are usually unknown. Having an automatic topology discovery tool may significantly benefit MPIap- plications on such systems. In this paper, we present an MPI tool that automatically discovers the switch level topology of an Ethernet switched cluster. The tool infers the switch level topology based on end-to-end measurements. In comparison to the traditional Ethernet topology discovery methods/tools [1, 3, 6, 8] that rely on Simple Network Management Protocol (SNMP) queries to discover the topology, our tool offers several advantages. First, our tool does not require the switches to support SNMP and thus works on clusters connected by managed and/or unmanaged switches (the kind that does not support SNMP). Notice that unmanaged switches are much cheaper and are deployed much more widely than managed switches. Second, since our tool infers the topology based on end-to-end measurements, it does not require special privilege: anyone who can run MPI programs on a set of machines (that form a cluster) can use our tool to discover the switch level topology for the set of machines. The limitation of this tool is that it works best on homogeneous clusters with similar machines and switches, and may not be able to handle heterogeneous clusters. However, most MPI applications run on homogeneous clusters where this tool can effectively discover the topologies. We tested the tool on various homogeneous clusters connected by fast Ethernet (100Mbps) or Giga-bit Ethernet (1Gbps) switches with different configurations. The results show that the tool is robust in discovering the switch level topology. With this tool, the automatic routine generators [4, 5, 12, 13] can use the topology information of a cluster and automatically produce efficient topology spe-
2 cific MPI collective routines. The tool can be downloaded at xyuan/mpi. The rest of this paper is organized as follows: Section 2 discusses the related concepts in Ethernet switched clusters; Section 3 presents our topology discovery tool; Section 4 evaluates the components of the tool as well as the tool as a whole on different cluster configurations; Section 5 discusses the related work and Section 6 concludes the paper. 2 Ethernet switched clusters An Ethernet switched cluster consists of workstations/personal computers connected by Ethernet switches. We will use the term computers to refer to workstations or personal computers. Since our tool works best on homogeneous clusters, we will assume that the clusters are homogeneous with the same type of computers, links, and switches. The links are fully duplex: computers connected to an Ethernet switch can send and receive at the full link speed simultaneously. Each machine is equipped with one Ethernet card. Ethernet switches use a spanning tree algorithm to determine forwarding paths that follow a tree structure [15]. Thus, the switch level physical topology of the network is always a tree. The topology can be modeled as a directed graph G = (V, E) with nodes V corresponding to switches and computers, and edges E corresponding to unidirectional channels. Let S be the set of all switches in the cluster and M be the set of all computers (V = S M). Let u, v V, adirectededge (u, v) E if and only if there is a link between nodes u and v. We will call the physical connection between nodes u and v link (u, v), which corresponds to two directed edges (u, v) and (v,u). Since the network topology is a tree, the graph is also a tree with computers being leaves and switches being internal nodes. Notice that aswitchmayalsobealeaf,butwillnotparticipateinany communication and will be excluded from the topology. We define the hop-count distance between two end machines to be the number of switches in the path between them (the path is unique in a tree). Figure 1 shows an example cluster. n0 n1 n4 n5 n2 s0 s1 s2 n3 switch computer Figure 1. Ethernet switched cluster example 3 An automatic MPI topology discovery tool The topology discovery tool is an MPI program that automatically infers the switch level topology from end-to-end measurements. Two key features in Ethernet switched clusters facilitate the determination of the switch level topology from end-to-end measurements. First, Ethernet switches use the store-and-forward switching mechanism. Thus, when a packet travels through a switch, a notable latency is introduced. Let the link bandwidth be B and the packet size be msize, the store-and-forward latency is at least 2 msize B for receiving the packet from the input port ( msize B and msize B for sending the packet to the output port). Consider 1Gbps Ethernet switches. When sending a 1000-byte packet (msize = 1000B), eachswitchadds at least s = 16μs. Such a latency is reflected in the round-trip time (RTT) between each pair of machines. Hence, we can measure the RTT between the pairs of machines and obtain a hop-count distance from these measurements. Second, the switch level topology of an Ethernet switched cluster is always a tree [15]. When the hopcount distance between each pair of computers is decided, the tree topology can be uniquely determined (we will prove this and give an algorithm to compute the tree topology from the hop-count distances). The topology discovery process consists of two steps. First, the end-to-end round-trip time (RTT) between each pair of machines is measured and the hop-count distance is derived. The result of this step is a hop-count matrix with each entry in the matrix recording the hop-count between a pair of machines. Once the hop-count matrix is obtained, in the second step, the switch level (tree) topology is computed. In the following, we will first discuss how to compute the network topology from the hop-count matrix, and then describe how to obtain the hop-count matrix. 3.1 Computing the tree topology from the hop-count matrix Let the number of machines in the cluster be N and let the machines be numbered from 0 to N 1. The hop-count matrix, HC, isann N matrix. Each entry in the hopcount matrix, HC i,j, 0 i 6= j N 1, records the hop-count distance from machine i to machine j. Before we present the algorithm to compute the tree topology from the hop count matrix, we will show an example that a general topology (not a tree topology) cannot be uniquely determined from the hop-count matrix. Consider the 3-machine cluster in Figure 2 (a). Let us assume that the paths are bi-directional and path(0, 1) = 0 s0 s1 1; path(0, 2) = 0 s0 s3 s2 2; and path(1, 2) = 1 s1 s2 2. Thus,
3 0 HC = s3 s0 s1 s2 1 (a) 2 s0 s1 s Figure 2. An example This matrix is also the hop-count matrix for the topology shown in Figure 2 (b). This example shows that multiple (general) topologies can correspond to one hop-count matrix. Hence, a hop-count matrix cannot uniquely determine a corresponding topology. Fortunately, this is not that case when the topology is a tree. Next, we will show that the tree topology corresponding to a hop-count matrix is unique and can be derived from the hop-count matrix, which is the theoretical foundation of our tool. Lemma 1: LetG =(S M,E) be a tree topology, S being the set of switches, M being the set of machines, and E being the set of edges. When S 2, there exists a switch s S such that s is connected to exactly one other switch. Proof: We will provide a constructive algorithm to find such a switch (internal node) that connects to only one other switch. Starting from any switch, s 0. If this switch connects only to one other switch, then s = s 0 is found. Otherwise, pick any one of the switches connected to s 0. Let this switch be s 1.Ifs 1 connects only to one switch (s 0 ), then s = s 1 is found. Otherwise, s 1 connects to at least two switches. We will consider the switch that connects to s 1, but is not s 0.Lettheswitchbes 2.Ifs 2 connects to one switch (s 1 ), then the switch is found. Otherwise, there are at least two switches connected to s 2 with one not equal to s 1. We can proceed and consider that switch. This process is then repeated. Since the topology is a tree, each time a new switch will be considered (otherwise, a loop is formed, which contradicts the assumption that the topology is a tree). Since there are a finite number of switches in the topology, the process will eventually stop with the switch that connects to only one other switch being found. 2 Definition 1: LetM be the set of machines, tree topology G 1 =(S 1 M, E 1 ) is said to be equivalent to tree topology G 2 =(S 2 M,E 2 ) if and only if there exists a one-toone mapping function F : S 1 M S 2 M such that (1) for all u M, F(u) =u; (2)forany(u, v) E 1, (F(u),F(v)) E 2 ; and (3) for any (F(u),F(v)) E 2, (u, v) E 1. Theorem 1: LetM be the set of machines. Let G 1 =(S 1 M, E 1 ) and G 2 = (S 2 M,E 2 ) be two tree topologies connecting the machines M. LetH(G) be the hop-count (b) matrix for topology G. If H(G 1 )=H(G 2 ), then G 1 is equivalent to G 2. Proof: We prove this by induction on the size of S 1 (the number of switches in G 1 ). Base case, when there is only one switch in S 1. In this case, we have H(G 1 ) i,j = H(G 2 ) i,j =1, 0 i 6= j N 1. If G 2 also has one switch, then G 1 and G 2 is equivalent since there is only one waytoconnectthen machines (each with one Ethernet card) to that switch. If G 2 has more than one switch, there are at least two switches in G 2 that are directly connected to machines (switches can only be internal nodes in the tree). Let s 1 and s 2 be two such switches in G 2 and let s 1 directly connect to machine m 1 and s 2 directly connect to machine m 2. We have H(G 2 ) m1,m 2 2 since the path from machine m 1 to machine m 2 must pass through switches s 1 and s 2. This contradicts the fact that H(G 2 ) m1,m 2 =1. Hence, G 2 can only have one switch and is equivalent to G 1. Induction case: assume that when S 1 k, when H(G 1 ) = H(G 2 ), G 1 and G 2 are equivalent. Consider thecasewhen S 1 = k +1. From Lemma 1, we can find aswitchs S 1 that connects to only one other switch in G 1. This switch s must connect to at least one machine (otherwise it is a leaf, which is not allowed). Let the set of machines directly connect to s be M s. Clearly, in G 2,there must exist a switch x directly connecting to all machines in M s (otherwise, H(G 1 ) 6= H(G 2 )). Moreover, switch x must connect to one other switch in G 2. We will prove this statement by contradiction. Assume that switch x connects to more than one other switch in G 2. Let switches y and z be two such switches that directly connect to x. Let machine x m M s be a machine directly connected to switch x. Since switches are internal nodes, there must exist a machine whose path to x m goes through switch y. Let this machine be y m and the switch directly connected to y m be switch ly. Similarly, there exists a machine whose path to x m goes through switch z. We denote this machine as machine z m and the switch it directly connects to be switch lz. Let us assume that in G 1,machiney m directly connects to switch t and machine z m directly connects to switch r. Machine x m M s directly connects to switch s in G 1. Figure 3 depicts the connectivity among these machines/switches in G 1 and G 2. x m s a (a) G 1 t r y m z m x m (b) G 2 Figure 3. The connectivity of the related nodes in G 1 and G 2 x y z ly lz y m z m
4 In G 1 shown in Figure 3 (a), path(x m,y m ) and path(x m,z m ) share some switches (the first one is switch s). We denote the last switch in both paths be switch a. Hence, path(x m,y m )=x m s... a... t y m. path(x m,z m )=x m s... a... r z m.sinceg 1 is a tree topology, there is a unique path between any two nodes. Hence, the unique path between y m and z m is path(y m,z m )=y m t... a... r z m.sinces only connects to one other switch, a 6= s and H(G 1 ) ym,z m <H(G 1 ) xm,y m + H(G 1 ) xm,z m 1. In G 2 showninfigure3(b),path(x m,y m )=x m x y... ly y m ; path(x m,z m ) = x m x z... lz z m ; and the unique path from y m to z m is path(y m,z m )=y m ly... y x z... lz z m. Hence, H(G 2 ) ym,z m = H(G 2 ) xm,y m + H(G 2 ) xm,z m 1. Under the assumption that H(G 1 ) = H(G 2 ), we have H(G 1 ) xm,y m = H(G 2 ) xm,y m and H(G 1 ) xm,z m = H(G 2 ) xm,z m. Hence, H(G 1 ) ym,z m <H(G 1 ) xm,y m + H(G 1 ) xm,z m 1 = H(G 2 ) xm,y m + H(G 2 ) xm,z m 1 = H(G 2 ) ym,z m This contradicts that fact that H(G 1 )=H(G 2 ). Hence, switch x can only connect to one other switch in G 2. Both G 1 and G 2 has the same sub-tree: M s s in G 1 and M s x in G 2. We can construct a reduced graph G 1 by removing M s from G 1 and changing switch s to a machine, and a reduced graph G 2 by removing M s from G 2 and changing switch x to a machine. Clearly, H(G 1 )=H(G 2 ) when H(G 1 )=H(G 2 ). From the induction hypothesis, G 1 and G 2 are equivalent. Using the one-to-one mapping function, f, that maps the switches in G 1 to the switches in G 2,we can construct a one-to-one mapping function, g that maps switches in G 1 to the switches in G 2 : g(b) =f(b) for all b S 1 s and g(s) =x; andforallu M, g(u) =u. Clearly, (1) for any (u, v) E 1, (g(u),g(v)) E 2 ;and (2) for any (g(u),g(v)) E 2, (u, v) E 1. Thus, G 1 is equivalent to G 2. Hence, when H(G 1 )=H(G 2 ), G 1 is equivalent to G 2. 2 Theorem 1 states that the tree topology has a unique hopcount matrix. Next, we will present an algorithm that computes the tree topology from the hop-count matrix. The idea of the algorithm is as follows: when the hop count distances between all pairs of machines are 1 s, the topology consists of one switch that connects to all nodes. When the hopcount distance for some pair of machines is larger than 1, there is more than one switch in the topology. In this case, the algorithm finds one switch that connects to only one switch in the topology (Lemma 1), and constructs the subtree that consists of this switch and attached machines to this switch. After that, the algorithm reduces the size of the problem by treating the newly found sub-tree as a single machine (with one less switch than the original problem), recomputes the hop-count matrix for the reduced graph, and repeats the process until all switches are computed. This algorithm is shown in Figure 4. Input: the hop-count matrix, HC Output: the tree topology stored in an array parent[..]. Switches are numbered from N to N + S 1 in the result. (1) WorkHC = HC; size = N; (2) for(i =0;i<N; i ++)NodeNum[i] =i; (3) newswitch = N; (4) loop: (5) largestd = max i,j {WorkHC i,j } (6) if (largestd == 1) then /* the last switch left */ (7) for (i =0; i<size; i ++) parent[nodenum[i]] = newswitch; (8) exit; /* done */ (9) end if /* More than two switches left */ (10) Find a node n such that there exists a j, WorkHC n,j == largestd (11) Find all nodes whose distances to n are 1 s. (12) Let the set of nodes be A. (13) for all i A parent[nodenum[i]] = newswitch; (14) Reduce the tree by removing all nodes in A and adding a new leaf node newswitch, recompute WorkHC,andNodeNum. (15) newswitch ++ (16) end loop Figure 4. Algorithm for tree topology Figure 5 shows how the topology in Figure 1 is derived by the algorithm. The WorkHC stores the hop-count matrix in Figure 5 (a) in the first iteration. The nodes involved in the largest hop count distance are connected to switch s0 or s2. Assume that a node connected to s0 is selected. The sub-tree consisting of n0, n1, s0 is computed and the topology is reduced to Figure 5 (b) (WorkHC stores the hopcount distance for this topology in the second iteration). The newly added node is the blank circle. The hop-count distances for the reduced graph are derived from the old hop-count distances. For example, distance from the new node to n4 is equal to the distance from n0 to n4 minus 1. In the next iteration, the sub-tree of n0, n1, s0, n4, n5,and s1 is discovered and is reduced to Figure 5 (c), which is discovered in the last iteration. 3.2 Obtaining the hop-count matrix The algorithm in Figure 4 derives the topology from the hop-count matrix. The remaining issue is to obtain the hopcount matrix. In theory, it is fairly straight-forward to ob-
5 n0 s0 n1 n4 s1 n5 n2 s2 n3 (a) iteration 1 n4 s1 n5 n2 n3 s2 (b) iteration 2 n2 s2 n3 (c) iteration 3 Figure 5. An example for topology calculation tain the hop-count matrix from end-to-end measurements in a homogeneous cluster. The key issue is to filter out practical measurement anomalies, which may be caused by system noises, asynchronized system events, and other factors. However, automatically deriving the hop-count matrix from end-to-end measurements for a heterogeneous cluster is a much more difficult problem that we are still working on. Therefore, our tool works the best on homogeneous clusters. This tool uses three steps to obtain the hop-count matrix. Step 1: Raw RTT measurements. In this step, the tool uses MPI Send and MPI Recv to obtain the end-to-end RTT measurements. The main challenge is to ensure that the empirical measurement results are sufficiently accurate. Step 2: Data clustering. Statistical methods are applied to filter out the noises in the RTT measurements and to group the raw RTT measurement results into classes with each class corresponding to a hop-count distance. Step 3: Hop-count inference. Hop-count distances are inferred from the classified RTT results. Raw RTT measurements The RTT between each pair of machines is basically the ping-pong time between a pair of machines. Since the accuracy of the empirical measurements has a very significant impact on the effectiveness of the tool, our tool incorporates several mechanisms to ensure the ping-pong time measurements are accurate. RTT is measured with a reasonably large packet size. The default packet size in our tool is 1400 Bytes to ensure the latency from a 1Gbps (and 100Mbps) switch is greater than the measurement noise. Each RTT measurement is obtained by measuring multiple iterations of message round trips and using the average of the multiple iterations. Measuring multiple iterations increases the total time and improves the measurement accuracy. The tool uses a parameter NUMRTT to control the number of iterations in each RTT measurement. A statistical method is incorporated to ensure that the measured results are consistent. To determine one RTT between a pair of machines, a set of X RTT samples are measured. The average and the 95% confidence interval of the samples are computed. If the confidence interval is small (the ratio between the interval and the average is smaller than a predefined value), the results are consistent and accepted. If the confident interval is large, the results are inconsistent and rejected. In this case, the number (X) of samples to be measured is doubled and the process is repeated. The RTT value is decided when the measurement results are sufficiently consistent or when a predefined large number (1000) of samples are taken (in this case, the measurement results may still be inconsistant, but we will use the results regardless). Having a threshold for the 95% confidence interval statistically guarantees the quality of the RTT measurements. The tool has two parameters related to the statistical method: THRESHOLD that is used to determine whether the 95% confidence interval is sufficiently small, and INITPINGS that is the number of samples taken initially. The final measurement results are stored in the RawRTT array at the sender. After all RawRTT entries are measured, the results are gathered into machine 0, which will perform the later steps. Figure 6 shows an example RawRTT matrix that is obtained for the example topology in Figure 1. RawRTT = Figure 6. A round-trip-time matrix Data clustering As can be seen from Figure 6, the raw RTTs collected in the first step have noises. For example, in Figure 6, RawRTT[0][2] = 0.157ms and RawRTT[0][3] = 0.156ms. As discussed earlier, the difference in RTT values with different hop-count distances is at least 16μs = 0.016ms when the packet size is 1000B. Hence, the difference in the two measurements is due to measurement noises. The data clustering step tries to clean up measurement noises and group similar measurements into the same class, with each class covering a range of data values that correspond to a fixed hop-count distance. The technique to locate these classes is traditionally known as data clustering [14]. In the tool, we use a simple data clustering algorithm shown in Figure 7. In this algorithm, the data are sorted first. After that, data clustering is done in two phases. The first phase (lines (4) to (15)), raw clustering, puts measurements that differ by less than a threshold, (measurementnoise) intothesame
6 class. After raw clustering, in the second phase, the algorithm further merges adjacent classes whose inter-class distance is not sufficiently larger than the maximum intraclass distances. The inter-class distance is the difference between the centers of two classes. The maximum intraclass distance is the maximum difference between two data points in a class. In the tool, the inter-class distance is at least INTER THRESHOLD times larger than the maximum intra-class distances (otherwise, the two adjacent classes are merged into one class). INTER THRESHOLD has a default value of 4. Note that standard data clustering algorithms such as the k-means algorithm [14] can also be used. We choose the simple algorithm because it is efficient and effective in handling data in our tool. Input: RawRTT Matrix; Output: RTT classes (1) Sort all RTTs in an array, sorted[..]. The sorted array has N*(N-1) elements (2) measurementnoise = NUMRTT * (3) numclass = 0; /* initial clustering */ (4) class[numclass].lowerbound = sorted[0]; (5) prev = sorted[0]; (6) for (i=0; i<n (N 1); i++) (7) if (sorted[i] prev < measurementnoise) (8) prev = sorted[i]; /* expand the class */ (9) else (10) class[numclass].upperbound = prev; (11) numclass++; (12) class[numclass].lowerbound = sorted[i]; (13) prev = sorted[i] (14) end if (15) end for /* further clustering */ (16) do (17) for each class[i] (18) if ((intra-distance(i) is less than inter-distance(i, i+1)/inter THRESHOLD) (19) Merge class[i] and class[i+1] (20) end if (21) end for (22) until no more merge can happen Figure 7. Data clustering algorithm For the round-trip-time matrix in Figure 6, the clustering algorithm will group the data into three classes: Class 0 covers range [0.131, 0.134]; Class 1 covers range [0.155, 0.160]; and Class 2 covers range [0.173, 0.177]. Finally, the RTT measurements in a class are replaced with the center of the class so they are numerically the same. Inferring hop-count distances There are two ways that the hop-count distances can be inferred in the tool. When the hop-count distances between some machines in the cluster are known, the hopcount distances can be correlated with RTT times off-line. The RTT times (that correspond to hop-count distances) can be merged into the RawRTT times during data clustering. In this case, the hop-count distance for a class can be assigned to the known hop-count distance that corresponds to the RTT time inside the class. The hop-count distances are determined when data clustering is done. This approach can be applied when the hop-count distances of some machines in the cluster are known, which is common in practice. When the tool does not have any additional information, it uses a heuristic to infer the hop-count distance automatically. First, the heuristic computes the gaps between the center of adjacent classes. For the example from the previous section, the gap between Class 1 and Class 0 is = 0.025; and the gap between Class 2 and Class 1 is = The tool then assumes that the smallest gap, in this example, is the time introduced by one switch. The tool further assumes that the smallest time measured corresponds to the hop-count of one switch (two machines connecting to one switch): class[0].hops =1. The formula to compute the hop-count for other classes is class[i].hops= class[i 1].hops+ smallest gap class[i].center class[i 1].center + 2 b c smallest gap In the example, we have smallest gap = class[1].hops = 1 + b / c = 2, and class[2].hops =2+b / c =3. Hence, the hop-count matrix for this example is shown in Figure 8. HC = Figure 8. A Hop Matrix This heuristic does not work in all cases. It works when (1) there exists one switch connecting more than one machine; and (2) there exist machines that are two switches apart. Most practical systems consist of a small number of switches and satisfy these two conditions. 4 Evaluation We evaluate the topology discovery tool on homogeneous clusters with various configurations. Figure 9 shows these topologies. We will refer to the topologies in Figure 9 as topologies (1), (2), (3), (4), and (5). Topology (1) contains 16 machines connected by a single switch. Topologies
7 n0 n1 n15 n0 n1 n15 n16 n17 n31 n0 n1 n7 n8 n9 n15 n16 n17 n23 n24 n25 n31 S0 S0 S1 S0 S1 S2 S3 (1) (2) (3) n0 n1n2 n3 n8 n9n10 n11 n16 n17 n18 n19 n4 S0 S1 S2 n5n6 n7 n12 n13 n14 n15 n20 n21 n22 n23 n0 n1 n7 n8 n9 S0 S1 n24 n25 n26 n31 S2 n24 n31 n25 S3 n30 n26 n27 n29 n28 (4) (5) Figure 9. Topologies for evaluation (2), (3), (4), and (5) are 32-machine clusters with different network connectivity. The machines are Dell Dimension 2400 with a 2.8 GHz P4 processor, 640MB of memory, and 40GB of disk space. All machines run Linux (Fedora) with kernel. The Ethernet card in each machine is Broadcom BCM Gbps/100Mbps/10Mbps card with the driver from Broadcom. Both Giga-bit Ethernet (1Gbps) and fast Ethernet (100Mbps) switches are used in the test (each configuration containsonetype of switch). The Gigabit switches are Dell Powerconnect 2624 (24-port 1Gbps Ethernet switches) and the fast Ethernet switches are Dell Powerconnect 2224 (24-port 100Mbps Ethernet switches). In the evaluation, we have defined the following parameters. The RTT measurement scheme has a packet size of 1400 Bytes; NUMRTT= 5; INITPINGS= 26; the 95% confidence interval THRESHOLD is 3% of the average. In the data clustering scheme, the noise threshold is measurementnoise=0.001*numrtt (NUMRTT micro-seconds) and the inter-cluster distance is INTER THRESHOLD=4 times the maximum intra-cluster distance. The tool successfully recognizes all the topologies in Figure 9 with both 1Gbps and 100Mbps switches, which demonstrates the tool is robust in discovering switch level topologies. In the following, we will present timing results for the tool. Each result is an average of five executions. All times are in the unit of second. Table 1 shows the total execution time of the tool for each configuration in Figure 9. The second column contains the results with 100Mbps Ethernet switches; and the third column are the results with 1Gbps switches. As can be seen from the table, it takes about 120 seconds to discover the topology with 32 nodes using 100Mbps switches and less than 35 seconds for 1Gbps switches. We further breakdown the execution time of each phase in the tool in Tables 2 and 3, which show the times for raw RTT measurement (RTT), data clustering and hop count inferring (DCHC), and topology calculation (TC). As can be seen from the table, for all cases, more than 99.9% of the total time is spent on the RTT measurements phase. This is due to the O(N 2 ) entries in the RTT matrix that need to be measured and the extensive statistical techniques that are incorporated in the tool to ensure the accuracy of the measurement results. If there is a need to improve the execution time, one can focus on optimizing the RTT measurement methods in this tool. In the experiments, most (> 99%) ofthertt measurements take 26 samples to reach the confidence interval threshold. Few measurements require 52 samples. No measurements require more than 104 samples. The DCHC and TC phases run on machine 0 and do not involve communication. When N is small, the time for these two phases is insignificant. Topology 100Mbps 1Gbps (1) s 6.935s (2) s s (3) s s (4) s s (5) s s Table 1. Total run time for configurations Topology RTT DCHC TC (1) s.007s.005s (2) s.013s.008s (3) s.056s.010s (4) s.016s.010s (5) s.013s.008s Table 2. The execution time for each phase (100Mbps switches) 5 Related work DeterminingtheswitchleveltopologyinanEthernet cluster is a challenging and important problem. Most existing methods [3, 1, 8, 6] and commercial tools [7, 16]
8 Topology RTT DCHC TC (1) 6.926s.005s.004s (2) s.010s.007s (3) s.022s.020s (4) s.020s.011s (5) s.010s.007s Table 3. The execution time for each phase (1Gbps switches) rely on the Simple Network Management Protocol (SNMP) to query the Management Information Base (MIB) such as the address forwarding tables in the switches in order to obtain the topology information. These techniques are mainly designed for system management and assume the system administrator privilege. Moreover, they also require the switches to support SNMP, which is a layer-3 protocol and is not supported by simple layer-2 switches. In [2], the authors discussed the problems with the SNMP based approaches in general and advocated an end-system based approach. The technique proposed in [2], however, is also designed with the assumption that the system administrative privilege is granted. Moreover, the techniques to derive the topology in [2] are totally different from ours. Due to various difficulties in topology discovery, none of the existing methods can reliably discover topologies in all cases. In this paper, we present a new topology discovery tool that allows a regular user to discover the system topology for homogeneous clusters where MPI applications typically run. 6 Conclusion We introduce an MPI topology discovery tool for Ethernet switched clusters. The tool infers the topology from end-to-end measurements. We describe the theoretical foundation and the algorithms used in the tool. This tool is proved to be effective in discovering the topology and can be used with automatic MPI routine generators to produce efficient topology specific MPI collective routines. Acknowledgement This work is supported by National Science Foundation (NSF) grants: CCF , CCF , CCF References [1] Y. Bejerano, Y. Breitbart, M. GaroFalakis, R. Rastogi, Physical Topology Discovery for Large Multisubnet Networks, IEEE INFOCOM, pp , April [2] R. Black, A. Donnelly, C. Fournet, Ethernet Topology Discovery without Network Assistance. IEEE ICNP, [3] Yuri Breitbart, Minos Garofalakis, Ben Jai, Cliff Martin, Rajeev Rastogi, and Avi Silberschatz, Topology Discovery in Heterogenous IP Networks: The Net- Inventory System IEEE/ACM Transactions on Networking, 12(3): , June [4] A. Faraj, X. Yuan, P. Patarasuk, A Message Scheduling Scheme for All-to-all Personalized Communication on Ethernet Switched Clusters, IEEE Transactions on Parallel and Distributed Systems, 18(2): , [5] A. Faraj, P. Patarasuk, and X. Yuan, Bandwidth Efficient All-to-all Broadcast on Switched Clusters, International Journal of Parallel Programming, Accepted. [6] Hasan Gobjuka, Yuri Breitbart, Ethernet Topology Discovery for Networks with Incomplete Information, IEEE ICCCN, pp , Aug [7] HP OpenView. Web page at [8] Bruce Lowekamp, David O Hallaron, Thomas Gross, Topology Discovery for Large Ethernet Networks, ACM SIGCOMM Computer Communication Review, 31(4): , [9] The MPI Forum, The MPI-2: Extensions to the Message Passing Interface, Available at [10] MPICH - A Portable Implementation of MPI. [11] Open MPI: Open Source High Performance Computing, [12] P. Patarasuk, A. Faraj, and X. Yuan, Pipelined Broadcast on Ethernet Switched Clusters, IEEE IPDPS, April [13] P. Patarasuk and X. Yuan, Bandwidth Efficient Allreduce Operation on Tree Topologies, IEEE IPDPS Workshop on High-Level Parallel Programming Models and Supportive Environments, March [14] P. Tan, M. Steinbach, and V. Kumar, Introduction to Data Mining, Addison-Wesley, [15] Andrew Tanenbaum, Computer Networks, 4th Edition, [16] IBM Tivoli. Web page at Feb
Finding Ethernet-Type Network Topology is Not Easy
Finding Ethernet-Type Network Topology is Not Easy Hassan Gobjuka, Yuri Breitbart Department of Computer Science Kent State University Kent, OH 44242 {hgobjuka,yuri}@cs.kent.edu 1 Abstract In this paper
A Flexible Cluster Infrastructure for Systems Research and Software Development
Award Number: CNS-551555 Title: CRI: Acquisition of an InfiniBand Cluster with SMP Nodes Institution: Florida State University PIs: Xin Yuan, Robert van Engelen, Kartik Gopalan A Flexible Cluster Infrastructure
IP NETWORK MONITORING AND AUTOMATIC ANOMALY DETECTION
IP NETWORK MONITORING AND AUTOMATIC ANOMALY DETECTION Michael de Castro, DEEC, Instituto Superior Técnico, Lisboa, Portugal November 2008 Abstract. This paper presents a new tool designed to assist in
Additional Information: A link to the conference website is available at: http://www.curtin.edu.my/cutse2008/index.html
Citation: Veeramani, S. and Gopal, Lenin. 2008. Network monitoring tool, in Curtin University of Technology (ed), Curtin University of Technology Science and Engineering International Conference CUTSE
Network Discovery Tool
Bestin Jose Mahesh Kumar Shyamsunder Computer Science and Engineering Indian Institue of Technology Bombay November 28, 2004 1 Problem Statement 2 Introduction to Netdisco Features of NetDisco 3 4 5 Problem
IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 12, NO. 3, JUNE 2004 401
IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 12, NO. 3, JUNE 2004 401 Topology Discovery in Heterogeneous IP Networks: The NetInventory System Yuri Breitbart, Minos Garofalakis, Member, IEEE, Ben Jai, Cliff
Load Balancing Mechanisms in Data Center Networks
Load Balancing Mechanisms in Data Center Networks Santosh Mahapatra Xin Yuan Department of Computer Science, Florida State University, Tallahassee, FL 33 {mahapatr,xyuan}@cs.fsu.edu Abstract We consider
Network (Tree) Topology Inference Based on Prüfer Sequence
Network (Tree) Topology Inference Based on Prüfer Sequence C. Vanniarajan and Kamala Krithivasan Department of Computer Science and Engineering Indian Institute of Technology Madras Chennai 600036 [email protected],
APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM
152 APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM A1.1 INTRODUCTION PPATPAN is implemented in a test bed with five Linux system arranged in a multihop topology. The system is implemented
Scalability and Classifications
Scalability and Classifications 1 Types of Parallel Computers MIMD and SIMD classifications shared and distributed memory multicomputers distributed shared memory computers 2 Network Topologies static
Distributed Dynamic Load Balancing for Iterative-Stencil Applications
Distributed Dynamic Load Balancing for Iterative-Stencil Applications G. Dethier 1, P. Marchot 2 and P.A. de Marneffe 1 1 EECS Department, University of Liege, Belgium 2 Chemical Engineering Department,
Decentralized Utility-based Sensor Network Design
Decentralized Utility-based Sensor Network Design Narayanan Sadagopan and Bhaskar Krishnamachari University of Southern California, Los Angeles, CA 90089-0781, USA [email protected], [email protected]
D1.2 Network Load Balancing
D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June [email protected],[email protected],
Operating Systems and Networks Sample Solution 1
Spring Term 2014 Operating Systems and Networks Sample Solution 1 1 byte = 8 bits 1 kilobyte = 1024 bytes 10 3 bytes 1 Network Performance 1.1 Delays Given a 1Gbps point to point copper wire (propagation
A Catechistic Method for Traffic Pattern Discovery in MANET
A Catechistic Method for Traffic Pattern Discovery in MANET R. Saranya 1, R. Santhosh 2 1 PG Scholar, Computer Science and Engineering, Karpagam University, Coimbatore. 2 Assistant Professor, Computer
International Journal of Advanced Research in Computer Science and Software Engineering
Volume 2, Issue 9, September 2012 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com An Experimental
Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Res. Lett. Inf. Math. Sci., 2003, Vol.5, pp 1-10 Available online at http://iims.massey.ac.nz/research/letters/ 1 Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
4 High-speed Transmission and Interoperability
4 High-speed Transmission and Interoperability Technology 4-1 Transport Protocols for Fast Long-Distance Networks: Comparison of Their Performances in JGN KUMAZOE Kazumi, KOUYAMA Katsushi, HORI Yoshiaki,
Performance Evaluation of Linux Bridge
Performance Evaluation of Linux Bridge James T. Yu School of Computer Science, Telecommunications, and Information System (CTI) DePaul University ABSTRACT This paper studies a unique network feature, Ethernet
Graph Theory Applications in Network Security
Graph Theory Applications in Network Security Jonathan Webb1, Fernando Docemmilli2, and Mikhail Bonin3 Theory Lab - Central Queensland University Wayville SA 5034 E-mail addresses: (1) [email protected]
Acronis Backup & Recovery 10 Advanced Server Virtual Edition. Quick Start Guide
Acronis Backup & Recovery 10 Advanced Server Virtual Edition Quick Start Guide Table of contents 1 Main components...3 2 License server...3 3 Supported operating systems...3 3.1 Agents... 3 3.2 License
Echidna: Efficient Clustering of Hierarchical Data for Network Traffic Analysis
Echidna: Efficient Clustering of Hierarchical Data for Network Traffic Analysis Abdun Mahmood, Christopher Leckie, Parampalli Udaya Department of Computer Science and Software Engineering University of
FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM
International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT
Schedule 2e. Additional terms for Ethernet services
1. SERVICE DESCRIPTION The Interoute Ethernet Clear and Ethernet Flex Services offer Circuits carrying telecommunications traffic between Network Termination Points on the Interoute Ethernet Network and/
Self-Compressive Approach for Distributed System Monitoring
Self-Compressive Approach for Distributed System Monitoring Akshada T Bhondave Dr D.Y Patil COE Computer Department, Pune University, India Santoshkumar Biradar Assistant Prof. Dr D.Y Patil COE, Computer
System Interconnect Architectures. Goals and Analysis. Network Properties and Routing. Terminology - 2. Terminology - 1
System Interconnect Architectures CSCI 8150 Advanced Computer Architecture Hwang, Chapter 2 Program and Network Properties 2.4 System Interconnect Architectures Direct networks for static connections Indirect
Interconnect Efficiency of Tyan PSC T-630 with Microsoft Compute Cluster Server 2003
Interconnect Efficiency of Tyan PSC T-630 with Microsoft Compute Cluster Server 2003 Josef Pelikán Charles University in Prague, KSVI Department, [email protected] Abstract 1 Interconnect quality
Load balancing in a heterogeneous computer system by self-organizing Kohonen network
Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.
Analysis and Implementation of Cluster Computing Using Linux Operating System
IOSR Journal of Computer Engineering (IOSRJCE) ISSN: 2278-0661 Volume 2, Issue 3 (July-Aug. 2012), PP 06-11 Analysis and Implementation of Cluster Computing Using Linux Operating System Zinnia Sultana
Enhance Service Delivery and Accelerate Financial Applications with Consolidated Market Data
White Paper Enhance Service Delivery and Accelerate Financial Applications with Consolidated Market Data What You Will Learn Financial market technology is advancing at a rapid pace. The integration of
Ethernet Topology Discovery: A Survey
Journal of Communication and Computer 10 (2013) 951-959 Kamal A. Ahmat Department of Information Technology, City University of New York, New York 10075, USA Received: May 10, 2012 / Accepted: June 12,
Architecting Low Latency Cloud Networks
Architecting Low Latency Cloud Networks Introduction: Application Response Time is Critical in Cloud Environments As data centers transition to next generation virtualized & elastic cloud architectures,
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking
Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Burjiz Soorty School of Computing and Mathematical Sciences Auckland University of Technology Auckland, New Zealand
10 Port L2 Managed Gigabit Ethernet Switch with 2 Open SFP Slots - Rack Mountable
10 Port L2 Managed Gigabit Ethernet Switch with 2 Open SFP Slots - Rack Mountable StarTech ID: IES101002SFP The IES101002SFP 10-port Ethernet switch delivers flexibility and control of your network by
Clustering UE 141 Spring 2013
Clustering UE 141 Spring 013 Jing Gao SUNY Buffalo 1 Definition of Clustering Finding groups of obects such that the obects in a group will be similar (or related) to one another and different from (or
Using High Availability Technologies Lesson 12
Using High Availability Technologies Lesson 12 Skills Matrix Technology Skill Objective Domain Objective # Using Virtualization Configure Windows Server Hyper-V and virtual machines 1.3 What Is High Availability?
High-Speed TCP Performance Characterization under Various Operating Systems
High-Speed TCP Performance Characterization under Various Operating Systems Y. Iwanaga, K. Kumazoe, D. Cavendish, M.Tsuru and Y. Oie Kyushu Institute of Technology 68-4, Kawazu, Iizuka-shi, Fukuoka, 82-852,
Avaya ExpertNet Lite Assessment Tool
IP Telephony Contact Centers Mobility Services WHITE PAPER Avaya ExpertNet Lite Assessment Tool April 2005 avaya.com Table of Contents Overview... 1 Network Impact... 2 Network Paths... 2 Path Generation...
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing N.F. Huysamen and A.E. Krzesinski Department of Mathematical Sciences University of Stellenbosch 7600 Stellenbosch, South
Stability of QOS. Avinash Varadarajan, Subhransu Maji {avinash,smaji}@cs.berkeley.edu
Stability of QOS Avinash Varadarajan, Subhransu Maji {avinash,smaji}@cs.berkeley.edu Abstract Given a choice between two services, rest of the things being equal, it is natural to prefer the one with more
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck
Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering
Contents. Connection Guide. What is Dante?... 2. Connections... 4. Network Set Up... 6. System Examples... 9. Copyright 2015 ROLAND CORPORATION
Contents What is Dante?............................................. 2 Outline.................................................. 2 Fundamental............................................ 3 Required Network
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed
A Measurement Study of Peer-to-Peer File Sharing Systems
CSF641 P2P Computing 點 對 點 計 算 A Measurement Study of Peer-to-Peer File Sharing Systems Stefan Saroiu, P. Krishna Gummadi, and Steven D. Gribble Department of Computer Science and Engineering University
Cisco Active Network Abstraction 4.0
Cisco Active Network Abstraction 4.0 Product Overview Cisco Active Network Abstraction (ANA) is a flexible, vendor-neutral network resource management solution for a multitechnology, multiservice network
N5 NETWORKING BEST PRACTICES
N5 NETWORKING BEST PRACTICES Table of Contents Nexgen N5 Networking... 2 Overview of Storage Networking Best Practices... 2 Recommended Switch features for an iscsi Network... 2 Setting up the iscsi Network
Frequently Asked Questions
Frequently Asked Questions 1. Q: What is the Network Data Tunnel? A: Network Data Tunnel (NDT) is a software-based solution that accelerates data transfer in point-to-point or point-to-multipoint network
Ethernet Topology Discovery: A Survey
Ethernet Topology Discovery: A Survey Kamal A. Ahmat CITY UNIVERSITY OF NEW YORK/Information Technology New York, USA Email: [email protected] ABSTRACT Ethernet networks have undergone impressive
Introduction to LAN/WAN. Network Layer
Introduction to LAN/WAN Network Layer Topics Introduction (5-5.1) Routing (5.2) (The core) Internetworking (5.5) Congestion Control (5.3) Network Layer Design Isues Store-and-Forward Packet Switching Services
QoS Switching. Two Related Areas to Cover (1) Switched IP Forwarding (2) 802.1Q (Virtual LANs) and 802.1p (GARP/Priorities)
QoS Switching H. T. Kung Division of Engineering and Applied Sciences Harvard University November 4, 1998 1of40 Two Related Areas to Cover (1) Switched IP Forwarding (2) 802.1Q (Virtual LANs) and 802.1p
COMP3420: Advanced Databases and Data Mining. Classification and prediction: Introduction and Decision Tree Induction
COMP3420: Advanced Databases and Data Mining Classification and prediction: Introduction and Decision Tree Induction Lecture outline Classification versus prediction Classification A two step process Supervised
Schedule 2t. Additional terms for Fast Trade - Online Ordering
1. SERVICE DESCRIPTION Fast Trade is a high quality bandwidth service specifically designed for customers who require extremely low latency and low packet loss on links between two Network Termination
IP Network Topology Discovery Using SNMP
IP Network Topology Discovery Using SNMP Suman Pandey #1, Mi-Jung Choi #2, Sung-Joo Lee #3, James W. Hong #4 # Dept. of Computer Science and Engineering, POSTECH, Korea 1 [email protected], 2 [email protected],
Computer Network. Interconnected collection of autonomous computers that are able to exchange information
Introduction Computer Network. Interconnected collection of autonomous computers that are able to exchange information No master/slave relationship between the computers in the network Data Communications.
5 Performance Management for Web Services. Rolf Stadler School of Electrical Engineering KTH Royal Institute of Technology. [email protected].
5 Performance Management for Web Services Rolf Stadler School of Electrical Engineering KTH Royal Institute of Technology [email protected] April 2008 Overview Service Management Performance Mgt QoS Mgt
An Active Packet can be classified as
Mobile Agents for Active Network Management By Rumeel Kazi and Patricia Morreale Stevens Institute of Technology Contact: rkazi,[email protected] Abstract-Traditionally, network management systems
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
SIDN Server Measurements
SIDN Server Measurements Yuri Schaeffer 1, NLnet Labs NLnet Labs document 2010-003 July 19, 2010 1 Introduction For future capacity planning SIDN would like to have an insight on the required resources
Performance Monitoring of Parallel Scientific Applications
Performance Monitoring of Parallel Scientific Applications Abstract. David Skinner National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory This paper introduces an infrastructure
MAXIMIZING RESTORABLE THROUGHPUT IN MPLS NETWORKS
MAXIMIZING RESTORABLE THROUGHPUT IN MPLS NETWORKS 1 M.LAKSHMI, 2 N.LAKSHMI 1 Assitant Professor, Dept.of.Computer science, MCC college.pattukottai. 2 Research Scholar, Dept.of.Computer science, MCC college.pattukottai.
IBM Tivoli Composite Application Manager for Microsoft Applications: Microsoft Hyper-V Server Agent Version 6.3.1 Fix Pack 2.
IBM Tivoli Composite Application Manager for Microsoft Applications: Microsoft Hyper-V Server Agent Version 6.3.1 Fix Pack 2 Reference IBM Tivoli Composite Application Manager for Microsoft Applications:
Dynamic Congestion-Based Load Balanced Routing in Optical Burst-Switched Networks
Dynamic Congestion-Based Load Balanced Routing in Optical Burst-Switched Networks Guru P.V. Thodime, Vinod M. Vokkarane, and Jason P. Jue The University of Texas at Dallas, Richardson, TX 75083-0688 vgt015000,
WHITE PAPER OCTOBER 2014. CA Unified Infrastructure Management for Networks
WHITE PAPER OCTOBER 2014 CA Unified Infrastructure Management for Networks 2 WHITE PAPER: CA UNIFIED INFRASTRUCTURE MANAGEMENT FOR NETWORKS ca.com Table of Contents Solution Overview 3 Specialized Probes
Firewall Verification and Redundancy Checking are Equivalent
Firewall Verification and Redundancy Checking are Equivalent H. B. Acharya University of Texas at Austin [email protected] M. G. Gouda National Science Foundation University of Texas at Austin [email protected]
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 by Tan, Steinbach, Kumar 1 What is Cluster Analysis? Finding groups of objects such that the objects in a group will
Chapter 14: Distributed Operating Systems
Chapter 14: Distributed Operating Systems Chapter 14: Distributed Operating Systems Motivation Types of Distributed Operating Systems Network Structure Network Topology Communication Structure Communication
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining 4/8/2004 Hierarchical
Layer 3 Network + Dedicated Internet Connectivity
Layer 3 Network + Dedicated Internet Connectivity Client: One of the IT Departments in a Northern State Customer's requirement: The customer wanted to establish CAN connectivity (Campus Area Network) for
diversifeye Application Note
diversifeye Application Note Test Performance of IGMP based Multicast Services with emulated IPTV STBs Shenick Network Systems Test Performance of IGMP based Multicast Services with emulated IPTV STBs
Performance of Software Switching
Performance of Software Switching Based on papers in IEEE HPSR 2011 and IFIP/ACM Performance 2011 Nuutti Varis, Jukka Manner Department of Communications and Networking (COMNET) Agenda Motivation Performance
Scalable Data Analysis in R. Lee E. Edlefsen Chief Scientist UserR! 2011
Scalable Data Analysis in R Lee E. Edlefsen Chief Scientist UserR! 2011 1 Introduction Our ability to collect and store data has rapidly been outpacing our ability to analyze it We need scalable data analysis
Reminder: Complexity (1) Parallel Complexity Theory. Reminder: Complexity (2) Complexity-new
Reminder: Complexity (1) Parallel Complexity Theory Lecture 6 Number of steps or memory units required to compute some result In terms of input size Using a single processor O(1) says that regardless of
technical brief Optimizing Performance in HP Web Jetadmin Web Jetadmin Overview Performance HP Web Jetadmin CPU Utilization utilization.
technical brief in HP Overview HP is a Web-based software application designed to install, configure, manage and troubleshoot network-connected devices. It includes a Web service, which allows multiple
Chapter 15: Distributed Structures. Topology
1 1 Chapter 15: Distributed Structures Topology Network Types Operating System Concepts 15.1 Topology Sites in the system can be physically connected in a variety of ways; they are compared with respect
Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach
Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach S. M. Ashraful Kadir 1 and Tazrian Khan 2 1 Scientific Computing, Royal Institute of Technology (KTH), Stockholm, Sweden [email protected],
Paolo Costa [email protected]
joint work with Ant Rowstron, Austin Donnelly, and Greg O Shea (MSR Cambridge) Hussam Abu-Libdeh, Simon Schubert (Interns) Paolo Costa [email protected] Paolo Costa CamCube - Rethinking the Data Center
Dual Mechanism to Detect DDOS Attack Priyanka Dembla, Chander Diwaker 2 1 Research Scholar, 2 Assistant Professor
International Association of Scientific Innovation and Research (IASIR) (An Association Unifying the Sciences, Engineering, and Applied Research) International Journal of Engineering, Business and Enterprise
First Midterm for ECE374 02/25/15 Solution!!
1 First Midterm for ECE374 02/25/15 Solution!! Instructions: Put your name and student number on each sheet of paper! The exam is closed book. You have 90 minutes to complete the exam. Be a smart exam
Scaling 10Gb/s Clustering at Wire-Speed
Scaling 10Gb/s Clustering at Wire-Speed InfiniBand offers cost-effective wire-speed scaling with deterministic performance Mellanox Technologies Inc. 2900 Stender Way, Santa Clara, CA 95054 Tel: 408-970-3400
CHAPTER 4 PROPOSED GRID NETWORK MONITORING ARCHITECTURE AND SYSTEM DESIGN
39 CHAPTER 4 PROPOSED GRID NETWORK MONITORING ARCHITECTURE AND SYSTEM DESIGN This chapter discusses about the proposed Grid network monitoring architecture and details of the layered architecture. This
A Framework for End-to-End Proactive Network Management
A Framework for End-to-End Proactive Network Management S. Hariri, Y. Kim, P. Varshney, Department of Electrical Engineering and Computer Science Syracuse University, Syracuse, NY 13244 {hariri, yhkim,varshey}@cat.syr.edu
CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING
CHAPTER 6 CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING 6.1 INTRODUCTION The technical challenges in WMNs are load balancing, optimal routing, fairness, network auto-configuration and mobility
How To Monitor And Test An Ethernet Network On A Computer Or Network Card
3. MONITORING AND TESTING THE ETHERNET NETWORK 3.1 Introduction The following parameters are covered by the Ethernet performance metrics: Latency (delay) the amount of time required for a frame to travel
Detection of Promiscuous Nodes Using ARP Packets
Detection of Promiscuous Nodes Using ARP Packets Version 1.0 Written by: 31Aug01 Daiji Sanai Translated by: Kelvin KingPang Tsang http://www.securityfriday.com 1 Contents Abstract...3
Network Virtualization and Data Center Networks 263-3825-00 Data Center Virtualization - Basics. Qin Yin Fall Semester 2013
Network Virtualization and Data Center Networks 263-3825-00 Data Center Virtualization - Basics Qin Yin Fall Semester 2013 1 Walmart s Data Center 2 Amadeus Data Center 3 Google s Data Center 4 Data Center
Traffic Engineering for Multiple Spanning Tree Protocol in Large Data Centers
Traffic Engineering for Multiple Spanning Tree Protocol in Large Data Centers Ho Trong Viet, Yves Deville, Olivier Bonaventure, Pierre François ICTEAM, Université catholique de Louvain (UCL), Belgium.
Professor: Ian Foster TAs: Xuehai Zhang, Yong Zhao. Winter Quarter. www.classes.cs.uchicago.edu/classes/archive/2003/winter/54001-1
Professor: Ian oster Ts: Xuehai Zhang, Yong Zhao Winter Quarter www.classes.cs.uchicago.edu/classes/archive//winter/541-1 alculate the total time required to transfer a 1 KB file (RTT=1 ms, packet size
Clustering. 15-381 Artificial Intelligence Henry Lin. Organizing data into clusters such that there is
Clustering 15-381 Artificial Intelligence Henry Lin Modified from excellent slides of Eamonn Keogh, Ziv Bar-Joseph, and Andrew Moore What is Clustering? Organizing data into clusters such that there is
131-1. Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10
1/10 131-1 Adding New Level in KDD to Make the Web Usage Mining More Efficient Mohammad Ala a AL_Hamami PHD Student, Lecturer m_ah_1@yahoocom Soukaena Hassan Hashem PHD Student, Lecturer soukaena_hassan@yahoocom
Ethernet (LAN switching)
Ethernet ( switching) 1 Outline Interconnection devices Bridges/ switches vs. Routers Bridges Learning Bridges Transparent bridges 2 1 Bridges/ switches Interconnect multiple, possibly with different type
ENC Enterprise Network Center. Intuitive, Real-time Monitoring and Management of Distributed Devices. Benefits. Access anytime, anywhere
Scalability management up to 2,000 devices Network and device auto-discovery Firmware upgrade/backup and device configurations Performance alerts and monitoring ZyXEL switch specialized in RMON management
