Performance Comparison of Dynamic Load-Balancing Strategies for Distributed Computing

Size: px
Start display at page:

Download "Performance Comparison of Dynamic Load-Balancing Strategies for Distributed Computing"

Transcription

1 Performance Comparison of Dynamic Load-Balancing Strategies for Distributed Computing A. Cortés, A. Ripoll, M.A. Senar and E. Luque Computer Architecture and Operating Systems Group Universitat Autònoma de Barcelona Bellaterra (Barcelona).Spain {a.cortes, a.ripoll, m.a.senar, e.luque}#cc.uab.es Abstract * The DASUD (Diffusion Algorithm Searching Unbalanced Domains) algorithm belongs to the nearestneighbours class and operates in a diffusion scheme where a processor balances its load with all its neighbours. DASUD detects unbalanced domains and performs local exchange of load between processors to achieve global balancing. The DASUD algorithm has been evaluated by comparison with another well-known strategy, namely, the SID (Sender Initiated Diffusion) algorithm across a range of networks topologies including ring, torus and hypercube where the number of processors varies from 8 to 128. From the experiments we have observed that DASUD outperforms the other strategy as it provides the best trade-off between the balance degree obtained at the final state and the number of iterations required to reach such state. DASUD is able to coerce any initial load distribution into a highly balanced global state and also exhibits good scalability properties. 1. Introduction The load-balancing problem in parallel computation is concerned with how to distribute the workload of a computation among the available processors so that each processor would have the same or nearly the same amount of work to do. In most cases, load balancing is done prior to execution and is done only once called static load balancing or static mapping-. This static balancing can be quite effective for computations that have predictable runtime behaviours [1]. For computations whose run-time behaviour is non-deterministic or not so predictable, however, performing load balancing only once in the beginning is insufficient. For these cases, it might be better * This work was supported by the CICYT under contract TIC to perform the load balancing more than once or periodically during run-time because the problem s variable behaviour more closely matches available computational resources. For example, in data parallel applications, the computational requirements associated with different parts of a problem domain may change as the computation proceeds. This occurs when the behaviour of the physical system being modelled changes with time. Such adaptive data parallel computations appear frequently in scientific and engineering applications such as those in molecular dynamics and computational fluid dynamics. This paper is about load balancing in distributed memory message-passing parallel computers. Each processor has its own address space and has to communicate with other processors by message passing. In general, a direct point-to-point interconnection network is used for the communications. Many commercial parallel computers are of this class, including the Intel Paragon, the Thinking Machine CM-5, IBM SP2, Origin 2000, and Cray T3D/T3E. The focus is on nearest-neighbours load balancing methods in which every processor at every step communicates simultaneously with all its nearest neighbours in order to reach a local balance. Nearestneighbour methods are iterative in nature because a global balanced state could be reached through processor s successive local operations. Nevertheless, the proposal strategies [2,3,4,5,6] assume that workloads are infinitely divisible and hence represent the workload of a processor by a real number. This assumption is valid in parallel programs that exploit very fine grain parallelism. To cover medium and large grain parallelism, the algorithm must be able to handle indivisible tasks. Under this more realistic assumption, the previous strategies may fail to guarantee global load balance. A new algorithm DASUD (Diffusion Algorithm Searching Unbalanced Domains) flexible in terms of allowing one to control the balancing qualities, effective for preserving communication locality and easily scaled in 1

2 parallel computers with a direct communication network, was proposed in [7,8]. This paper compares the DASUD algorithm with a well-known nearest-neighbour SID (Sender Initiated Diffusion) algorithm..the performance characteristics of DASUD have been evaluated using simulation experiments. The results illustrate the benefits offered by DASUD with regard to the balance quality (the maximum difference of load between processors) and the efficiency of the algorithm measured as the number of steps and communication cost required to drive an initial workload distribution into a uniform distribution. The rest of the paper is organised as follows. In section 2 the DASUD strategy is described. The simulation results concerning the goodness of DASUD strategy are reported in section 3 and, finally, in section 4 the main conclusions are presented. 2. Description of the DASUD strategy DASUD is an asynchronous nearest-neighbours strategy based on the SID (Sender Initiated Diffusion) strategy proposed by Willebeek-LeMair et alter. in [5]. SID uses overlapping neighbourhood domains to achieve global load balancing over the network. A threshold identifies the overloaded processor (sender). A sender performs load balancing whenever each load level is greater than the threshold value. Once the sender is identified using the threshold, the next step is to determine the amount of load (number of tasks or data set) to transfer to the sender s neighbours. This dynamic load balancing strategy uses local state information to guide load distribution. The processor selection and task-transfers policies are distributed in nature: all processors in the network have the responsibility of achieving global local balance. This strategy assumes that workloads are infinitely divisible and hence represents the workload of a processor by a real number. Nevertheless, the SID algorithm can be adapted to the integer workload model using floor and ceiling functions. However, this integer approach may fail to guarantee a global balanced situation. Although the load of each processor may differ by only one unit at most from that of its neighbours, the global load balance may be very poor. DASUD was developed to solve this problem detecting unbalanced domains and it performs local exchange of load between processors to achieve a global load balanced state (where the maximum load difference between any two processors is one unit). The behaviour of DASUD can be summarised in figure 2.1 where the load of a processor i at time t is defined as w i (t) and w(t) = (w 1 (t),...w n (t)) represents the global load vector at time t. Each processor executes the same group of operations at each iteration of the load-balancing algorithm. First, each processor sends its load information to all its neighbours and receives the load information from all its neighbours as well (lines 4.1). Then, it computes the load average of its domain, including the load of all its neighbours and itself, and it also computes its value d ii (t) (line 4.2). If processor i is an overloaded processor d ii will be a negative value (d ii <0). Otherwise, if processor i is an under-loaded processor d ii will be a non-negative value (d ii 0). An overloaded processor i (d ii < 0) performs load balancing by apportioning its excess load only to deficient neighbours. So, a new weight d + ij ()is t computed for all neighbours with deficit of load (line 4.4) and the total amount of load deficits is computed on td to evaluate the proportion of processor i s load excess that is assigned to neighbour j, P ij (t) (line 4.5). The amount of i s load excess to be sent is finally computed as s ij (t) and sent to processor j (line 4.6). (1) while (not converged) do begin DASUD Algorithm (2) for ALL processors i (3) parbegin (4.1) Exchange load information with neighbours SID Algorithm (4.2) compute: w i (t) and d ii(t) = w i (t) - w i(t) (4.3) if (processor i has excess of load) then (4.4) evaluate the load deficits of processor i s neighbours: + if (d ij (t)>0) then dij ( t) = dij( t) else d + ij ( t) = 0 + (4.5) and its portion of excess load to be moved: Pij ( t) = dij td (4.6) send s ij(t) = floor ( - (P ij(t) * d ii(t)) units of load to processor j (4.7) if (s ij(t) = 0 ƒor all neighbours) then compute: w i max (t) ; w max vi (t ) ; w min i (t) ; w min vi (t) ; (4.8) if (( wi max (t) - wi min (t) )> 1) then begin (4.9) if (((w ii(t) = wi max (t) ) and ( wvi max (t) = wvi min (t) )) then distributes the excess load unit by unit among the deficient neighbors. (4.10) if ((w i(t) = wi max (t) ) and ( wvi max (t) J wvi min (t) )) then send one unit of load to one of the lowest loaded neighbour (4.11) if (w i(t) wi max (t) ) then send an instruction message to a highest loaded neighbour processor saying it must send a unit of load to one of the lowest loaded processor. end (4.12) if no units of load have been already sent then receive and sort all instruction messages and send a unit of load to the first processor of that list (5) delete the rest of instruction messages (6) parend (7) end Figure 2.1 DASUD algorithm 2

3 DASUD incorporates new features to the previous steps that detect whether a domain is balanced or not. For that purpose the following four parameters are evaluated (line 4.7): a) maximum load value of the domain: wi max (t), b) minimum load value of the domain: wi min (t), c) maximum load value of processor neighbours: wvi max (t), d) minimum load value of processor neighbours: wvi min (t). If the maximum load difference between the neighbours and its own is more than one (line 4.8), there is a load imbalance and units of load will be distributed based on the following three actions: Action 1: If processor i is the processor with maximum load of its domain and all its immediate neighbours have the same load, then processor i will distribute α= ( w max i ()-w t min i () t - 1) units of load to its neighbours (line 4.9). Action 2: If processor i is the processor with maximum load of its domain but not all the neighbours have the same load, then one unit of load is sent to one processor of the less loaded processors (line 4.10). Action 3: if the domain of processor i is not balanced but processor i is not the most loaded processor, then processor i will instruct one of its neighbours with maximum load to send a load unit to one of its neighbours with minimum load (line 4.11) Finally, each processor that has not sent any load unit in the previous actions waits for instruction messages generated by other neighbours in action 3. If instruction messages arrive, they are sorted and one unit of load is sent to the first processor in the list (line 4.12). A detailed justification for all the actions carried out on the instruction messages are out of the scope of this paper because they are needed to ensure the convergence of DASUD. Details of the formal proof of DASUD convergence can be found in [8]. 3. The experimental study In this section, we compare SID and DASUD algorithms with respect to their stability and efficiency. The stability (or balance quality) measures the ability of an algorithm to coerce any initial load distribution into an equilibrium state, i.e., to reach the global uniform distribution state. The efficiency measure is reflected by the time incurred in the load communication steps and the number of balancing steps required by the algorithm to drive an initial workload distribution into a stable distribution. To see the effect of DASUD over SID, different processor networks were simulated with different and representative initial load distributions. The following k- ary n-cube topologies have been used: k-ary 1-Cube (ring), 2-ary n-cube (hypercube) and k-ary 2-Cube (2- dimensional torus). The sizes of these communication networks were: 8, 16, 32, 64 and 128 processors (notice that, in order to have square k-ary 2-Cube, instead of 8, 32 and 128 processors, the sizes of these topologies have been changed by 9 (3x3), 36 (6x6) and 121 (11x11) respectively). Synthetic load distributions consist of a set of initial load distributions, w(0). The total workload is denoted as L. So, we can evaluate a priori the expected final load at each processor, i.e. the global load average, L/n or L/n, where n is the size of the topology. In our experiments the problem size was chosen as L = Initial load distributions were classified into two main groups: likely distributions and pathological distributions. Likely distributions cover all the situations that are assumed to appear in real scenarios where most of the processors start from an initial load that is not zero. In this case, each element w i (0) has been obtained by random generation from one of four uniform distributions patterns. The four patterns used in likely distributions were the following: varying 25% from the global load average: i w i (0) [ L/n-0.25*L/n, L/n+0.25*L/n] varying 50% from the global load average: i w i (0) [ L/n-0.50*L/n, L/n+0.50*L/n] varying 75% from the global load average: i w i (0) [ L/n-0.75*L/n, L/n+0.75*L/n] varying 100% from the global load average: i w i (0) [ L/n-L/n, L/n+L/n] The 25% variation pattern corresponds to the situation where all processors have a similar load at the beginning and these loads are close to the global average, i.e., the initial situation is quite balanced. On the other hand, the 100% variation pattern corresponds to the situation where the difference of load between processors at the beginning is considerable. 50% and 75% variation patterns constitute intermediate situations between the other two. For every likely distribution pattern, 10 different initial load distributions were used. The group of pathological distributions was also used in order to evaluate the behaviour of the strategies under extreme initial distributions. In these distributions a significant amount of processors has a zero initial load. These scenarios seem less likely to appear in practice but we have used them for the sake of completeness in the evaluation of the strategies. The pathological distributions were classified in four groups: A spiked initial load distribution, where all the load is located on a single processor: 25% of idle processors. 50% of idle processors. 75% of idle processors. 3

4 In addition to the above mentioned distributions, each one was scattered using two different shapes: a single mountain shape and a chain shape defined as follows: 1. Single Mountain (SM), where load values from the initial load distribution have been scattered by drawing a single mountain surface (see figure 3.1). 2. Chain, where load values from the initial load distribution have been scattered by drawing multiple mountain surfaces (see figure 3.1). Figure 3.1 Single Mountain and Chain shapes As a consequence, we have evaluated not only the influence of the values of initial load distribution, but also the influence of how these values are collocated onto the processors. To sum up, the total number of distributions tested for a given processor network was 87, which were obtained in the following way: 10 likely distributions * 4 patterns * 2 shapes + 3 pathological distributions * 2 shapes + 1 spiked pathological distribution. The simulation process was run until a global termination detection was accomplished. This termination condition can be a limit on the number of simulation steps set beforehand or the detection that no load movements have been carried out from one step to the next one. In our experiments, simulations were stopped when no new load movements were performed from one step to the next one. Although the simulation did not mimic the truly asynchronous behaviour of the algorithms, their results can still help us to understand the performance of the algorithms since the final load imbalances, for instance, are similar whether the algorithm is implemented synchronously or asynchronously. The main difference is in the convergence speed. 3.1 Stability analysis. As we have mention above, the stability reflects the ability of an algorithm in bounding any initial load distribution into an equilibrium state, that is, to reach the global uniform distribution state. Since we are dealing with integer load values, the final balanced state will be the one where the maximum load difference between any two processors of the topology should be zero or one depending on L and the number of processors. If L is an exact multiple of n, the optimal final balanced state is the one where the maximum load difference between any two processors of the system is zero. Otherwise, it should be one. Figure 3.2 shows the maximum load differences (dif_max) reached by DASUD and SID algorithms for all topologies and architectures sizes when the initial load distribution varies from 25% to 100% from the global load average (i.e. all the likely distributions). The effect of pathological distributions is shown in figure 3.2 when the number of idle processors varies from 25% to n-1. These global results indicate that DASUD strategy outperforms SID strategy in all cases. On the average, the maximum load difference obtained by SID was more than 4 times the maximum load difference obtained by DASUD. Moreover, the maximum load differences obtained by SID grew worse as the initial unbalance degree increases. Moreover, tables 3.1 and 3.1 show the standard deviation of load with respect to the load average obtained by both strategies for all the load distributions used in our experiments. As it can be seen, DASUD achieves a deviation that is very low for all topologies and distributions (always less than 1 for hypercubes and torus and less than 7 for rings). In contrast, SID exhibits a high deviation for all cases (always more than 4 times the deviation obtained by DASUD). From the results shown in figure 3.2 and tables 3.1 and 3.1, we can conclude that, on the one hand, DASUD achieves a smaller maximum difference than SID on average and, on the other hand, all the processors have a better final state, i.e., the overall system is closer to the optimal balanced state. Below we analyse the influence of some parameters in the final results. First, we compare the behaviour of SID and DASUD with respect to the topology and we give an upper bound derived experimentally for the maximum difference that DASUD can obtain for a given topology. Then, as DASUD proved to be the strategy that achieved a better stability, we give a more detailed information about the influence of the load distribution and its shape on its final results. 3.1.a. Influence of the topology on the stability. The topology influence is shown in figures 3.3 and 3.3 for likely and pathological initial load distributions, respectively. The maximum load difference obtained by SID is always greater than the one obtained by DASUD for all topologies. Moreover, DASUD demonstrates an additional quality: for hypercubes and torus topologies the maximum load difference keeps nearly constant for any system size and load distribution pattern (on the average, the maximum difference was 1.5). For rings, the maximum difference on the average was a somewhat higher and a slight increment was obtained when the initial unbalance was 75% or 100%, but it was always less than 10. By contrast, SID algorithm always obtained a higher maximum difference in every case and, additionally, the difference increases, as the initial unbalance grew higher. 4

5 Figure 3.2 Maximum load difference for SID and DASUD algorithms considering likely initial load distributions, pathological initial load distributions. Likely distributions Pathological distributions NªP Hyper Torus Ring Likely distributions Pathological distributions NªP Hyper Torus Ring Table 3.1 Standard deviation obtained on average by DASUD and SID for likely and pathological distributions Figure 3.3 Influence of the topology: Maximum load difference for SID and DASUD algorithms considering likely initial load distributions and pathological initial load distributions. 5

6 As figure 3.3 shows, the worst results for both strategies were obtained for the ring topology. There are two reason for those results. On the one hand, the ring topology exhibits a small number of neighbours for any processor. As a consequence, the load movement through the network is slowed down. On the other hand, a platform effect appears. As we have already mentioned, the perfect balance is achieved when the maximum difference is one. However, due to the fully distributed nature of our method a local termination condition is achieved when the maximum difference is equal to one or zero for every domain of processors. This behaviour could lead to a situation where loads are finally laid out in a platform fashion. Figure 3.4 shows the platform effect that consists of obtaining a global unbalance caused by the existence of overlapped domains, in spite of achieving locally balanced domains. In the worst case, this effect will spread by the shortest path between two processors located at the maximum distance, i. e. by a path with a distance equal to the diameter of the architecture (d). Table 3.2 shows the value of the diameter for different topologies with different sizes. If we observe the values of the maximum difference obtained, we can derive an upper bound for the maximum difference achieved at the end of the balancing process that depends on the parameter d. We call this bound β (also d shown in table 3.2) and it is equal to: β = Figure 3.4 Global state unbalanced, local domain balanced ( platform effect ). n=nº processors 8(9) 16 32(36) (121) Topol. Diam. d β d β d β d β d β Hyper log n Torus n * 2 Ring n Table 3.2 Diameter of some topologies and its corresponding β bound. 5 d=1 β=1 d=2 d=3 d=4 d= β=1 β=2 β=2 β=3 Figure 3.5 Platform effect of β Figure 3.5 shows the variation of β as the value of d increases. Table 3.3 shows the maximum value for the maximum difference obtained by DASUD in all our tests. As it can be seen, in the worst case DASUD always achieves a maximum difference lower than the corresponding value of β. This means that, even for highly pathological initial distributions in a ring topology where there are a small number of intersections between multiple domains, DASUD is able to obtain a final maximum difference bounded by half of the diameter of the architecture. % load variation Likely distributions idle processors Pathological distributions) % % Hypercube 50% % % % % n % % Torus 50% % % % % n % % Ring 50% % % % % n Table 3.3 Maximum dif_max on average for likely and pathological distributions. 6

7 3.1.b. Influence of the distribution on DASUD s stability. Table 3.4 shows for likely and pathological initial distributions, the influence of the size of the architecture on the final balance when the DASUD algorithm is applied. We observed in our experiments that DASUD has the same behaviour for all topologies, so we only present the results obtained for hypercubes. From these results, we can conclude that as the number of processors increases the maximum difference obtained at the end likewise increases. This result is due to the completely distributed nature of our policy, where only local information from immediate neighbours is used during the balancing process. Moreover, the increment of the maximum difference observed is not very significant; for instance, on average, the maximum difference was always less than 3 when the number of processors was 128 for both likely and pathological distributions. Hypercube (dif_max) Nº likely distributions pathological distributions of P.. 25% 50% 75% 100% 25% 50% 75% n Table 3.4 Maximum load difference for DASUD algorithm considering likely and pathological initial load distributions for hypercubes from 8 to 128 processors. 3.1.c. Influence of the shape on DASUD s stability. As a final consideration for the stability analysis we have observed the results obtained according to the original shape used in the initial load distribution. For all the experiments we always considered two different shapes for every initial load distribution: a Single Mountain (SM) shape and a Chain shape. For all topologies we observed that the final maximum load difference depends on how the load distribution was scattered through the system. Table 3.5 shows this dependency for hypercubes for both likely and pathological distributions and, additionally, takes into account the variation from global load average and the percentage of idle processors respectively. One can observe that for the chain shape initial scattering, the final state obtained is slightly more balanced than the final state obtained when the initial scattering corresponds to the single mountain shape. This behaviour can be explained because with the single mountain shape the platform effect has a great influence. With the chain shape, the workload is scattered onto various high-load areas surrounded by low-load. As a consequence, the number of platforms that appears is low and they have fewer levels. These results were also obtained for ring and torus topologies and we therefore do not include the corresponding tables, as no additional information would thereby be provided. Hypercube (dif_max by shapes) likely distributions pathological distributions 25% 50% 75% 100% 25% 50% 75% n-1 SM Chain Table 3.5 Influence of the shape on the initial distribution: dif_max 3.2 Efficiency analysis. In this section we analyse the efficiency of the DASUD and SID algorithms. The efficiency reflects the time required to either reduce the variance of processors loads or arrive at the equilibrium state. In order to have a measure of the time needed by both strategies to reach the final load distribution, we measure the number of simulation steps to reach a final stable distribution and we introduce the parameter u to measure the load movements incurred in the balancing process. For a given step of simulation process, s, the maximum amount of load moved from any processor to one of its neighbours is called max_load(s). According to our synchronous simulation paradigm, step s will not end until max_load(s) units of loads have been moved from the corresponding processor to its neighbour. Therefore, the duration of each step depends directly on the value of max_load(s). We assume a communication model where a processor is able to communicate with all its nearestneighbours simultaneously. The time required to send one unit of load from one processor to any one of its nearestneighbours is called per-hop time (t h ). So, if we multiply the addition of all max_load(s) where s varies from 1 to the number of simulation steps, by t h, we obtain the total time required for the global simulation process that we call u. u = t * h s= last _ step s= 1 max _ load( s) For simplicity, we assume that t h is equal to one. Figures 3.6 and 3.6 show SID and DASUD efficiency for likely and pathological initial load distributions in terms of u s and simulation steps respectively. These results summarise the time required for both strategies to reach the termination condition, for all topologies. As we can observe, the DASUD algorithm needs more time to achieve the stable final state than the SID algorithm independently of the number of processors. These results were to be logically expected because the DASUD algorithm is an extension of the SID algorithm, which tries to improve the balance degree of the final load distributions by detecting unbalanced domains and arranging these situations by performing some extra load movements. The additional time needed by DASUD is, 7

8 however, moderate on average. And, bearing in mind the results from the stability analysis, we can conclude that DASUD exhibits a better trade-off between the degree of global balance and the time needed to achieve it. Like in the stability analysis, we compare below the efficiency of SID and DASUD with respect to the topology and we give a more detailed information about the influence of the load distribution shape on DASUD s final results. 3.2.a. Influence of the topology on the efficiency. Figures 3.7 and 3.8 give more detailed information about the time in terms of u s and the number of steps needed on average for likely and pathological distributions. As can be observed for torus and hypercubes the time needed by DASUD was moderately higher than the time needed by SID. In particular, for likely distributions the time of DASUD was on average twice the time for SID, but the maximum difference obtained by SID was more than 4 times the maximum difference obtained by DASUD. For pathological distributions the time for DASUD was 30% more than SID, while the maximum difference obtained by SID was more than 7 times the maximum difference obtained by DASUD. In that sense, the improvement of the final load balancing obtained by DASUD was not only due to the increase of the number of steps. DASUD obtains a better final load balancing because it moves more load at each step and takes advantage of the overlapping of load movements. Higher differences between DASUD and SID in the time and number of steps were obtained for rings, and especially for pathological distributions. This could be explained again because of the better load balancing obtained at the end. While SID suffers from the platform effect at the initial steps of the balancing process and is unable to move load to the less loaded processors, DASUD is able to significantly overcome the platform effect and it can move load during several additional steps. However, due to the small connectivity exhibited by ring topologies, the movement of load performed by DASUD is very small once a certain balance degree has been achieved, and the strategy goes through a lot of steps where very few loads are being moved at each step. We can also observe a positive effect of system size on the performance of both strategies for a fixed size of the problem (L). The time required to achieve a stable state decreases as the system size increases. This characteristic remains constant for any percentage of load average variation. Notice that as the number of processors increases, the value of the load average decreases. Consequently, the total number of load units to be moved among the system also decreases. But, in spite of this, both algorithms, DASUD and SID, demonstrate the same behaviour. It has been observed that if the global load average remains constant as the number of processors increases (i.e. the value of L varies as the number of processors increases), the time required to reach the final state keeps more or less constant for any system size. 3.2.b. Influence of the shapes on DASUD s efficiency. The effect of how the initial load distribution is scattered through the processors has also been investigated in order to observe whether it has any influence on the total execution time or not. For all topologies it has been observed that the time behaviour follows the same pattern, for this reason, we only show the results for hypercubes topologies. In table 3.6 we show the time required to reach a stable state using DASUD algorithm on average depending on how the initial load distribution is scattered to processors. Each value is the mean value for all sizes of hypercubes. For 50%, 75% and 100% variations on load average and pathological initial distributions, the value of u for chain scattering is bigger than the one obtained when single mountain scattering is applied. This is attributable to the kind of load movements generated in both cases. When single mountain scattering is applied, all local load movements have the same global direction, from heavily loaded processors to lightly loaded ones, because this kind of load scattering generates a local gradient equal to the global one. Consequently, all local load movements are productive movements. On the other hand, when chain scattering is used, some processors can see themselves as locally load-maximum while not being globally-maximum. This is a consequence of the distributed nature of the algorithm Hypercube Topology (time -u s - by shapes) likely distributions pathological distributions 25% 50% 75% 100% 25% 50% 75% n-1 SM Chain Table 3.6 Influence of the shape on the initial distribution: time incurred in the load movement. 8

9 Figure 3.6 Efficiency in terms of: time incurred in the load movement and number of steps. Figure 3.7 Efficiency results for likely distributions: time incurred in th load movements and number of step. Figure 3.8 Efficiency results for pathol. distributions: time spent in load movement and number of steps 9

10 As we can see in table 3.6, the previous reasoning does not comply with 25% load average variation. In such a situation the maximum load difference between any two processors in the initial load distribution is not too big. This being so, the local load movements generated by any processors tend to be the productive ones and no penalty for unnecessary load thrashing is produced. We have also investigated the influence of the initial load scattering on the number of steps needed by the balance process to reach the termination condition. Table 3.7 shows the average of such number of steps for all sizes of hypercubes. For single mountain shapes, the number of steps is higher than for chain shapes. This characteristic is independent of the initial unbalanced degree and of the kind of distribution (likely or pathological). For all likely distributions the number of steps required for single mountain shapes is approximately twice the number of steps required for chain distributions. Hypercube Topology (steps by shapes) likely distributions pathological distributions 25% 50% 75% 100% 25% 50% 75% n-1 SM Chain Table 3.7 Influence of the shape on the initial distribution: number of steps. Bearing in mind the information set out in tables 3.5, 3.6 and 3.7, we can deduce that, starting from a chain-form collocation, DASUD achieves a more balanced final state that attained from starting with a Single Mountain collocation. And furthermore, it requires a smaller number of steps, since in each step a greater load quantity is moved. 4. Conclusions. In this paper, we have compared two algorithms, DASUD (Diffusion Algorithm Searching Unbalanced Domains) and SID (Sender Initiated Diffusion) for dynamic load balancing in parallel systems. The comparison was carried out by considering a large set of load distributions that exhibit different degrees of initial workload unbalancing as well as different shapes of workload unbalancing. These distributions were applied to ring, torus and hypercube topologies, and the number of processors ranged from 8 to 128. The experiments were conducted to analyse the balancing degree achieved by both strategies at the final state, the time incurred in the load movement and the number of balancing steps. From these experiments we have observed that DASUD outperforms the SID strategy as it provides the best tradeoff between the global balance degree obtained at the final state and the number of iterations required to reach such a state. For the most common topologies (torus and hypercubes) DASUD and SID spent on the average a similar number of balancing steps, while the maximum difference achieved by SID was more than 4 times larger than the maximum difference obtained by DASUD. This behaviour was observed independently of the initial unbalanced degree, the scattering of the loads or the number of processors. Moreover, DASUD not only obtained a smaller value for the maximum difference, but also achieved a better balanced degree for all the processors in the system, as all processors had a final load that was very close to the optimal load average. References [1] G.C.Fox, M.A. Johnson, G.A. Lyzenga,S.W. Otto, J.K. Salmon and D.W. Walkeer, Solving Problems on Concurrent Processors, vol. 1, Prentice-Hall, [2] S. H. Hosseini, B. Litow, M. Malkawi, J. McPherson, and K. Vairavan, Analysis of a Graph Coloring Based Distributed Load Balancing Algorithm, Journal of Parallel and Distributed Computing 10, 1990, pp [3] V. Kumar, A. Y. Grama and N. R. Vempaty, Scalable load balancing techniques for parallel computers, J. of Par. and Distrib. Comput., 22(1), 1994, pp [4] R. Subramain, I. D. Scherson, An Analysis of Diffusive Load-Balancing, In Proceedings of 6th ACM Symposium on Parallel Algorithms and Architectures, 1994 [5] M. Willebeek-LeMair, A. P. Reeves, Strategies for Dynamic Load Balancing on Highly Parallel Computers, IEEE Transactions on Parallel and Distributed Systems, vol. 4, No. 9, September 1993, pp [6] C. Z. Xu and F. C. M. Lau, Load Balancing Parallel Computers - Theory and Practice, Kluwer Academic Publishers, 1997 [7] A. Cortés, A. Ripoll, M. A. Senar and E. Luque, Dynamic Load Balancing Strategy for Scalable Parallel Systems, PARCO 97, [8] A. Cortés, A. Ripoll, M.A.Senar, F. Cedó and E. Luque, On the convergence of SID and DASUD load-balancing algorithms, Technical Report, UAB,

DECENTRALIZED LOAD BALANCING IN HETEROGENEOUS SYSTEMS USING DIFFUSION APPROACH

DECENTRALIZED LOAD BALANCING IN HETEROGENEOUS SYSTEMS USING DIFFUSION APPROACH DECENTRALIZED LOAD BALANCING IN HETEROGENEOUS SYSTEMS USING DIFFUSION APPROACH P.Neelakantan Department of Computer Science & Engineering, SVCET, Chittoor pneelakantan@rediffmail.com ABSTRACT The grid

More information

Load Balancing Between Heterogenous Computing Clusters

Load Balancing Between Heterogenous Computing Clusters Load Balancing Between Heterogenous Computing Clusters Siu-Cheung Chau Dept. of Physics and Computing, Wilfrid Laurier University, Waterloo, Ontario, Canada, N2L 3C5 e-mail: schau@wlu.ca Ada Wai-Chee Fu

More information

Load Balancing between Computing Clusters

Load Balancing between Computing Clusters Load Balancing between Computing Clusters Siu-Cheung Chau Dept. of Physics and Computing, Wilfrid Laurier University, Waterloo, Ontario, Canada, NL 3C5 e-mail: schau@wlu.ca Ada Wai-Chee Fu Dept. of Computer

More information

An Ecient Dynamic Load Balancing using the Dimension Exchange. Ju-wook Jang. of balancing load among processors, most of the realworld

An Ecient Dynamic Load Balancing using the Dimension Exchange. Ju-wook Jang. of balancing load among processors, most of the realworld An Ecient Dynamic Load Balancing using the Dimension Exchange Method for Balancing of Quantized Loads on Hypercube Multiprocessors * Hwakyung Rim Dept. of Computer Science Seoul Korea 11-74 ackyung@arqlab1.sogang.ac.kr

More information

Preserving Message Integrity in Dynamic Process Migration

Preserving Message Integrity in Dynamic Process Migration Preserving Message Integrity in Dynamic Process Migration E. Heymann, F. Tinetti, E. Luque Universidad Autónoma de Barcelona Departamento de Informática 8193 - Bellaterra, Barcelona, Spain e-mail: e.heymann@cc.uab.es

More information

Feedback guided load balancing in a distributed memory environment

Feedback guided load balancing in a distributed memory environment Feedback guided load balancing in a distributed memory environment Constantinos Christofi August 18, 2011 MSc in High Performance Computing The University of Edinburgh Year of Presentation: 2011 Abstract

More information

A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters

A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters Abhijit A. Rajguru, S.S. Apte Abstract - A distributed system can be viewed as a collection

More information

Experiments on the local load balancing algorithms; part 1

Experiments on the local load balancing algorithms; part 1 Experiments on the local load balancing algorithms; part 1 Ştefan Măruşter Institute e-austria Timisoara West University of Timişoara, Romania maruster@info.uvt.ro Abstract. In this paper the influence

More information

Parallel Scalable Algorithms- Performance Parameters

Parallel Scalable Algorithms- Performance Parameters www.bsc.es Parallel Scalable Algorithms- Performance Parameters Vassil Alexandrov, ICREA - Barcelona Supercomputing Center, Spain Overview Sources of Overhead in Parallel Programs Performance Metrics for

More information

Hyper Node Torus: A New Interconnection Network for High Speed Packet Processors

Hyper Node Torus: A New Interconnection Network for High Speed Packet Processors 2011 International Symposium on Computer Networks and Distributed Systems (CNDS), February 23-24, 2011 Hyper Node Torus: A New Interconnection Network for High Speed Packet Processors Atefeh Khosravi,

More information

Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems

Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems G.Rajina #1, P.Nagaraju #2 #1 M.Tech, Computer Science Engineering, TallaPadmavathi Engineering College, Warangal,

More information

A Comparison of General Approaches to Multiprocessor Scheduling

A Comparison of General Approaches to Multiprocessor Scheduling A Comparison of General Approaches to Multiprocessor Scheduling Jing-Chiou Liou AT&T Laboratories Middletown, NJ 0778, USA jing@jolt.mt.att.com Michael A. Palis Department of Computer Science Rutgers University

More information

Distributed Dynamic Load Balancing for Iterative-Stencil Applications

Distributed Dynamic Load Balancing for Iterative-Stencil Applications Distributed Dynamic Load Balancing for Iterative-Stencil Applications G. Dethier 1, P. Marchot 2 and P.A. de Marneffe 1 1 EECS Department, University of Liege, Belgium 2 Chemical Engineering Department,

More information

A New Nature-inspired Algorithm for Load Balancing

A New Nature-inspired Algorithm for Load Balancing A New Nature-inspired Algorithm for Load Balancing Xiang Feng East China University of Science and Technology Shanghai, China 200237 Email: xfeng{@ecusteducn, @cshkuhk} Francis CM Lau The University of

More information

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age.

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age. Volume 3, Issue 10, October 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Load Measurement

More information

APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM

APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM 152 APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM A1.1 INTRODUCTION PPATPAN is implemented in a test bed with five Linux system arranged in a multihop topology. The system is implemented

More information

Grid Computing Approach for Dynamic Load Balancing

Grid Computing Approach for Dynamic Load Balancing International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-1 E-ISSN: 2347-2693 Grid Computing Approach for Dynamic Load Balancing Kapil B. Morey 1*, Sachin B. Jadhav

More information

Load Balancing. Load Balancing 1 / 24

Load Balancing. Load Balancing 1 / 24 Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait

More information

Components: Interconnect Page 1 of 18

Components: Interconnect Page 1 of 18 Components: Interconnect Page 1 of 18 PE to PE interconnect: The most expensive supercomputer component Possible implementations: FULL INTERCONNECTION: The ideal Usually not attainable Each PE has a direct

More information

A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture

A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture Yangsuk Kee Department of Computer Engineering Seoul National University Seoul, 151-742, Korea Soonhoi

More information

A Review of Customized Dynamic Load Balancing for a Network of Workstations

A Review of Customized Dynamic Load Balancing for a Network of Workstations A Review of Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester

More information

International journal of Engineering Research-Online A Peer Reviewed International Journal Articles available online http://www.ijoer.

International journal of Engineering Research-Online A Peer Reviewed International Journal Articles available online http://www.ijoer. RESEARCH ARTICLE ISSN: 2321-7758 GLOBAL LOAD DISTRIBUTION USING SKIP GRAPH, BATON AND CHORD J.K.JEEVITHA, B.KARTHIKA* Information Technology,PSNA College of Engineering & Technology, Dindigul, India Article

More information

Analysis of Micromouse Maze Solving Algorithms

Analysis of Micromouse Maze Solving Algorithms 1 Analysis of Micromouse Maze Solving Algorithms David M. Willardson ECE 557: Learning from Data, Spring 2001 Abstract This project involves a simulation of a mouse that is to find its way through a maze.

More information

A Network Flow Approach in Cloud Computing

A Network Flow Approach in Cloud Computing 1 A Network Flow Approach in Cloud Computing Soheil Feizi, Amy Zhang, Muriel Médard RLE at MIT Abstract In this paper, by using network flow principles, we propose algorithms to address various challenges

More information

Load balancing in a heterogeneous computer system by self-organizing Kohonen network

Load balancing in a heterogeneous computer system by self-organizing Kohonen network Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.

More information

Dynamic Load Balancing in a Network of Workstations

Dynamic Load Balancing in a Network of Workstations Dynamic Load Balancing in a Network of Workstations 95.515F Research Report By: Shahzad Malik (219762) November 29, 2000 Table of Contents 1 Introduction 3 2 Load Balancing 4 2.1 Static Load Balancing

More information

An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems

An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems Ardhendu Mandal and Subhas Chandra Pal Department of Computer Science and Application, University

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

Resource Allocation Schemes for Gang Scheduling

Resource Allocation Schemes for Gang Scheduling Resource Allocation Schemes for Gang Scheduling B. B. Zhou School of Computing and Mathematics Deakin University Geelong, VIC 327, Australia D. Walsh R. P. Brent Department of Computer Science Australian

More information

DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM

DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM 1 Introduction In parallel distributed computing system, due to the lightly loaded and overloaded nodes that cause load imbalance, could affect

More information

A Novel Switch Mechanism for Load Balancing in Public Cloud

A Novel Switch Mechanism for Load Balancing in Public Cloud International OPEN ACCESS Journal Of Modern Engineering Research (IJMER) A Novel Switch Mechanism for Load Balancing in Public Cloud Kalathoti Rambabu 1, M. Chandra Sekhar 2 1 M. Tech (CSE), MVR College

More information

A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems*

A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* Junho Jang, Saeyoung Han, Sungyong Park, and Jihoon Yang Department of Computer Science and Interdisciplinary Program

More information

A Hybrid Load Balancing Policy underlying Cloud Computing Environment

A Hybrid Load Balancing Policy underlying Cloud Computing Environment A Hybrid Load Balancing Policy underlying Cloud Computing Environment S.C. WANG, S.C. TSENG, S.S. WANG*, K.Q. YAN* Chaoyang University of Technology 168, Jifeng E. Rd., Wufeng District, Taichung 41349

More information

A SIMULATOR FOR LOAD BALANCING ANALYSIS IN DISTRIBUTED SYSTEMS

A SIMULATOR FOR LOAD BALANCING ANALYSIS IN DISTRIBUTED SYSTEMS Mihai Horia Zaharia, Florin Leon, Dan Galea (3) A Simulator for Load Balancing Analysis in Distributed Systems in A. Valachi, D. Galea, A. M. Florea, M. Craus (eds.) - Tehnologii informationale, Editura

More information

Load Balancing on Massively Parallel Networks: A Cellular Automaton Approach

Load Balancing on Massively Parallel Networks: A Cellular Automaton Approach Load Balancing on Massively Parallel Networks: A Cellular Automaton Approach by Brian Dickens at the Thomas Jefferson High School for Science and Technology supervising instructor: Donald Hyatt originally

More information

The Liquid Model Load Balancing Method

The Liquid Model Load Balancing Method P APER ACCEPTED FOR THE JOURNAL OF P ARALLEL ALGORITHMS AND APPLICATIONS, SPECIAL ISSUE ON ALGORITHMS FOR ENHANCED MESH ARCHITECTURES The Liquid Model Load Balancing Method Dominik Henrich Institute for

More information

A STUDY OF TASK SCHEDULING IN MULTIPROCESSOR ENVIROMENT Ranjit Rajak 1, C.P.Katti 2, Nidhi Rajak 3

A STUDY OF TASK SCHEDULING IN MULTIPROCESSOR ENVIROMENT Ranjit Rajak 1, C.P.Katti 2, Nidhi Rajak 3 A STUDY OF TASK SCHEDULING IN MULTIPROCESSOR ENVIROMENT Ranjit Rajak 1, C.P.Katti, Nidhi Rajak 1 Department of Computer Science & Applications, Dr.H.S.Gour Central University, Sagar, India, ranjit.jnu@gmail.com

More information

Dynamic Multi-User Load Balancing in Distributed Systems

Dynamic Multi-User Load Balancing in Distributed Systems Dynamic Multi-User Load Balancing in Distributed Systems Satish Penmatsa and Anthony T. Chronopoulos The University of Texas at San Antonio Dept. of Computer Science One UTSA Circle, San Antonio, Texas

More information

Hardware-Aware Analysis and. Presentation Date: Sep 15 th 2009 Chrissie C. Cui

Hardware-Aware Analysis and. Presentation Date: Sep 15 th 2009 Chrissie C. Cui Hardware-Aware Analysis and Optimization of Stable Fluids Presentation Date: Sep 15 th 2009 Chrissie C. Cui Outline Introduction Highlights Flop and Bandwidth Analysis Mehrstellen Schemes Advection Caching

More information

Jan F. Prins. Work-efficient Techniques for the Parallel Execution of Sparse Grid-based Computations TR91-042

Jan F. Prins. Work-efficient Techniques for the Parallel Execution of Sparse Grid-based Computations TR91-042 Work-efficient Techniques for the Parallel Execution of Sparse Grid-based Computations TR91-042 Jan F. Prins The University of North Carolina at Chapel Hill Department of Computer Science CB#3175, Sitterson

More information

Load Balancing in Distributed Data Base and Distributed Computing System

Load Balancing in Distributed Data Base and Distributed Computing System Load Balancing in Distributed Data Base and Distributed Computing System Lovely Arya Research Scholar Dravidian University KUPPAM, ANDHRA PRADESH Abstract With a distributed system, data can be located

More information

RESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009

RESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009 An Algorithm for Dynamic Load Balancing in Distributed Systems with Multiple Supporting Nodes by Exploiting the Interrupt Service Parveen Jain 1, Daya Gupta 2 1,2 Delhi College of Engineering, New Delhi,

More information

Lecture 2 Parallel Programming Platforms

Lecture 2 Parallel Programming Platforms Lecture 2 Parallel Programming Platforms Flynn s Taxonomy In 1966, Michael Flynn classified systems according to numbers of instruction streams and the number of data stream. Data stream Single Multiple

More information

Key Words: Dynamic Load Balancing, and Distributed System

Key Words: Dynamic Load Balancing, and Distributed System DYNAMIC ROTATING LOAD BALANCING ALGORITHM IN DISTRIBUTED SYSTEMS ROSE SULEIMAN AL DAHOUD ALI ISSA OTOUM Al-Zaytoonah University Al-Zaytoonah University Neelain University rosesuleiman@yahoo.com aldahoud@alzaytoonah.edu.jo

More information

Decentralized Utility-based Sensor Network Design

Decentralized Utility-based Sensor Network Design Decentralized Utility-based Sensor Network Design Narayanan Sadagopan and Bhaskar Krishnamachari University of Southern California, Los Angeles, CA 90089-0781, USA narayans@cs.usc.edu, bkrishna@usc.edu

More information

How To Balance In Cloud Computing

How To Balance In Cloud Computing A Review on Load Balancing Algorithms in Cloud Hareesh M J Dept. of CSE, RSET, Kochi hareeshmjoseph@ gmail.com John P Martin Dept. of CSE, RSET, Kochi johnpm12@gmail.com Yedhu Sastri Dept. of IT, RSET,

More information

Load Balancing on a Grid Using Data Characteristics

Load Balancing on a Grid Using Data Characteristics Load Balancing on a Grid Using Data Characteristics Jonathan White and Dale R. Thompson Computer Science and Computer Engineering Department University of Arkansas Fayetteville, AR 72701, USA {jlw09, drt}@uark.edu

More information

CHAPTER 5 WLDMA: A NEW LOAD BALANCING STRATEGY FOR WAN ENVIRONMENT

CHAPTER 5 WLDMA: A NEW LOAD BALANCING STRATEGY FOR WAN ENVIRONMENT 81 CHAPTER 5 WLDMA: A NEW LOAD BALANCING STRATEGY FOR WAN ENVIRONMENT 5.1 INTRODUCTION Distributed Web servers on the Internet require high scalability and availability to provide efficient services to

More information

Various Schemes of Load Balancing in Distributed Systems- A Review

Various Schemes of Load Balancing in Distributed Systems- A Review 741 Various Schemes of Load Balancing in Distributed Systems- A Review Monika Kushwaha Pranveer Singh Institute of Technology Kanpur, U.P. (208020) U.P.T.U., Lucknow Saurabh Gupta Pranveer Singh Institute

More information

THE DESIGN OF AN EFFICIENT LOAD BALANCING ALGORITHM EMPLOYING BLOCK DESIGN. Ilyong Chung and Yongeun Bae. 1. Introduction

THE DESIGN OF AN EFFICIENT LOAD BALANCING ALGORITHM EMPLOYING BLOCK DESIGN. Ilyong Chung and Yongeun Bae. 1. Introduction J. Appl. Math. & Computing Vol. 14(2004), No. 1-2, pp. 343-351 THE DESIGN OF AN EFFICIENT LOAD BALANCING ALGORITHM EMPLOYING BLOCK DESIGN Ilyong Chung and Yongeun Bae Abstract. In order to maintain load

More information

System Interconnect Architectures. Goals and Analysis. Network Properties and Routing. Terminology - 2. Terminology - 1

System Interconnect Architectures. Goals and Analysis. Network Properties and Routing. Terminology - 2. Terminology - 1 System Interconnect Architectures CSCI 8150 Advanced Computer Architecture Hwang, Chapter 2 Program and Network Properties 2.4 System Interconnect Architectures Direct networks for static connections Indirect

More information

How To Compare Load Sharing And Job Scheduling In A Network Of Workstations

How To Compare Load Sharing And Job Scheduling In A Network Of Workstations A COMPARISON OF LOAD SHARING AND JOB SCHEDULING IN A NETWORK OF WORKSTATIONS HELEN D. KARATZA Department of Informatics Aristotle University of Thessaloniki 546 Thessaloniki, GREECE Email: karatza@csd.auth.gr

More information

Topological Properties

Topological Properties Advanced Computer Architecture Topological Properties Routing Distance: Number of links on route Node degree: Number of channels per node Network diameter: Longest minimum routing distance between any

More information

Comparison on Different Load Balancing Algorithms of Peer to Peer Networks

Comparison on Different Load Balancing Algorithms of Peer to Peer Networks Comparison on Different Load Balancing Algorithms of Peer to Peer Networks K.N.Sirisha *, S.Bhagya Rekha M.Tech,Software Engineering Noble college of Engineering & Technology for Women Web Technologies

More information

DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS

DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS Journal homepage: www.mjret.in DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS ISSN:2348-6953 Rahul S. Wankhade, Darshan M. Marathe, Girish P. Nikam, Milind R. Jawale Department of Computer Engineering,

More information

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518 International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 Load Balancing Heterogeneous Request in DHT-based P2P Systems Mrs. Yogita A. Dalvi Dr. R. Shankar Mr. Atesh

More information

CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING

CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING CHAPTER 6 CROSS LAYER BASED MULTIPATH ROUTING FOR LOAD BALANCING 6.1 INTRODUCTION The technical challenges in WMNs are load balancing, optimal routing, fairness, network auto-configuration and mobility

More information

Lecture 4 Online and streaming algorithms for clustering

Lecture 4 Online and streaming algorithms for clustering CSE 291: Geometric algorithms Spring 2013 Lecture 4 Online and streaming algorithms for clustering 4.1 On-line k-clustering To the extent that clustering takes place in the brain, it happens in an on-line

More information

A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster

A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906

More information

SIMULATION OF LOAD BALANCING ALGORITHMS: A Comparative Study

SIMULATION OF LOAD BALANCING ALGORITHMS: A Comparative Study SIMULATION OF LOAD BALANCING ALGORITHMS: A Comparative Study Milan E. Soklic Abstract This article introduces a new load balancing algorithm, called diffusive load balancing, and compares its performance

More information

Interconnection Network

Interconnection Network Interconnection Network Recap: Generic Parallel Architecture A generic modern multiprocessor Network Mem Communication assist (CA) $ P Node: processor(s), memory system, plus communication assist Network

More information

?kt. An Unconventional Method for Load Balancing. w = C ( t m a z - ti) = p(tmaz - 0i=l. 1 Introduction. R. Alan McCoy,*

?kt. An Unconventional Method for Load Balancing. w = C ( t m a z - ti) = p(tmaz - 0i=l. 1 Introduction. R. Alan McCoy,* ENL-62052 An Unconventional Method for Load Balancing Yuefan Deng,* R. Alan McCoy,* Robert B. Marr,t Ronald F. Peierlst Abstract A new method of load balancing is introduced based on the idea of dynamically

More information

Fair Scheduling Algorithm with Dynamic Load Balancing Using In Grid Computing

Fair Scheduling Algorithm with Dynamic Load Balancing Using In Grid Computing Research Inventy: International Journal Of Engineering And Science Vol.2, Issue 10 (April 2013), Pp 53-57 Issn(e): 2278-4721, Issn(p):2319-6483, Www.Researchinventy.Com Fair Scheduling Algorithm with Dynamic

More information

Performance of networks containing both MaxNet and SumNet links

Performance of networks containing both MaxNet and SumNet links Performance of networks containing both MaxNet and SumNet links Lachlan L. H. Andrew and Bartek P. Wydrowski Abstract Both MaxNet and SumNet are distributed congestion control architectures suitable for

More information

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2 On the Traffic Capacity of Cellular Data Networks T. Bonald 1,2, A. Proutière 1,2 1 France Telecom Division R&D, 38-40 rue du Général Leclerc, 92794 Issy-les-Moulineaux, France {thomas.bonald, alexandre.proutiere}@francetelecom.com

More information

Journal of Theoretical and Applied Information Technology 20 th July 2015. Vol.77. No.2 2005-2015 JATIT & LLS. All rights reserved.

Journal of Theoretical and Applied Information Technology 20 th July 2015. Vol.77. No.2 2005-2015 JATIT & LLS. All rights reserved. EFFICIENT LOAD BALANCING USING ANT COLONY OPTIMIZATION MOHAMMAD H. NADIMI-SHAHRAKI, ELNAZ SHAFIGH FARD, FARAMARZ SAFI Department of Computer Engineering, Najafabad branch, Islamic Azad University, Najafabad,

More information

Keywords Load balancing, Dispatcher, Distributed Cluster Server, Static Load balancing, Dynamic Load balancing.

Keywords Load balancing, Dispatcher, Distributed Cluster Server, Static Load balancing, Dynamic Load balancing. Volume 5, Issue 7, July 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Hybrid Algorithm

More information

USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS

USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS Natarajan Meghanathan Jackson State University, 1400 Lynch St, Jackson, MS, USA natarajan.meghanathan@jsums.edu

More information

Change Management in Enterprise IT Systems: Process Modeling and Capacity-optimal Scheduling

Change Management in Enterprise IT Systems: Process Modeling and Capacity-optimal Scheduling Change Management in Enterprise IT Systems: Process Modeling and Capacity-optimal Scheduling Praveen K. Muthusamy, Koushik Kar, Sambit Sahu, Prashant Pradhan and Saswati Sarkar Rensselaer Polytechnic Institute

More information

Energy Constrained Resource Scheduling for Cloud Environment

Energy Constrained Resource Scheduling for Cloud Environment Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering

More information

Joint Optimization of Overlapping Phases in MapReduce

Joint Optimization of Overlapping Phases in MapReduce Joint Optimization of Overlapping Phases in MapReduce Minghong Lin, Li Zhang, Adam Wierman, Jian Tan Abstract MapReduce is a scalable parallel computing framework for big data processing. It exhibits multiple

More information

An Overview of CORBA-Based Load Balancing

An Overview of CORBA-Based Load Balancing An Overview of CORBA-Based Load Balancing Jian Shu, Linlan Liu, Shaowen Song, Member, IEEE Department of Computer Science Nanchang Institute of Aero-Technology,Nanchang, Jiangxi, P.R.China 330034 dylan_cn@yahoo.com

More information

Load Balancing in MapReduce Based on Scalable Cardinality Estimates

Load Balancing in MapReduce Based on Scalable Cardinality Estimates Load Balancing in MapReduce Based on Scalable Cardinality Estimates Benjamin Gufler 1, Nikolaus Augsten #, Angelika Reiser 3, Alfons Kemper 4 Technische Universität München Boltzmannstraße 3, 85748 Garching

More information

PERFORMANCE STUDY AND SIMULATION OF AN ANYCAST PROTOCOL FOR WIRELESS MOBILE AD HOC NETWORKS

PERFORMANCE STUDY AND SIMULATION OF AN ANYCAST PROTOCOL FOR WIRELESS MOBILE AD HOC NETWORKS PERFORMANCE STUDY AND SIMULATION OF AN ANYCAST PROTOCOL FOR WIRELESS MOBILE AD HOC NETWORKS Reza Azizi Engineering Department, Bojnourd Branch, Islamic Azad University, Bojnourd, Iran reza.azizi@bojnourdiau.ac.ir

More information

A Survey Of Various Load Balancing Algorithms In Cloud Computing

A Survey Of Various Load Balancing Algorithms In Cloud Computing A Survey Of Various Load Balancing Algorithms In Cloud Computing Dharmesh Kashyap, Jaydeep Viradiya Abstract: Cloud computing is emerging as a new paradigm for manipulating, configuring, and accessing

More information

Towards a Load Balancing in a Three-level Cloud Computing Network

Towards a Load Balancing in a Three-level Cloud Computing Network Towards a Load Balancing in a Three-level Cloud Computing Network Shu-Ching Wang, Kuo-Qin Yan * (Corresponding author), Wen-Pin Liao and Shun-Sheng Wang Chaoyang University of Technology Taiwan, R.O.C.

More information

MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS

MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS Priyesh Kanungo 1 Professor and Senior Systems Engineer (Computer Centre), School of Computer Science and

More information

CLOUD COMPUTING PARTITIONING ALGORITHM AND LOAD BALANCING ALGORITHM

CLOUD COMPUTING PARTITIONING ALGORITHM AND LOAD BALANCING ALGORITHM CLOUD COMPUTING PARTITIONING ALGORITHM AND LOAD BALANCING ALGORITHM Anisaara Nadaph 1 and Prof. Vikas Maral 2 1 Department of Computer Engineering, K.J College of Engineering and Management Research Pune

More information

Adaptive MAP Selection with Load Balancing Mechanism for the Hierarchical Mobile IPv6

Adaptive MAP Selection with Load Balancing Mechanism for the Hierarchical Mobile IPv6 Tamkang Journal of Science and Engineering, Vol. 12, No. 4, pp. 481 487 (2009) 481 Adaptive MAP Selection with Load Balancing Mechanism for the Hierarchical Mobile IPv6 Ying-Hong Wang, Chih-Peng Hsu* and

More information

Design and Implementation of a Massively Parallel Version of DIRECT

Design and Implementation of a Massively Parallel Version of DIRECT Design and Implementation of a Massively Parallel Version of DIRECT JIAN HE Department of Computer Science, Virginia Polytechnic Institute and State University, Blacksburg, VA 24061, USA ALEX VERSTAK Department

More information

How To Balance In A Distributed System

How To Balance In A Distributed System 6 IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, VOL. 11, NO. 1, JANUARY 2000 How Useful Is Old Information? Michael Mitzenmacher AbstractÐWe consider the problem of load balancing in dynamic distributed

More information

Distributed Load Balancing for Machines Fully Heterogeneous

Distributed Load Balancing for Machines Fully Heterogeneous Internship Report 2 nd of June - 22 th of August 2014 Distributed Load Balancing for Machines Fully Heterogeneous Nathanaël Cheriere nathanael.cheriere@ens-rennes.fr ENS Rennes Academic Year 2013-2014

More information

Decentralized Dynamic Load Balancing: The Particles Approach

Decentralized Dynamic Load Balancing: The Particles Approach appered in: Information Sciences, Vol 84, Issue 1-2 (May 1995) S 115-128 Decentralized Dynamic Load Balancing: The Particles Approach Hans-Ulrich Heiss Department of Informatics and Automation Technical

More information

Analysis of an Artificial Hormone System (Extended abstract)

Analysis of an Artificial Hormone System (Extended abstract) c 2013. This is the author s version of the work. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purpose or for creating

More information

A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster

A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster , pp.11-20 http://dx.doi.org/10.14257/ ijgdc.2014.7.2.02 A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster Kehe Wu 1, Long Chen 2, Shichao Ye 2 and Yi Li 2 1 Beijing

More information

How To Balance A Single Failure With A Single Power Of Two (Sus)

How To Balance A Single Failure With A Single Power Of Two (Sus) Ranking Service Units for Providing and Protecting Highly Available Services with Load Balancing A. Kanso, F. Khendek, A. Hamou-Lhadj Electrical and Computer Engineering Department Concordia University

More information

Proposal of Dynamic Load Balancing Algorithm in Grid System

Proposal of Dynamic Load Balancing Algorithm in Grid System www.ijcsi.org 186 Proposal of Dynamic Load Balancing Algorithm in Grid System Sherihan Abu Elenin Faculty of Computers and Information Mansoura University, Egypt Abstract This paper proposed dynamic load

More information

Optimal Load Balancing in a Beowulf Cluster. Daniel Alan Adams. A Thesis. Submitted to the Faculty WORCESTER POLYTECHNIC INSTITUTE

Optimal Load Balancing in a Beowulf Cluster. Daniel Alan Adams. A Thesis. Submitted to the Faculty WORCESTER POLYTECHNIC INSTITUTE Optimal Load Balancing in a Beowulf Cluster by Daniel Alan Adams A Thesis Submitted to the Faculty of WORCESTER POLYTECHNIC INSTITUTE in partial fulfillment of the requirements for the Degree of Master

More information

Comparative Study of Load Balancing Algorithms

Comparative Study of Load Balancing Algorithms IOSR Journal of Engineering (IOSRJEN) e-issn: 2250-3021, p-issn: 2278-8719 Vol. 3, Issue 3 (Mar. 2013), V2 PP 45-50 Comparative Study of Load Balancing Algorithms Jyoti Vashistha 1, Anant Kumar Jayswal

More information

Performance Evaluation of Mobile Agent-based Dynamic Load Balancing Algorithm

Performance Evaluation of Mobile Agent-based Dynamic Load Balancing Algorithm Performance Evaluation of Mobile -based Dynamic Load Balancing Algorithm MAGDY SAEB, CHERINE FATHY Computer Engineering Department Arab Academy for Science, Technology & Maritime Transport Alexandria,

More information

Bargaining Solutions in a Social Network

Bargaining Solutions in a Social Network Bargaining Solutions in a Social Network Tanmoy Chakraborty and Michael Kearns Department of Computer and Information Science University of Pennsylvania Abstract. We study the concept of bargaining solutions,

More information

A Comparison of Dynamic Load Balancing Algorithms

A Comparison of Dynamic Load Balancing Algorithms A Comparison of Dynamic Load Balancing Algorithms Toufik Taibi 1, Abdelouahab Abid 2 and Engku Fariez Engku Azahan 2 1 College of Information Technology, United Arab Emirates University, P.O. Box 17555,

More information

IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE OPERATORS

IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE OPERATORS Volume 2, No. 3, March 2011 Journal of Global Research in Computer Science RESEARCH PAPER Available Online at www.jgrcs.info IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE

More information

Multi-service Load Balancing in a Heterogeneous Network with Vertical Handover

Multi-service Load Balancing in a Heterogeneous Network with Vertical Handover 1 Multi-service Load Balancing in a Heterogeneous Network with Vertical Handover Jie Xu, Member, IEEE, Yuming Jiang, Member, IEEE, and Andrew Perkis, Member, IEEE Abstract In this paper we investigate

More information

Dynamic load balancing of parallel cellular automata

Dynamic load balancing of parallel cellular automata Dynamic load balancing of parallel cellular automata Marc Mazzariol, Benoit A. Gennart, Roger D. Hersch Ecole Polytechnique Fédérale de Lausanne, EPFL * ABSTRACT We are interested in running in parallel

More information

A Power Efficient QoS Provisioning Architecture for Wireless Ad Hoc Networks

A Power Efficient QoS Provisioning Architecture for Wireless Ad Hoc Networks A Power Efficient QoS Provisioning Architecture for Wireless Ad Hoc Networks Didem Gozupek 1,Symeon Papavassiliou 2, Nirwan Ansari 1, and Jie Yang 1 1 Department of Electrical and Computer Engineering

More information

Adaptive Processor Allocation for Moldable Jobs in Computational Grid

Adaptive Processor Allocation for Moldable Jobs in Computational Grid 10 International Journal of Grid and High Performance Computing, 1(1), 10-21, January-March 2009 Adaptive Processor Allocation for Moldable Jobs in Computational Grid Kuo-Chan Huang, National Taichung

More information

An Optimized Load-balancing Scheduling Method Based on the WLC Algorithm for Cloud Data Centers

An Optimized Load-balancing Scheduling Method Based on the WLC Algorithm for Cloud Data Centers Journal of Computational Information Systems 9: 7 (23) 689 6829 Available at http://www.jofcis.com An Optimized Load-balancing Scheduling Method Based on the WLC Algorithm for Cloud Data Centers Lianying

More information

Stability of QOS. Avinash Varadarajan, Subhransu Maji {avinash,smaji}@cs.berkeley.edu

Stability of QOS. Avinash Varadarajan, Subhransu Maji {avinash,smaji}@cs.berkeley.edu Stability of QOS Avinash Varadarajan, Subhransu Maji {avinash,smaji}@cs.berkeley.edu Abstract Given a choice between two services, rest of the things being equal, it is natural to prefer the one with more

More information