A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture

Size: px
Start display at page:

Download "A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture"

Transcription

1 A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture Yangsuk Kee Department of Computer Engineering Seoul National University Seoul, , Korea Soonhoi Ha* Department of Computer Engineering Seoul National University Seoul, , Korea Abstract The working condition of a multicomputer system based on message passing communication is changeable and unpredictable. Any good algorithm for the system must be adaptive to the dynamic change of working condition. In this paper, we propose a new algorithm called RAS(Reservation And Work stealing). RAS treats load balancing, fault-tolerance and processor selection problems by work stealing and reservation based distribution. According to our experiment on the IBM SP2 with a matrix multiplication program, RAS is proven superior to the weighted factoring under the shared running environment and comparable under the dedicated running environment. Keywords: dynamic load balancing, fault tolerance, processor selection 1 Introduction As the performance of general purpose processors and networks improves, the multicomputer system emerges as a viable platform for high-performance computing. Interprocessor communication of the system is mainly achieved by explicit message passing. Due to the high communication overhead of message passing, only coarse-grain parallelism is exploited on the system. Hence, we are specially aimed at executing data parallel applications which have coarse-grain parallelism. In a time-shared environment in which a system is not dedicated to the program, a program may show very dynamic behavior during its running period. The operating system and other user processes usually interfere the program. The change of loads severely affects the overall performance since the running time of an application depends on the speed of the slowest node. Therefore, balancing loads among nodes is essential to achieving shorter running time. In addition, there are several other considerations in exploiting parallelism on a multicomputer system. First, a multicomputer system, particularly NOW/COW, generally does not guarantee the reliability of nodes. Few previous dynamic load balancing approaches, if any, do allow the failure of nodes and links. To guarantee the termination of a program, an algorithm should tolerate the dynamic failures. Second, it is not always true that the performance increases in proportion with the number of nodes, because high communication overhead allows the master to manage only a limited number of slaves without performance degradation. Superfluous nodes may only raise the scheduling burden of the master. Furthermore, the master may make a wrong decision by sending data to slower nodes. Therefore, it is necessary to select faster nodes up to as many as a certain threshold. We propose a new method, RAS (Reservation And work Stealing), to exploit coarsegrain data parallelism on a multicomputer system by solving load-balancing, fault-tolerance and node selection problems at the same time. In this paper, we compare RAS with the weighted factoring algorithm with a data parallel application, matrix multiplication. Our

2 experiments are performed on the IBM SP2 using MPI. We first review some prior approaches to dynamic load-balancing problem on several systems. Then, we present the proposed algorithm in details. Last, we report our experimental results, which are used to demonstrate the robustness of our algorithm under various running conditions. 2 Previous Works The simplest way to share N data among P processors is to divide the whole data into N P chunks. However, this method is not appropriate for a multicomputer system, because it does not consider the dynamic change of running environment and communication overhead. Hence, several approaches have been proposed to solve the problems. One approach is to balance workloads by reducing the amount of data as the scheduling progresses. GSS [1] allocates 1 P of the remaining data to the next idle node. The size of chunk for the ith request is G i = (1 1 P )i N P. Factoring [2] schedules data in batches of P chunks sized equally. In practice, the total number of data per batch is half the remaining data. The size of chunks in the ith batch is F i = ( 1 2 )i+1 N P. Originally both algorithms are proposed to a shared memory system. When both algorithms are applied to a multicomputer system, large chunks assigned to the slowest node at the beginning of the computation can delay the overall finishing time, because the remaining data may not be sufficient to achieve a good load balance. Weighed factoring [3] is a variant of factoring which was proposed for multicomputer systems. Weighted factoring is identical to factoring except that the size of chunks in a batch is determined in proportion to the weight(w j ) of nodes. The size of the jth chunk in the ith batch is F ij = ( 1 2 )i+1 N W j k=p. Even k=1 W k though the weighted factoring considered heterogeneity of nodes, it still can not reduce the Algorithm Chunk Size Static Chunk GSS Factoring Weighted Factoring (P=4, N=100, Weights are ) Table 1: Sample chunk sizes effect of superfluous nodes as mentioned in [3]. Table 1 illustrates the sample chunk sizes according to these algorithms. The other approach is work stealing wherein work migrates dynamically from heavily loaded nodes to lightly loaded ones. Hybrid Scheduling [4] combines a static scheduling method with a dynamic one. The master evenly distributes data to the slaves at the static level. Then, the slaves divide the given data into chunks using an algorithm like GSS at the dynamic level. After the execution of each chunk, all slaves periodically send messages to the master to inform of their performance. When the master detects certain imbalance of workloads, it sends direction to the slaves to redistribute the data. Hamdi s approach [5] is similar to hybrid scheduling. The master periodically polls the slaves to collect workload information and partitions data into several regions. To avoid unnecessary communication, a slave sends a request for the data in the region to the slave who has the data, when the requesting slave has no data to compute. Both algorithms experience high scheduling overhead to collect load information, because lock-step synchronization and periodic polling are burden-some operations. 3 RAS Algorithm We propose a new adaptive algorithm, RAS, for multicomputer systems. To reduce or hide communication overhead, RAS adopts overlapped communication. In a message passing environment, sending a large message is desirable compared with multiple small messages, because it amortizes high start-up cost. However, as shown in figure 1, large message delays the start time of the slaves. In order to achieve high performance, there has to be

3 a trade-off between the start-up cost and the start time. Communication overhead is proportional to the size of message, if the size is large enough [6]. From an experiment as shown in figure 2, we can find the point on which linearity is broken; the message larger than 1KB gives us relatively early start time without significant overhead. slaves and redistributes data when load imbalance is detected. As the master is responsible for both scheduling and distribution of data, RAS allows several slaves to compute a data redundantly. This mechanism makes RAS tolerable to the failure of slaves. Figure 3: RAS algorithm flowchart Figure 1: Large message vs. Small message 3.1 Performance Estimation of Slaves Figure 2: Overhead per 1KB of MPI Send on SP2 RAS consists of three phases as shown in figure 3. In the first phase, the master estimates the performance of slaves in terms of their computation latency. After the estimation, the master distributes data on the basis of the performance. In this distribution phase, the master can select faster slaves in the slave pool as many as profitably used. In the last phase, the master monitors the workloads of The performance estimation of slaves should be carried out in advance to schedule data precisely. We divide the whole data into basic chunks which are the unit of computation. In case of matrix multiplication, one row or column can be a basic chunk. If the size of a basic chunk is smaller than 1KB, several basic chunks are packed into a basic message to reduce the communication overhead. The size of a basic message is determined as 1KB S basic S basic, where S basic denotes the size of a basic chunk in bytes. When RAS starts up, the master sends one basic message to each slave for estimation. The first basic message sent to each slave is called a probe. The average execution time of a basic message is used to represent the performance. The master wants the slaves to report their performance every computation of a basic message for accurate estimation. However, frequent reports will be burden-some to both the slaves and the network. As a compromise, a slave reports the result of computation piggybacked

4 with its average execution time, when it meets the following conditions. When a slave finishs computation of the probe. When a slave receives the End Of Data Distribution message. When a slave produces results larger than 1KB. When the master interrupts a slave to steal data. When a slave has no data to compute. The estimation is used both in the distribution phase and in the work stealing phase. The master continues to check the arrival of reply to the probe until the end of the distribution phase, if not all the slaves report their performances. 3.2 Data Distribution on the Basis of Reservation In the distribution phase, the master sends data in the form of basic message. The master converts the average execution time into the sending period of basic message and distributes basic messages on the basis of the period. The sending period(p i ) of the ith slave is deter- Texec(i) T overhead, where T exec (i) denotes the mined as average execution time of a basic message in the ith slave and T overhead does the overhead of sending operation for a basic message. The number of MPI Send operations is used to represent the sending period. When the ith basic message is sent to the jth slave, the (i+p j )th basic message is reserved for the slave. There are two possible ambiguous cases during this reservation. One is the case that several slaves try to reserve the same data. In this case, the master selects the fastest slave and sends the data to the slave. The other slaves who fail to get the data reserve the next data. The other is the case that no slave reserves a data. This case is divided into two sub-cases. When there is no slave who replies Figure 4: Scenario of data distribution based on reservation the result of the probe, the master sends data in a round-robin fashion to all slaves. When there are more than one slaves that responded to the result of the probe, the master sends data in a round-robin fashion to the responded slaves. These two cases make some slaves receive superfluous data. The master does not use these superfluous data for future reservation. This distribution scheme enables the master to select faster slaves. When the number of slaves exceeds a certain threshold, all data are reserved by some slaves. Since the master sends data to faster slaves, slower ones repeatedly fail to get data. A scenario of data distribution based on the reservation is shown in figure 4, where the sending periods of slaves are 6,9,5 and 7, respectively. The first four data M1, M2, M3 and M4 are probes which are used for the performance estimation. Since M5 and M6 are not reserved, the master distributes them in a round-robin fashion to slave1 and slave2. As the master receives a reply from slave1 during sending M7, it sends M8 to slave1 and reserves M14 for slave1. After that, the master receives a reply from slave3, and it sends M9 to slave3 and reserves M14 for slave3. In case of M14, there are two slaves which reserve the data. The master selects the fastest slave, slave3 and sends the data to the slave. The slave1 who fails to get the data, reserves the next data

5 M Work Stealing Even though the master tries to distribute data in proportion to the speed of slaves, dynamic change of running condition makes the finishing time of slaves inconsistent. Superfluous data distributed in a round-robin style can be another source of load imbalance. The master redistributes data by work stealing when it detects any significant load imbalance. When a slave S i has no basic chunks to compute, the master selects the slave S M who has the most amount of the remaining basic chunks. The master steals a portion of the basic chunks from S M andsendthemto S i. The amount of basic chunks to steal is T given by exec(m) T D exec(i)+t exec(m) M, where D M is the amount of the remaining basic chunks of the richest slave. When the amount of stealing basic chunks is less than 1, the master redundantly sends one basic chunk of S M to S i to avoid the interrupt overhead of the richest slave. When a slave is heavily loaded, the slave can not reply its computational result. In that case, the master can send the duplicated data, which was sent to the slave, to other slaves. With this mechanism, RAS can be tolerant to the failure of slaves only if the master is alive. We adopted a software interrupt to implement work stealing mechanism. As MPI dose not define the interrupt function between processes running on different nodes, we implement a interrupt using asynchronous I/O with socket [7]. 4 Experimental Result We compare RAS with the weighted factoring(wf) under two running environment, dedicated and shared running environment. We check the net elapsed time of matrix multiplication excluding the initialization time of MPI. To make a shared running condition, we execute a simple parallel version of matrix multiplication program as a background process. We take the mean value of running times after 10 executions of both programs. When we refer to the number of processors, it means the total number of processors allocated including the master. The result shows that RAS is a little better than the weighted factoring even under the dedicated running condition and superior under the shared running condition. It means that adaptation to the dynamic change of working condition is important, even though it requires additional scheduling overhead. Note the results of the experiment with 16 processors shown in figure 7 and figure 8. RAS balances loads very adaptively under the dynamic environment, while the weighted factoring suffers from load imbalance. The main reason is that the master sends too many data to the slower slaves. It is useful to observe the number of stealing, since RAS balances loads by work stealing. Table 2 and table 3 show the number of stealing under the various configurations. The number of interrupts excluding the number of redundant computations is used to represent the number of stealing. The number of stealing implies the degree of decision inaccuracy. If the master distributes data on the basis of correct information, the number of stealing decreases. The number of stealing is related to the number of processors and the variance of environment. When the master schedules data incorrectly, it tries to balance loads with additional stealing. As expected, the number of stealing increases as the number of processors grows and the environment becomes more dynamic. To test real fault tolerance, we must rewrite the error handling routines of MPI. For simplicity, we modeled a sudden failure or a surge of load of a node with infinite loop. We let a slave fall into infinite loop after program execution. Then the master can not obtain any information about the slave. As the master is responsible for distributing data, it redundantly sends the data sent to the slave to others. Even though the weight factoring was not terminated, RAS was successfully terminated

6 Figure 5: 100x100 matrix multiplication under the dedicated running environment Figure 7: 100x100 matrix multiplication under the shared running environment Figure 6: 200x200 matrix multiplication under the dedicated running environment Figure 8: 200x200 matrix multiplication under the shared running environment without significant performance degradation. In this paper, we applied RAS to parallelizing a matrix multiplication problem. RAS showed good performance under both dedicated and shared running environments, even though the degree of difference is changeable according to the running environment. Furthermore, RAS was fault-tolerant and adaptive to the number of slaves. For more rigorous evaluation, it is necessary to execute various applications. We also have a plan to apply RAS to PC cluster and cluster of workstations. In addition, it is necessary to give user friedly interface of library for application programmer. matrix size number of processors x x Table 2: The number of stealing under the dedicated running environment matrix size number of processors x x Table 3: The number of stealing under the shared running environment

7 References [1] Constantine D. Polychronopoulos and David J. Kuck. Guided Self-scheduling: A Practical Scheduling Scheme for Parallel Supercomputers. IEEE Transaction on Computers, Vol.C-36, No.12: , DEC [2] Susan Flynn Hummel, Edith Schonberg, and Lawrence E.Flynn. Factoring: A Method for Scheduling Parallel Loops. Communication of the ACM, Vol.35, No.8:90 101, Aug [3] Susan Flynn Hummel, Jeanette Schmidt, R.N.Uma, and Joel Wein. Load-Sharing in Heterogeneous Systems via Weighted Factoring. SPAA, pages , [4] Oscar Plata and Francisco F. Rivera. Combining Static and Dynamic Scheduling on Distributed-Memory Multiprocessors. ICS, pages , [5] Mounir Hamdi and Chi-Kin Lee. Dynamic Load Balancing of Data Parallel Applications on a Distributed Network. ICS, pages , [6] Zhiwei Xu and Kai Hwang. Modeling Communication Overhead: MPI and MPL Performance on the IBM SP2. IEEE Parallel and Distributed Technology, Spring:9 23, [7] Message Passing Interface Forum. MPI : A Message-Passing Interface Standard. May, 1994.

A Review of Customized Dynamic Load Balancing for a Network of Workstations

A Review of Customized Dynamic Load Balancing for a Network of Workstations A Review of Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester

More information

Dynamic Load Balancing in a Network of Workstations

Dynamic Load Balancing in a Network of Workstations Dynamic Load Balancing in a Network of Workstations 95.515F Research Report By: Shahzad Malik (219762) November 29, 2000 Table of Contents 1 Introduction 3 2 Load Balancing 4 2.1 Static Load Balancing

More information

High Performance Cluster Support for NLB on Window

High Performance Cluster Support for NLB on Window High Performance Cluster Support for NLB on Window [1]Arvind Rathi, [2] Kirti, [3] Neelam [1]M.Tech Student, Department of CSE, GITM, Gurgaon Haryana (India) arvindrathi88@gmail.com [2]Asst. Professor,

More information

Locality-Preserving Dynamic Load Balancing for Data-Parallel Applications on Distributed-Memory Multiprocessors

Locality-Preserving Dynamic Load Balancing for Data-Parallel Applications on Distributed-Memory Multiprocessors JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 18, 1037-1048 (2002) Short Paper Locality-Preserving Dynamic Load Balancing for Data-Parallel Applications on Distributed-Memory Multiprocessors PANGFENG

More information

Load balancing. David Bindel. 12 Nov 2015

Load balancing. David Bindel. 12 Nov 2015 Load balancing David Bindel 12 Nov 2015 Inefficiencies in parallel code Poor single processor performance Typically in the memory system Saw this in matrix multiply assignment Overhead for parallelism

More information

PERFORMANCE EVALUATION OF THREE DYNAMIC LOAD BALANCING ALGORITHMS ON SPMD MODEL

PERFORMANCE EVALUATION OF THREE DYNAMIC LOAD BALANCING ALGORITHMS ON SPMD MODEL PERFORMANCE EVALUATION OF THREE DYNAMIC LOAD BALANCING ALGORITHMS ON SPMD MODEL Najib A. Kofahi Associate Professor Department of Computer Sciences Faculty of Information Technology and Computer Sciences

More information

Hectiling: An Integration of Fine and Coarse Grained Load Balancing Strategies 1

Hectiling: An Integration of Fine and Coarse Grained Load Balancing Strategies 1 Copyright 1998 IEEE. Published in the Proceedings of HPDC 7 98, 28 31 July 1998 at Chicago, Illinois. Personal use of this material is permitted. However, permission to reprint/republish this material

More information

Cellular Computing on a Linux Cluster

Cellular Computing on a Linux Cluster Cellular Computing on a Linux Cluster Alexei Agueev, Bernd Däne, Wolfgang Fengler TU Ilmenau, Department of Computer Architecture Topics 1. Cellular Computing 2. The Experiment 3. Experimental Results

More information

RevoScaleR Speed and Scalability

RevoScaleR Speed and Scalability EXECUTIVE WHITE PAPER RevoScaleR Speed and Scalability By Lee Edlefsen Ph.D., Chief Scientist, Revolution Analytics Abstract RevoScaleR, the Big Data predictive analytics library included with Revolution

More information

A Dynamic Approach for Load Balancing using Clusters

A Dynamic Approach for Load Balancing using Clusters A Dynamic Approach for Load Balancing using Clusters ShwetaRajani 1, RenuBagoria 2 Computer Science 1,2,Global Technical Campus, Jaipur 1,JaganNath University, Jaipur 2 Email: shwetarajani28@yahoo.in 1

More information

Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6

Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6 Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6 Winter Term 2008 / 2009 Jun.-Prof. Dr. André Brinkmann Andre.Brinkmann@uni-paderborn.de Universität Paderborn PC² Agenda Multiprocessor and

More information

An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems

An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems Ardhendu Mandal and Subhas Chandra Pal Department of Computer Science and Application, University

More information

Customized Dynamic Load Balancing for a Network of Workstations

Customized Dynamic Load Balancing for a Network of Workstations Customized Dynamic Load Balancing for a Network of Workstations Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester, Rochester NY 4627 zaki,wei,srini

More information

Load balancing. Prof. Richard Vuduc Georgia Institute of Technology CSE/CS 8803 PNA: Parallel Numerical Algorithms [L.26] Thursday, April 17, 2008

Load balancing. Prof. Richard Vuduc Georgia Institute of Technology CSE/CS 8803 PNA: Parallel Numerical Algorithms [L.26] Thursday, April 17, 2008 Load balancing Prof. Richard Vuduc Georgia Institute of Technology CSE/CS 8803 PNA: Parallel Numerical Algorithms [L.26] Thursday, April 17, 2008 1 Today s sources CS 194/267 at UCB (Yelick/Demmel) Intro

More information

A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment

A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed

More information

Client/Server Computing Distributed Processing, Client/Server, and Clusters

Client/Server Computing Distributed Processing, Client/Server, and Clusters Client/Server Computing Distributed Processing, Client/Server, and Clusters Chapter 13 Client machines are generally single-user PCs or workstations that provide a highly userfriendly interface to the

More information

Overlapping Data Transfer With Application Execution on Clusters

Overlapping Data Transfer With Application Execution on Clusters Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm reid@cs.toronto.edu stumm@eecg.toronto.edu Department of Computer Science Department of Electrical and Computer

More information

Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach

Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach S. M. Ashraful Kadir 1 and Tazrian Khan 2 1 Scientific Computing, Royal Institute of Technology (KTH), Stockholm, Sweden smakadir@csc.kth.se,

More information

The International Journal Of Science & Technoledge (ISSN 2321 919X) www.theijst.com

The International Journal Of Science & Technoledge (ISSN 2321 919X) www.theijst.com THE INTERNATIONAL JOURNAL OF SCIENCE & TECHNOLEDGE Efficient Parallel Processing on Public Cloud Servers using Load Balancing Manjunath K. C. M.Tech IV Sem, Department of CSE, SEA College of Engineering

More information

MOSIX: High performance Linux farm

MOSIX: High performance Linux farm MOSIX: High performance Linux farm Paolo Mastroserio [mastroserio@na.infn.it] Francesco Maria Taurino [taurino@na.infn.it] Gennaro Tortone [tortone@na.infn.it] Napoli Index overview on Linux farm farm

More information

Operatin g Systems: Internals and Design Principle s. Chapter 10 Multiprocessor and Real-Time Scheduling Seventh Edition By William Stallings

Operatin g Systems: Internals and Design Principle s. Chapter 10 Multiprocessor and Real-Time Scheduling Seventh Edition By William Stallings Operatin g Systems: Internals and Design Principle s Chapter 10 Multiprocessor and Real-Time Scheduling Seventh Edition By William Stallings Operating Systems: Internals and Design Principles Bear in mind,

More information

Load Balancing In Concurrent Parallel Applications

Load Balancing In Concurrent Parallel Applications Load Balancing In Concurrent Parallel Applications Jeff Figler Rochester Institute of Technology Computer Engineering Department Rochester, New York 14623 May 1999 Abstract A parallel concurrent application

More information

A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters

A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters Abhijit A. Rajguru, S.S. Apte Abstract - A distributed system can be viewed as a collection

More information

A Locally Cache-Coherent Multiprocessor Architecture

A Locally Cache-Coherent Multiprocessor Architecture A Locally Cache-Coherent Multiprocessor Architecture Kevin Rich Computing Research Group Lawrence Livermore National Laboratory Livermore, CA 94551 Norman Matloff Division of Computer Science University

More information

Customized Dynamic Load Balancing for a Network of Workstations 1

Customized Dynamic Load Balancing for a Network of Workstations 1 JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING 43, 156 162 (1997) ARTICLE NO. PC971339 Customized Dynamic Load Balancing for a Network of Workstations 1 Mohammed Javeed Zaki, Wei Li, and Srinivasan Parthasarathy

More information

Fault Tolerance in Hadoop for Work Migration

Fault Tolerance in Hadoop for Work Migration 1 Fault Tolerance in Hadoop for Work Migration Shivaraman Janakiraman Indiana University Bloomington ABSTRACT Hadoop is a framework that runs applications on large clusters which are built on numerous

More information

{emery,browne}@cs.utexas.edu ABSTRACT. Keywords scalable, load distribution, load balancing, work stealing

{emery,browne}@cs.utexas.edu ABSTRACT. Keywords scalable, load distribution, load balancing, work stealing Scalable Load Distribution and Load Balancing for Dynamic Parallel Programs E. Berger and J. C. Browne Department of Computer Science University of Texas at Austin Austin, Texas 78701 USA 01-512-471-{9734,9579}

More information

TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance

TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance M. Rangarajan, A. Bohra, K. Banerjee, E.V. Carrera, R. Bianchini, L. Iftode, W. Zwaenepoel. Presented

More information

Load Balancing Algorithms for Peer to Peer and Client Server Distributed Environments

Load Balancing Algorithms for Peer to Peer and Client Server Distributed Environments Load Balancing Algorithms for Peer to Peer and Client Server Distributed Environments Sameena Naaz Afshar Alam Ranjit Biswas Department of Computer Science Jamia Hamdard, New Delhi, India ABSTRACT Advancements

More information

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age.

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age. Volume 3, Issue 10, October 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Load Measurement

More information

Outline. Distributed DBMS

Outline. Distributed DBMS Outline Introduction Background Architecture Distributed Database Design Semantic Data Control Distributed Query Processing Distributed Transaction Management Data server approach Parallel architectures

More information

Load Balancing in Distributed Data Base and Distributed Computing System

Load Balancing in Distributed Data Base and Distributed Computing System Load Balancing in Distributed Data Base and Distributed Computing System Lovely Arya Research Scholar Dravidian University KUPPAM, ANDHRA PRADESH Abstract With a distributed system, data can be located

More information

Comparison on Different Load Balancing Algorithms of Peer to Peer Networks

Comparison on Different Load Balancing Algorithms of Peer to Peer Networks Comparison on Different Load Balancing Algorithms of Peer to Peer Networks K.N.Sirisha *, S.Bhagya Rekha M.Tech,Software Engineering Noble college of Engineering & Technology for Women Web Technologies

More information

DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS

DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS Journal homepage: www.mjret.in DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS ISSN:2348-6953 Rahul S. Wankhade, Darshan M. Marathe, Girish P. Nikam, Milind R. Jawale Department of Computer Engineering,

More information

Scalability and Classifications

Scalability and Classifications Scalability and Classifications 1 Types of Parallel Computers MIMD and SIMD classifications shared and distributed memory multicomputers distributed shared memory computers 2 Network Topologies static

More information

Resource Allocation Schemes for Gang Scheduling

Resource Allocation Schemes for Gang Scheduling Resource Allocation Schemes for Gang Scheduling B. B. Zhou School of Computing and Mathematics Deakin University Geelong, VIC 327, Australia D. Walsh R. P. Brent Department of Computer Science Australian

More information

Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems

Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems G.Rajina #1, P.Nagaraju #2 #1 M.Tech, Computer Science Engineering, TallaPadmavathi Engineering College, Warangal,

More information

A Review on an Algorithm for Dynamic Load Balancing in Distributed Network with Multiple Supporting Nodes with Interrupt Service

A Review on an Algorithm for Dynamic Load Balancing in Distributed Network with Multiple Supporting Nodes with Interrupt Service A Review on an Algorithm for Dynamic Load Balancing in Distributed Network with Multiple Supporting Nodes with Interrupt Service Payal Malekar 1, Prof. Jagruti S. Wankhede 2 Student, Information Technology,

More information

Load Balancing of Web Server System Using Service Queue Length

Load Balancing of Web Server System Using Service Queue Length Load Balancing of Web Server System Using Service Queue Length Brajendra Kumar 1, Dr. Vineet Richhariya 2 1 M.tech Scholar (CSE) LNCT, Bhopal 2 HOD (CSE), LNCT, Bhopal Abstract- In this paper, we describe

More information

Chapter 18: Database System Architectures. Centralized Systems

Chapter 18: Database System Architectures. Centralized Systems Chapter 18: Database System Architectures! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types 18.1 Centralized Systems! Run on a single computer system and

More information

Improved Hybrid Dynamic Load Balancing Algorithm for Distributed Environment

Improved Hybrid Dynamic Load Balancing Algorithm for Distributed Environment International Journal of Scientific and Research Publications, Volume 3, Issue 3, March 2013 1 Improved Hybrid Dynamic Load Balancing Algorithm for Distributed Environment UrjashreePatil*, RajashreeShedge**

More information

Grid Computing Approach for Dynamic Load Balancing

Grid Computing Approach for Dynamic Load Balancing International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-1 E-ISSN: 2347-2693 Grid Computing Approach for Dynamic Load Balancing Kapil B. Morey 1*, Sachin B. Jadhav

More information

Load Balancing in Fault Tolerant Video Server

Load Balancing in Fault Tolerant Video Server Load Balancing in Fault Tolerant Video Server # D. N. Sujatha*, Girish K*, Rashmi B*, Venugopal K. R*, L. M. Patnaik** *Department of Computer Science and Engineering University Visvesvaraya College of

More information

Thread level parallelism

Thread level parallelism Thread level parallelism ILP is used in straight line code or loops Cache miss (off-chip cache and main memory) is unlikely to be hidden using ILP. Thread level parallelism is used instead. Thread: process

More information

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next

More information

Contributions to Gang Scheduling

Contributions to Gang Scheduling CHAPTER 7 Contributions to Gang Scheduling In this Chapter, we present two techniques to improve Gang Scheduling policies by adopting the ideas of this Thesis. The first one, Performance- Driven Gang Scheduling,

More information

A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems*

A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* Junho Jang, Saeyoung Han, Sungyong Park, and Jihoon Yang Department of Computer Science and Interdisciplinary Program

More information

Chapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup

Chapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Chapter 12: Multiprocessor Architectures Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Objective Be familiar with basic multiprocessor architectures and be able to

More information

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518 International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 Load Balancing Heterogeneous Request in DHT-based P2P Systems Mrs. Yogita A. Dalvi Dr. R. Shankar Mr. Atesh

More information

How To Understand The Concept Of A Distributed System

How To Understand The Concept Of A Distributed System Distributed Operating Systems Introduction Ewa Niewiadomska-Szynkiewicz and Adam Kozakiewicz ens@ia.pw.edu.pl, akozakie@ia.pw.edu.pl Institute of Control and Computation Engineering Warsaw University of

More information

Various Schemes of Load Balancing in Distributed Systems- A Review

Various Schemes of Load Balancing in Distributed Systems- A Review 741 Various Schemes of Load Balancing in Distributed Systems- A Review Monika Kushwaha Pranveer Singh Institute of Technology Kanpur, U.P. (208020) U.P.T.U., Lucknow Saurabh Gupta Pranveer Singh Institute

More information

Source Code Transformations Strategies to Load-balance Grid Applications

Source Code Transformations Strategies to Load-balance Grid Applications Source Code Transformations Strategies to Load-balance Grid Applications Romaric David, Stéphane Genaud, Arnaud Giersch, Benjamin Schwarz, and Éric Violard LSIIT-ICPS, Université Louis Pasteur, Bd S. Brant,

More information

A Flexible Cluster Infrastructure for Systems Research and Software Development

A Flexible Cluster Infrastructure for Systems Research and Software Development Award Number: CNS-551555 Title: CRI: Acquisition of an InfiniBand Cluster with SMP Nodes Institution: Florida State University PIs: Xin Yuan, Robert van Engelen, Kartik Gopalan A Flexible Cluster Infrastructure

More information

RESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009

RESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009 An Algorithm for Dynamic Load Balancing in Distributed Systems with Multiple Supporting Nodes by Exploiting the Interrupt Service Parveen Jain 1, Daya Gupta 2 1,2 Delhi College of Engineering, New Delhi,

More information

The Probabilistic Model of Cloud Computing

The Probabilistic Model of Cloud Computing A probabilistic multi-tenant model for virtual machine mapping in cloud systems Zhuoyao Wang, Majeed M. Hayat, Nasir Ghani, and Khaled B. Shaban Department of Electrical and Computer Engineering, University

More information

A novel load balancing algorithm for computational grid

A novel load balancing algorithm for computational grid International Journal of Computational Intelligence Techniques, ISSN: 0976 0466 & E-ISSN: 0976 0474 Volume 1, Issue 1, 2010, PP-20-26 A novel load balancing algorithm for computational grid Saravanakumar

More information

Centralized Systems. A Centralized Computer System. Chapter 18: Database System Architectures

Centralized Systems. A Centralized Computer System. Chapter 18: Database System Architectures Chapter 18: Database System Architectures Centralized Systems! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types! Run on a single computer system and do

More information

Symmetric Multiprocessing

Symmetric Multiprocessing Multicore Computing A multi-core processor is a processing system composed of two or more independent cores. One can describe it as an integrated circuit to which two or more individual processors (called

More information

Efficient Parallel Processing on Public Cloud Servers Using Load Balancing

Efficient Parallel Processing on Public Cloud Servers Using Load Balancing Efficient Parallel Processing on Public Cloud Servers Using Load Balancing Valluripalli Srinath 1, Sudheer Shetty 2 1 M.Tech IV Sem CSE, Sahyadri College of Engineering & Management, Mangalore. 2 Asso.

More information

FAULT TOLERANCE FOR MULTIPROCESSOR SYSTEMS VIA TIME REDUNDANT TASK SCHEDULING

FAULT TOLERANCE FOR MULTIPROCESSOR SYSTEMS VIA TIME REDUNDANT TASK SCHEDULING FAULT TOLERANCE FOR MULTIPROCESSOR SYSTEMS VIA TIME REDUNDANT TASK SCHEDULING Hussain Al-Asaad and Alireza Sarvi Department of Electrical & Computer Engineering University of California Davis, CA, U.S.A.

More information

A Comparison of General Approaches to Multiprocessor Scheduling

A Comparison of General Approaches to Multiprocessor Scheduling A Comparison of General Approaches to Multiprocessor Scheduling Jing-Chiou Liou AT&T Laboratories Middletown, NJ 0778, USA jing@jolt.mt.att.com Michael A. Palis Department of Computer Science Rutgers University

More information

Overview and History of Operating Systems

Overview and History of Operating Systems Overview and History of Operating Systems These are the notes for lecture 1. Please review the Syllabus notes before these. Overview / Historical Developments An Operating System... Sits between hardware

More information

Keywords Load balancing, Dispatcher, Distributed Cluster Server, Static Load balancing, Dynamic Load balancing.

Keywords Load balancing, Dispatcher, Distributed Cluster Server, Static Load balancing, Dynamic Load balancing. Volume 5, Issue 7, July 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Hybrid Algorithm

More information

A Survey on Load Balancing Algorithms in Cloud Environment

A Survey on Load Balancing Algorithms in Cloud Environment A Survey on Load s in Cloud Environment M.Aruna Assistant Professor (Sr.G)/CSE Erode Sengunthar Engineering College, Thudupathi, Erode, India D.Bhanu, Ph.D Associate Professor Sri Krishna College of Engineering

More information

Optimizing the Virtual Data Center

Optimizing the Virtual Data Center Optimizing the Virtual Center The ideal virtual data center dynamically balances workloads across a computing cluster and redistributes hardware resources among clusters in response to changing needs.

More information

Kappa: A system for Linux P2P Load Balancing and Transparent Process Migration

Kappa: A system for Linux P2P Load Balancing and Transparent Process Migration Kappa: A system for Linux P2P Load Balancing and Transparent Process Migration Gaurav Mogre gaurav.mogre@gmail.com Avinash Hanumanthappa avinash947@gmail.com Alwyn Roshan Pais alwyn@nitk.ac.in Abstract

More information

Performance Metrics and Scalability Analysis. Performance Metrics and Scalability Analysis

Performance Metrics and Scalability Analysis. Performance Metrics and Scalability Analysis Performance Metrics and Scalability Analysis 1 Performance Metrics and Scalability Analysis Lecture Outline Following Topics will be discussed Requirements in performance and cost Performance metrics Work

More information

EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications

EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications ECE6102 Dependable Distribute Systems, Fall2010 EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications Deepal Jayasinghe, Hyojun Kim, Mohammad M. Hossain, Ali Payani

More information

Distributed Operating Systems. Cluster Systems

Distributed Operating Systems. Cluster Systems Distributed Operating Systems Cluster Systems Ewa Niewiadomska-Szynkiewicz ens@ia.pw.edu.pl Institute of Control and Computation Engineering Warsaw University of Technology E&IT Department, WUT 1 1. Cluster

More information

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Dr. Maurice Eggen Nathan Franklin Department of Computer Science Trinity University San Antonio, Texas 78212 Dr. Roger Eggen Department

More information

Proactive, Resource-Aware, Tunable Real-time Fault-tolerant Middleware

Proactive, Resource-Aware, Tunable Real-time Fault-tolerant Middleware Proactive, Resource-Aware, Tunable Real-time Fault-tolerant Middleware Priya Narasimhan T. Dumitraş, A. Paulos, S. Pertet, C. Reverte, J. Slember, D. Srivastava Carnegie Mellon University Problem Description

More information

A Clustered Approach for Load Balancing in Distributed Systems

A Clustered Approach for Load Balancing in Distributed Systems SSRG International Journal of Mobile Computing & Application (SSRG-IJMCA) volume 2 Issue 1 Jan to Feb 2015 A Clustered Approach for Load Balancing in Distributed Systems Shweta Rajani 1, Niharika Garg

More information

Comparative Study of Load Balancing Algorithms

Comparative Study of Load Balancing Algorithms IOSR Journal of Engineering (IOSRJEN) e-issn: 2250-3021, p-issn: 2278-8719 Vol. 3, Issue 3 (Mar. 2013), V2 PP 45-50 Comparative Study of Load Balancing Algorithms Jyoti Vashistha 1, Anant Kumar Jayswal

More information

MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS

MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS Priyesh Kanungo 1 Professor and Senior Systems Engineer (Computer Centre), School of Computer Science and

More information

DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM

DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM 1 Introduction In parallel distributed computing system, due to the lightly loaded and overloaded nodes that cause load imbalance, could affect

More information

A Study on the Application of Existing Load Balancing Algorithms for Large, Dynamic, Heterogeneous Distributed Systems

A Study on the Application of Existing Load Balancing Algorithms for Large, Dynamic, Heterogeneous Distributed Systems A Study on the Application of Existing Load Balancing Algorithms for Large, Dynamic, Heterogeneous Distributed Systems RUPAM MUKHOPADHYAY, DIBYAJYOTI GHOSH AND NANDINI MUKHERJEE Department of Computer

More information

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique Jyoti Malhotra 1,Priya Ghyare 2 Associate Professor, Dept. of Information Technology, MIT College of

More information

Efficient Data Replication Scheme based on Hadoop Distributed File System

Efficient Data Replication Scheme based on Hadoop Distributed File System , pp. 177-186 http://dx.doi.org/10.14257/ijseia.2015.9.12.16 Efficient Data Replication Scheme based on Hadoop Distributed File System Jungha Lee 1, Jaehwa Chung 2 and Daewon Lee 3* 1 Division of Supercomputing,

More information

Index Terms : Load rebalance, distributed file systems, clouds, movement cost, load imbalance, chunk.

Index Terms : Load rebalance, distributed file systems, clouds, movement cost, load imbalance, chunk. Load Rebalancing for Distributed File Systems in Clouds. Smita Salunkhe, S. S. Sannakki Department of Computer Science and Engineering KLS Gogte Institute of Technology, Belgaum, Karnataka, India Affiliated

More information

EFFICIENT SCHEDULING STRATEGY USING COMMUNICATION AWARE SCHEDULING FOR PARALLEL JOBS IN CLUSTERS

EFFICIENT SCHEDULING STRATEGY USING COMMUNICATION AWARE SCHEDULING FOR PARALLEL JOBS IN CLUSTERS EFFICIENT SCHEDULING STRATEGY USING COMMUNICATION AWARE SCHEDULING FOR PARALLEL JOBS IN CLUSTERS A.Neela madheswari 1 and R.S.D.Wahida Banu 2 1 Department of Information Technology, KMEA Engineering College,

More information

Fault-Tolerant Framework for Load Balancing System

Fault-Tolerant Framework for Load Balancing System Fault-Tolerant Framework for Load Balancing System Y. K. LIU, L.M. CHENG, L.L.CHENG Department of Electronic Engineering City University of Hong Kong Tat Chee Avenue, Kowloon, Hong Kong SAR HONG KONG Abstract:

More information

A Survey on Load Balancing and Scheduling in Cloud Computing

A Survey on Load Balancing and Scheduling in Cloud Computing IJIRST International Journal for Innovative Research in Science & Technology Volume 1 Issue 7 December 2014 ISSN (online): 2349-6010 A Survey on Load Balancing and Scheduling in Cloud Computing Niraj Patel

More information

Portable Parallel Programming for the Dynamic Load Balancing of Unstructured Grid Applications

Portable Parallel Programming for the Dynamic Load Balancing of Unstructured Grid Applications Portable Parallel Programming for the Dynamic Load Balancing of Unstructured Grid Applications Rupak Biswas MRJ Technology Solutions NASA Ames Research Center Moffett Field, CA 9435, USA rbiswas@nas.nasa.gov

More information

An Optimistic Parallel Simulation Protocol for Cloud Computing Environments

An Optimistic Parallel Simulation Protocol for Cloud Computing Environments An Optimistic Parallel Simulation Protocol for Cloud Computing Environments 3 Asad Waqar Malik 1, Alfred J. Park 2, Richard M. Fujimoto 3 1 National University of Science and Technology, Pakistan 2 IBM

More information

CSAR: Cluster Storage with Adaptive Redundancy

CSAR: Cluster Storage with Adaptive Redundancy CSAR: Cluster Storage with Adaptive Redundancy Manoj Pillai, Mario Lauria Department of Computer and Information Science The Ohio State University Columbus, OH, 4321 Email: pillai,lauria@cis.ohio-state.edu

More information

Chapter 1 - Web Server Management and Cluster Topology

Chapter 1 - Web Server Management and Cluster Topology Objectives At the end of this chapter, participants will be able to understand: Web server management options provided by Network Deployment Clustered Application Servers Cluster creation and management

More information

Scheduling in Distributed Systems

Scheduling in Distributed Systems Scheduling in Distributed Systems Dongning Liang, Pei-Jung Ho, Bao Liu Department of Computer Science and Engineering University of California, San Diego Abstract This paper presents several scheduling/coscheduling

More information

Mining Association Rules on Grid Platforms

Mining Association Rules on Grid Platforms UNIVERSITY OF TUNIS EL MANAR FACULTY OF SCIENCES OF TUNISIA Mining Association Rules on Grid Platforms Raja Tlili raja_tlili@yahoo.fr Yahya Slimani yahya.slimani@fst.rnu.tn CoreGrid 11 Plan Introduction

More information

Introduction to Cloud Computing

Introduction to Cloud Computing Introduction to Cloud Computing Parallel Processing I 15 319, spring 2010 7 th Lecture, Feb 2 nd Majd F. Sakr Lecture Motivation Concurrency and why? Different flavors of parallel computing Get the basic

More information

PARALLEL PROCESSING AND THE DATA WAREHOUSE

PARALLEL PROCESSING AND THE DATA WAREHOUSE PARALLEL PROCESSING AND THE DATA WAREHOUSE BY W. H. Inmon One of the essences of the data warehouse environment is the accumulation of and the management of large amounts of data. Indeed, it is said that

More information

- An Essential Building Block for Stable and Reliable Compute Clusters

- An Essential Building Block for Stable and Reliable Compute Clusters Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative

More information

Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems

Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems Fault Tolerance & Reliability CDA 5140 Chapter 3 RAID & Sample Commercial FT Systems - basic concept in these, as with codes, is redundancy to allow system to continue operation even if some components

More information

Study of Various Load Balancing Techniques in Cloud Environment- A Review

Study of Various Load Balancing Techniques in Cloud Environment- A Review International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-04 E-ISSN: 2347-2693 Study of Various Load Balancing Techniques in Cloud Environment- A Review Rajdeep

More information

Improved Dynamic Load Balance Model on Gametheory for the Public Cloud

Improved Dynamic Load Balance Model on Gametheory for the Public Cloud ISSN (Online): 2349-7084 GLOBAL IMPACT FACTOR 0.238 DIIF 0.876 Improved Dynamic Load Balance Model on Gametheory for the Public Cloud 1 Rayapu Swathi, 2 N.Parashuram, 3 Dr S.Prem Kumar 1 (M.Tech), CSE,

More information

The Optimistic Total Order Protocol

The Optimistic Total Order Protocol From Spontaneous Total Order to Uniform Total Order: different degrees of optimistic delivery Luís Rodrigues Universidade de Lisboa ler@di.fc.ul.pt José Mocito Universidade de Lisboa jmocito@lasige.di.fc.ul.pt

More information

Load Balancing. Load Balancing 1 / 24

Load Balancing. Load Balancing 1 / 24 Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait

More information

Maginatics Cloud Storage Platform for Elastic NAS Workloads

Maginatics Cloud Storage Platform for Elastic NAS Workloads Maginatics Cloud Storage Platform for Elastic NAS Workloads Optimized for Cloud Maginatics Cloud Storage Platform () is the first solution optimized for the cloud. It provides lower cost, easier administration,

More information

Parallel Scalable Algorithms- Performance Parameters

Parallel Scalable Algorithms- Performance Parameters www.bsc.es Parallel Scalable Algorithms- Performance Parameters Vassil Alexandrov, ICREA - Barcelona Supercomputing Center, Spain Overview Sources of Overhead in Parallel Programs Performance Metrics for

More information

Load Balancing in cloud computing

Load Balancing in cloud computing Load Balancing in cloud computing 1 Foram F Kherani, 2 Prof.Jignesh Vania Department of computer engineering, Lok Jagruti Kendra Institute of Technology, India 1 kheraniforam@gmail.com, 2 jigumy@gmail.com

More information