A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture
|
|
- Josephine Ross
- 8 years ago
- Views:
Transcription
1 A Robust Dynamic Load-balancing Scheme for Data Parallel Application on Message Passing Architecture Yangsuk Kee Department of Computer Engineering Seoul National University Seoul, , Korea Soonhoi Ha* Department of Computer Engineering Seoul National University Seoul, , Korea Abstract The working condition of a multicomputer system based on message passing communication is changeable and unpredictable. Any good algorithm for the system must be adaptive to the dynamic change of working condition. In this paper, we propose a new algorithm called RAS(Reservation And Work stealing). RAS treats load balancing, fault-tolerance and processor selection problems by work stealing and reservation based distribution. According to our experiment on the IBM SP2 with a matrix multiplication program, RAS is proven superior to the weighted factoring under the shared running environment and comparable under the dedicated running environment. Keywords: dynamic load balancing, fault tolerance, processor selection 1 Introduction As the performance of general purpose processors and networks improves, the multicomputer system emerges as a viable platform for high-performance computing. Interprocessor communication of the system is mainly achieved by explicit message passing. Due to the high communication overhead of message passing, only coarse-grain parallelism is exploited on the system. Hence, we are specially aimed at executing data parallel applications which have coarse-grain parallelism. In a time-shared environment in which a system is not dedicated to the program, a program may show very dynamic behavior during its running period. The operating system and other user processes usually interfere the program. The change of loads severely affects the overall performance since the running time of an application depends on the speed of the slowest node. Therefore, balancing loads among nodes is essential to achieving shorter running time. In addition, there are several other considerations in exploiting parallelism on a multicomputer system. First, a multicomputer system, particularly NOW/COW, generally does not guarantee the reliability of nodes. Few previous dynamic load balancing approaches, if any, do allow the failure of nodes and links. To guarantee the termination of a program, an algorithm should tolerate the dynamic failures. Second, it is not always true that the performance increases in proportion with the number of nodes, because high communication overhead allows the master to manage only a limited number of slaves without performance degradation. Superfluous nodes may only raise the scheduling burden of the master. Furthermore, the master may make a wrong decision by sending data to slower nodes. Therefore, it is necessary to select faster nodes up to as many as a certain threshold. We propose a new method, RAS (Reservation And work Stealing), to exploit coarsegrain data parallelism on a multicomputer system by solving load-balancing, fault-tolerance and node selection problems at the same time. In this paper, we compare RAS with the weighted factoring algorithm with a data parallel application, matrix multiplication. Our
2 experiments are performed on the IBM SP2 using MPI. We first review some prior approaches to dynamic load-balancing problem on several systems. Then, we present the proposed algorithm in details. Last, we report our experimental results, which are used to demonstrate the robustness of our algorithm under various running conditions. 2 Previous Works The simplest way to share N data among P processors is to divide the whole data into N P chunks. However, this method is not appropriate for a multicomputer system, because it does not consider the dynamic change of running environment and communication overhead. Hence, several approaches have been proposed to solve the problems. One approach is to balance workloads by reducing the amount of data as the scheduling progresses. GSS [1] allocates 1 P of the remaining data to the next idle node. The size of chunk for the ith request is G i = (1 1 P )i N P. Factoring [2] schedules data in batches of P chunks sized equally. In practice, the total number of data per batch is half the remaining data. The size of chunks in the ith batch is F i = ( 1 2 )i+1 N P. Originally both algorithms are proposed to a shared memory system. When both algorithms are applied to a multicomputer system, large chunks assigned to the slowest node at the beginning of the computation can delay the overall finishing time, because the remaining data may not be sufficient to achieve a good load balance. Weighed factoring [3] is a variant of factoring which was proposed for multicomputer systems. Weighted factoring is identical to factoring except that the size of chunks in a batch is determined in proportion to the weight(w j ) of nodes. The size of the jth chunk in the ith batch is F ij = ( 1 2 )i+1 N W j k=p. Even k=1 W k though the weighted factoring considered heterogeneity of nodes, it still can not reduce the Algorithm Chunk Size Static Chunk GSS Factoring Weighted Factoring (P=4, N=100, Weights are ) Table 1: Sample chunk sizes effect of superfluous nodes as mentioned in [3]. Table 1 illustrates the sample chunk sizes according to these algorithms. The other approach is work stealing wherein work migrates dynamically from heavily loaded nodes to lightly loaded ones. Hybrid Scheduling [4] combines a static scheduling method with a dynamic one. The master evenly distributes data to the slaves at the static level. Then, the slaves divide the given data into chunks using an algorithm like GSS at the dynamic level. After the execution of each chunk, all slaves periodically send messages to the master to inform of their performance. When the master detects certain imbalance of workloads, it sends direction to the slaves to redistribute the data. Hamdi s approach [5] is similar to hybrid scheduling. The master periodically polls the slaves to collect workload information and partitions data into several regions. To avoid unnecessary communication, a slave sends a request for the data in the region to the slave who has the data, when the requesting slave has no data to compute. Both algorithms experience high scheduling overhead to collect load information, because lock-step synchronization and periodic polling are burden-some operations. 3 RAS Algorithm We propose a new adaptive algorithm, RAS, for multicomputer systems. To reduce or hide communication overhead, RAS adopts overlapped communication. In a message passing environment, sending a large message is desirable compared with multiple small messages, because it amortizes high start-up cost. However, as shown in figure 1, large message delays the start time of the slaves. In order to achieve high performance, there has to be
3 a trade-off between the start-up cost and the start time. Communication overhead is proportional to the size of message, if the size is large enough [6]. From an experiment as shown in figure 2, we can find the point on which linearity is broken; the message larger than 1KB gives us relatively early start time without significant overhead. slaves and redistributes data when load imbalance is detected. As the master is responsible for both scheduling and distribution of data, RAS allows several slaves to compute a data redundantly. This mechanism makes RAS tolerable to the failure of slaves. Figure 3: RAS algorithm flowchart Figure 1: Large message vs. Small message 3.1 Performance Estimation of Slaves Figure 2: Overhead per 1KB of MPI Send on SP2 RAS consists of three phases as shown in figure 3. In the first phase, the master estimates the performance of slaves in terms of their computation latency. After the estimation, the master distributes data on the basis of the performance. In this distribution phase, the master can select faster slaves in the slave pool as many as profitably used. In the last phase, the master monitors the workloads of The performance estimation of slaves should be carried out in advance to schedule data precisely. We divide the whole data into basic chunks which are the unit of computation. In case of matrix multiplication, one row or column can be a basic chunk. If the size of a basic chunk is smaller than 1KB, several basic chunks are packed into a basic message to reduce the communication overhead. The size of a basic message is determined as 1KB S basic S basic, where S basic denotes the size of a basic chunk in bytes. When RAS starts up, the master sends one basic message to each slave for estimation. The first basic message sent to each slave is called a probe. The average execution time of a basic message is used to represent the performance. The master wants the slaves to report their performance every computation of a basic message for accurate estimation. However, frequent reports will be burden-some to both the slaves and the network. As a compromise, a slave reports the result of computation piggybacked
4 with its average execution time, when it meets the following conditions. When a slave finishs computation of the probe. When a slave receives the End Of Data Distribution message. When a slave produces results larger than 1KB. When the master interrupts a slave to steal data. When a slave has no data to compute. The estimation is used both in the distribution phase and in the work stealing phase. The master continues to check the arrival of reply to the probe until the end of the distribution phase, if not all the slaves report their performances. 3.2 Data Distribution on the Basis of Reservation In the distribution phase, the master sends data in the form of basic message. The master converts the average execution time into the sending period of basic message and distributes basic messages on the basis of the period. The sending period(p i ) of the ith slave is deter- Texec(i) T overhead, where T exec (i) denotes the mined as average execution time of a basic message in the ith slave and T overhead does the overhead of sending operation for a basic message. The number of MPI Send operations is used to represent the sending period. When the ith basic message is sent to the jth slave, the (i+p j )th basic message is reserved for the slave. There are two possible ambiguous cases during this reservation. One is the case that several slaves try to reserve the same data. In this case, the master selects the fastest slave and sends the data to the slave. The other slaves who fail to get the data reserve the next data. The other is the case that no slave reserves a data. This case is divided into two sub-cases. When there is no slave who replies Figure 4: Scenario of data distribution based on reservation the result of the probe, the master sends data in a round-robin fashion to all slaves. When there are more than one slaves that responded to the result of the probe, the master sends data in a round-robin fashion to the responded slaves. These two cases make some slaves receive superfluous data. The master does not use these superfluous data for future reservation. This distribution scheme enables the master to select faster slaves. When the number of slaves exceeds a certain threshold, all data are reserved by some slaves. Since the master sends data to faster slaves, slower ones repeatedly fail to get data. A scenario of data distribution based on the reservation is shown in figure 4, where the sending periods of slaves are 6,9,5 and 7, respectively. The first four data M1, M2, M3 and M4 are probes which are used for the performance estimation. Since M5 and M6 are not reserved, the master distributes them in a round-robin fashion to slave1 and slave2. As the master receives a reply from slave1 during sending M7, it sends M8 to slave1 and reserves M14 for slave1. After that, the master receives a reply from slave3, and it sends M9 to slave3 and reserves M14 for slave3. In case of M14, there are two slaves which reserve the data. The master selects the fastest slave, slave3 and sends the data to the slave. The slave1 who fails to get the data, reserves the next data
5 M Work Stealing Even though the master tries to distribute data in proportion to the speed of slaves, dynamic change of running condition makes the finishing time of slaves inconsistent. Superfluous data distributed in a round-robin style can be another source of load imbalance. The master redistributes data by work stealing when it detects any significant load imbalance. When a slave S i has no basic chunks to compute, the master selects the slave S M who has the most amount of the remaining basic chunks. The master steals a portion of the basic chunks from S M andsendthemto S i. The amount of basic chunks to steal is T given by exec(m) T D exec(i)+t exec(m) M, where D M is the amount of the remaining basic chunks of the richest slave. When the amount of stealing basic chunks is less than 1, the master redundantly sends one basic chunk of S M to S i to avoid the interrupt overhead of the richest slave. When a slave is heavily loaded, the slave can not reply its computational result. In that case, the master can send the duplicated data, which was sent to the slave, to other slaves. With this mechanism, RAS can be tolerant to the failure of slaves only if the master is alive. We adopted a software interrupt to implement work stealing mechanism. As MPI dose not define the interrupt function between processes running on different nodes, we implement a interrupt using asynchronous I/O with socket [7]. 4 Experimental Result We compare RAS with the weighted factoring(wf) under two running environment, dedicated and shared running environment. We check the net elapsed time of matrix multiplication excluding the initialization time of MPI. To make a shared running condition, we execute a simple parallel version of matrix multiplication program as a background process. We take the mean value of running times after 10 executions of both programs. When we refer to the number of processors, it means the total number of processors allocated including the master. The result shows that RAS is a little better than the weighted factoring even under the dedicated running condition and superior under the shared running condition. It means that adaptation to the dynamic change of working condition is important, even though it requires additional scheduling overhead. Note the results of the experiment with 16 processors shown in figure 7 and figure 8. RAS balances loads very adaptively under the dynamic environment, while the weighted factoring suffers from load imbalance. The main reason is that the master sends too many data to the slower slaves. It is useful to observe the number of stealing, since RAS balances loads by work stealing. Table 2 and table 3 show the number of stealing under the various configurations. The number of interrupts excluding the number of redundant computations is used to represent the number of stealing. The number of stealing implies the degree of decision inaccuracy. If the master distributes data on the basis of correct information, the number of stealing decreases. The number of stealing is related to the number of processors and the variance of environment. When the master schedules data incorrectly, it tries to balance loads with additional stealing. As expected, the number of stealing increases as the number of processors grows and the environment becomes more dynamic. To test real fault tolerance, we must rewrite the error handling routines of MPI. For simplicity, we modeled a sudden failure or a surge of load of a node with infinite loop. We let a slave fall into infinite loop after program execution. Then the master can not obtain any information about the slave. As the master is responsible for distributing data, it redundantly sends the data sent to the slave to others. Even though the weight factoring was not terminated, RAS was successfully terminated
6 Figure 5: 100x100 matrix multiplication under the dedicated running environment Figure 7: 100x100 matrix multiplication under the shared running environment Figure 6: 200x200 matrix multiplication under the dedicated running environment Figure 8: 200x200 matrix multiplication under the shared running environment without significant performance degradation. In this paper, we applied RAS to parallelizing a matrix multiplication problem. RAS showed good performance under both dedicated and shared running environments, even though the degree of difference is changeable according to the running environment. Furthermore, RAS was fault-tolerant and adaptive to the number of slaves. For more rigorous evaluation, it is necessary to execute various applications. We also have a plan to apply RAS to PC cluster and cluster of workstations. In addition, it is necessary to give user friedly interface of library for application programmer. matrix size number of processors x x Table 2: The number of stealing under the dedicated running environment matrix size number of processors x x Table 3: The number of stealing under the shared running environment
7 References [1] Constantine D. Polychronopoulos and David J. Kuck. Guided Self-scheduling: A Practical Scheduling Scheme for Parallel Supercomputers. IEEE Transaction on Computers, Vol.C-36, No.12: , DEC [2] Susan Flynn Hummel, Edith Schonberg, and Lawrence E.Flynn. Factoring: A Method for Scheduling Parallel Loops. Communication of the ACM, Vol.35, No.8:90 101, Aug [3] Susan Flynn Hummel, Jeanette Schmidt, R.N.Uma, and Joel Wein. Load-Sharing in Heterogeneous Systems via Weighted Factoring. SPAA, pages , [4] Oscar Plata and Francisco F. Rivera. Combining Static and Dynamic Scheduling on Distributed-Memory Multiprocessors. ICS, pages , [5] Mounir Hamdi and Chi-Kin Lee. Dynamic Load Balancing of Data Parallel Applications on a Distributed Network. ICS, pages , [6] Zhiwei Xu and Kai Hwang. Modeling Communication Overhead: MPI and MPL Performance on the IBM SP2. IEEE Parallel and Distributed Technology, Spring:9 23, [7] Message Passing Interface Forum. MPI : A Message-Passing Interface Standard. May, 1994.
A Review of Customized Dynamic Load Balancing for a Network of Workstations
A Review of Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester
More informationDynamic Load Balancing in a Network of Workstations
Dynamic Load Balancing in a Network of Workstations 95.515F Research Report By: Shahzad Malik (219762) November 29, 2000 Table of Contents 1 Introduction 3 2 Load Balancing 4 2.1 Static Load Balancing
More informationHigh Performance Cluster Support for NLB on Window
High Performance Cluster Support for NLB on Window [1]Arvind Rathi, [2] Kirti, [3] Neelam [1]M.Tech Student, Department of CSE, GITM, Gurgaon Haryana (India) arvindrathi88@gmail.com [2]Asst. Professor,
More informationLocality-Preserving Dynamic Load Balancing for Data-Parallel Applications on Distributed-Memory Multiprocessors
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 18, 1037-1048 (2002) Short Paper Locality-Preserving Dynamic Load Balancing for Data-Parallel Applications on Distributed-Memory Multiprocessors PANGFENG
More informationLoad balancing. David Bindel. 12 Nov 2015
Load balancing David Bindel 12 Nov 2015 Inefficiencies in parallel code Poor single processor performance Typically in the memory system Saw this in matrix multiply assignment Overhead for parallelism
More informationPERFORMANCE EVALUATION OF THREE DYNAMIC LOAD BALANCING ALGORITHMS ON SPMD MODEL
PERFORMANCE EVALUATION OF THREE DYNAMIC LOAD BALANCING ALGORITHMS ON SPMD MODEL Najib A. Kofahi Associate Professor Department of Computer Sciences Faculty of Information Technology and Computer Sciences
More informationHectiling: An Integration of Fine and Coarse Grained Load Balancing Strategies 1
Copyright 1998 IEEE. Published in the Proceedings of HPDC 7 98, 28 31 July 1998 at Chicago, Illinois. Personal use of this material is permitted. However, permission to reprint/republish this material
More informationCellular Computing on a Linux Cluster
Cellular Computing on a Linux Cluster Alexei Agueev, Bernd Däne, Wolfgang Fengler TU Ilmenau, Department of Computer Architecture Topics 1. Cellular Computing 2. The Experiment 3. Experimental Results
More informationRevoScaleR Speed and Scalability
EXECUTIVE WHITE PAPER RevoScaleR Speed and Scalability By Lee Edlefsen Ph.D., Chief Scientist, Revolution Analytics Abstract RevoScaleR, the Big Data predictive analytics library included with Revolution
More informationA Dynamic Approach for Load Balancing using Clusters
A Dynamic Approach for Load Balancing using Clusters ShwetaRajani 1, RenuBagoria 2 Computer Science 1,2,Global Technical Campus, Jaipur 1,JaganNath University, Jaipur 2 Email: shwetarajani28@yahoo.in 1
More informationMultiprocessor Scheduling and Scheduling in Linux Kernel 2.6
Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6 Winter Term 2008 / 2009 Jun.-Prof. Dr. André Brinkmann Andre.Brinkmann@uni-paderborn.de Universität Paderborn PC² Agenda Multiprocessor and
More informationAn Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems
An Empirical Study and Analysis of the Dynamic Load Balancing Techniques Used in Parallel Computing Systems Ardhendu Mandal and Subhas Chandra Pal Department of Computer Science and Application, University
More informationCustomized Dynamic Load Balancing for a Network of Workstations
Customized Dynamic Load Balancing for a Network of Workstations Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester, Rochester NY 4627 zaki,wei,srini
More informationLoad balancing. Prof. Richard Vuduc Georgia Institute of Technology CSE/CS 8803 PNA: Parallel Numerical Algorithms [L.26] Thursday, April 17, 2008
Load balancing Prof. Richard Vuduc Georgia Institute of Technology CSE/CS 8803 PNA: Parallel Numerical Algorithms [L.26] Thursday, April 17, 2008 1 Today s sources CS 194/267 at UCB (Yelick/Demmel) Intro
More informationA Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed
More informationClient/Server Computing Distributed Processing, Client/Server, and Clusters
Client/Server Computing Distributed Processing, Client/Server, and Clusters Chapter 13 Client machines are generally single-user PCs or workstations that provide a highly userfriendly interface to the
More informationOverlapping Data Transfer With Application Execution on Clusters
Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm reid@cs.toronto.edu stumm@eecg.toronto.edu Department of Computer Science Department of Electrical and Computer
More informationParallel Ray Tracing using MPI: A Dynamic Load-balancing Approach
Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach S. M. Ashraful Kadir 1 and Tazrian Khan 2 1 Scientific Computing, Royal Institute of Technology (KTH), Stockholm, Sweden smakadir@csc.kth.se,
More informationThe International Journal Of Science & Technoledge (ISSN 2321 919X) www.theijst.com
THE INTERNATIONAL JOURNAL OF SCIENCE & TECHNOLEDGE Efficient Parallel Processing on Public Cloud Servers using Load Balancing Manjunath K. C. M.Tech IV Sem, Department of CSE, SEA College of Engineering
More informationMOSIX: High performance Linux farm
MOSIX: High performance Linux farm Paolo Mastroserio [mastroserio@na.infn.it] Francesco Maria Taurino [taurino@na.infn.it] Gennaro Tortone [tortone@na.infn.it] Napoli Index overview on Linux farm farm
More informationOperatin g Systems: Internals and Design Principle s. Chapter 10 Multiprocessor and Real-Time Scheduling Seventh Edition By William Stallings
Operatin g Systems: Internals and Design Principle s Chapter 10 Multiprocessor and Real-Time Scheduling Seventh Edition By William Stallings Operating Systems: Internals and Design Principles Bear in mind,
More informationLoad Balancing In Concurrent Parallel Applications
Load Balancing In Concurrent Parallel Applications Jeff Figler Rochester Institute of Technology Computer Engineering Department Rochester, New York 14623 May 1999 Abstract A parallel concurrent application
More informationA Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters
A Comparative Performance Analysis of Load Balancing Algorithms in Distributed System using Qualitative Parameters Abhijit A. Rajguru, S.S. Apte Abstract - A distributed system can be viewed as a collection
More informationA Locally Cache-Coherent Multiprocessor Architecture
A Locally Cache-Coherent Multiprocessor Architecture Kevin Rich Computing Research Group Lawrence Livermore National Laboratory Livermore, CA 94551 Norman Matloff Division of Computer Science University
More informationCustomized Dynamic Load Balancing for a Network of Workstations 1
JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING 43, 156 162 (1997) ARTICLE NO. PC971339 Customized Dynamic Load Balancing for a Network of Workstations 1 Mohammed Javeed Zaki, Wei Li, and Srinivasan Parthasarathy
More informationFault Tolerance in Hadoop for Work Migration
1 Fault Tolerance in Hadoop for Work Migration Shivaraman Janakiraman Indiana University Bloomington ABSTRACT Hadoop is a framework that runs applications on large clusters which are built on numerous
More information{emery,browne}@cs.utexas.edu ABSTRACT. Keywords scalable, load distribution, load balancing, work stealing
Scalable Load Distribution and Load Balancing for Dynamic Parallel Programs E. Berger and J. C. Browne Department of Computer Science University of Texas at Austin Austin, Texas 78701 USA 01-512-471-{9734,9579}
More informationTCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance
TCP Servers: Offloading TCP Processing in Internet Servers. Design, Implementation, and Performance M. Rangarajan, A. Bohra, K. Banerjee, E.V. Carrera, R. Bianchini, L. Iftode, W. Zwaenepoel. Presented
More informationLoad Balancing Algorithms for Peer to Peer and Client Server Distributed Environments
Load Balancing Algorithms for Peer to Peer and Client Server Distributed Environments Sameena Naaz Afshar Alam Ranjit Biswas Department of Computer Science Jamia Hamdard, New Delhi, India ABSTRACT Advancements
More informationKeywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age.
Volume 3, Issue 10, October 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Load Measurement
More informationOutline. Distributed DBMS
Outline Introduction Background Architecture Distributed Database Design Semantic Data Control Distributed Query Processing Distributed Transaction Management Data server approach Parallel architectures
More informationLoad Balancing in Distributed Data Base and Distributed Computing System
Load Balancing in Distributed Data Base and Distributed Computing System Lovely Arya Research Scholar Dravidian University KUPPAM, ANDHRA PRADESH Abstract With a distributed system, data can be located
More informationComparison on Different Load Balancing Algorithms of Peer to Peer Networks
Comparison on Different Load Balancing Algorithms of Peer to Peer Networks K.N.Sirisha *, S.Bhagya Rekha M.Tech,Software Engineering Noble college of Engineering & Technology for Women Web Technologies
More informationDYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS
Journal homepage: www.mjret.in DYNAMIC LOAD BALANCING SCHEME FOR ITERATIVE APPLICATIONS ISSN:2348-6953 Rahul S. Wankhade, Darshan M. Marathe, Girish P. Nikam, Milind R. Jawale Department of Computer Engineering,
More informationScalability and Classifications
Scalability and Classifications 1 Types of Parallel Computers MIMD and SIMD classifications shared and distributed memory multicomputers distributed shared memory computers 2 Network Topologies static
More informationResource Allocation Schemes for Gang Scheduling
Resource Allocation Schemes for Gang Scheduling B. B. Zhou School of Computing and Mathematics Deakin University Geelong, VIC 327, Australia D. Walsh R. P. Brent Department of Computer Science Australian
More informationScheduling Allowance Adaptability in Load Balancing technique for Distributed Systems
Scheduling Allowance Adaptability in Load Balancing technique for Distributed Systems G.Rajina #1, P.Nagaraju #2 #1 M.Tech, Computer Science Engineering, TallaPadmavathi Engineering College, Warangal,
More informationA Review on an Algorithm for Dynamic Load Balancing in Distributed Network with Multiple Supporting Nodes with Interrupt Service
A Review on an Algorithm for Dynamic Load Balancing in Distributed Network with Multiple Supporting Nodes with Interrupt Service Payal Malekar 1, Prof. Jagruti S. Wankhede 2 Student, Information Technology,
More informationLoad Balancing of Web Server System Using Service Queue Length
Load Balancing of Web Server System Using Service Queue Length Brajendra Kumar 1, Dr. Vineet Richhariya 2 1 M.tech Scholar (CSE) LNCT, Bhopal 2 HOD (CSE), LNCT, Bhopal Abstract- In this paper, we describe
More informationChapter 18: Database System Architectures. Centralized Systems
Chapter 18: Database System Architectures! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types 18.1 Centralized Systems! Run on a single computer system and
More informationImproved Hybrid Dynamic Load Balancing Algorithm for Distributed Environment
International Journal of Scientific and Research Publications, Volume 3, Issue 3, March 2013 1 Improved Hybrid Dynamic Load Balancing Algorithm for Distributed Environment UrjashreePatil*, RajashreeShedge**
More informationGrid Computing Approach for Dynamic Load Balancing
International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-1 E-ISSN: 2347-2693 Grid Computing Approach for Dynamic Load Balancing Kapil B. Morey 1*, Sachin B. Jadhav
More informationLoad Balancing in Fault Tolerant Video Server
Load Balancing in Fault Tolerant Video Server # D. N. Sujatha*, Girish K*, Rashmi B*, Venugopal K. R*, L. M. Patnaik** *Department of Computer Science and Engineering University Visvesvaraya College of
More informationThread level parallelism
Thread level parallelism ILP is used in straight line code or loops Cache miss (off-chip cache and main memory) is unlikely to be hidden using ILP. Thread level parallelism is used instead. Thread: process
More informationBENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
More informationContributions to Gang Scheduling
CHAPTER 7 Contributions to Gang Scheduling In this Chapter, we present two techniques to improve Gang Scheduling policies by adopting the ideas of this Thesis. The first one, Performance- Driven Gang Scheduling,
More informationA Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems*
A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* Junho Jang, Saeyoung Han, Sungyong Park, and Jihoon Yang Department of Computer Science and Interdisciplinary Program
More informationChapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup
Chapter 12: Multiprocessor Architectures Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Objective Be familiar with basic multiprocessor architectures and be able to
More informationInternational Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518
International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 Load Balancing Heterogeneous Request in DHT-based P2P Systems Mrs. Yogita A. Dalvi Dr. R. Shankar Mr. Atesh
More informationHow To Understand The Concept Of A Distributed System
Distributed Operating Systems Introduction Ewa Niewiadomska-Szynkiewicz and Adam Kozakiewicz ens@ia.pw.edu.pl, akozakie@ia.pw.edu.pl Institute of Control and Computation Engineering Warsaw University of
More informationVarious Schemes of Load Balancing in Distributed Systems- A Review
741 Various Schemes of Load Balancing in Distributed Systems- A Review Monika Kushwaha Pranveer Singh Institute of Technology Kanpur, U.P. (208020) U.P.T.U., Lucknow Saurabh Gupta Pranveer Singh Institute
More informationSource Code Transformations Strategies to Load-balance Grid Applications
Source Code Transformations Strategies to Load-balance Grid Applications Romaric David, Stéphane Genaud, Arnaud Giersch, Benjamin Schwarz, and Éric Violard LSIIT-ICPS, Université Louis Pasteur, Bd S. Brant,
More informationA Flexible Cluster Infrastructure for Systems Research and Software Development
Award Number: CNS-551555 Title: CRI: Acquisition of an InfiniBand Cluster with SMP Nodes Institution: Florida State University PIs: Xin Yuan, Robert van Engelen, Kartik Gopalan A Flexible Cluster Infrastructure
More informationRESEARCH PAPER International Journal of Recent Trends in Engineering, Vol 1, No. 1, May 2009
An Algorithm for Dynamic Load Balancing in Distributed Systems with Multiple Supporting Nodes by Exploiting the Interrupt Service Parveen Jain 1, Daya Gupta 2 1,2 Delhi College of Engineering, New Delhi,
More informationThe Probabilistic Model of Cloud Computing
A probabilistic multi-tenant model for virtual machine mapping in cloud systems Zhuoyao Wang, Majeed M. Hayat, Nasir Ghani, and Khaled B. Shaban Department of Electrical and Computer Engineering, University
More informationA novel load balancing algorithm for computational grid
International Journal of Computational Intelligence Techniques, ISSN: 0976 0466 & E-ISSN: 0976 0474 Volume 1, Issue 1, 2010, PP-20-26 A novel load balancing algorithm for computational grid Saravanakumar
More informationCentralized Systems. A Centralized Computer System. Chapter 18: Database System Architectures
Chapter 18: Database System Architectures Centralized Systems! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types! Run on a single computer system and do
More informationSymmetric Multiprocessing
Multicore Computing A multi-core processor is a processing system composed of two or more independent cores. One can describe it as an integrated circuit to which two or more individual processors (called
More informationEfficient Parallel Processing on Public Cloud Servers Using Load Balancing
Efficient Parallel Processing on Public Cloud Servers Using Load Balancing Valluripalli Srinath 1, Sudheer Shetty 2 1 M.Tech IV Sem CSE, Sahyadri College of Engineering & Management, Mangalore. 2 Asso.
More informationFAULT TOLERANCE FOR MULTIPROCESSOR SYSTEMS VIA TIME REDUNDANT TASK SCHEDULING
FAULT TOLERANCE FOR MULTIPROCESSOR SYSTEMS VIA TIME REDUNDANT TASK SCHEDULING Hussain Al-Asaad and Alireza Sarvi Department of Electrical & Computer Engineering University of California Davis, CA, U.S.A.
More informationA Comparison of General Approaches to Multiprocessor Scheduling
A Comparison of General Approaches to Multiprocessor Scheduling Jing-Chiou Liou AT&T Laboratories Middletown, NJ 0778, USA jing@jolt.mt.att.com Michael A. Palis Department of Computer Science Rutgers University
More informationOverview and History of Operating Systems
Overview and History of Operating Systems These are the notes for lecture 1. Please review the Syllabus notes before these. Overview / Historical Developments An Operating System... Sits between hardware
More informationKeywords Load balancing, Dispatcher, Distributed Cluster Server, Static Load balancing, Dynamic Load balancing.
Volume 5, Issue 7, July 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Hybrid Algorithm
More informationA Survey on Load Balancing Algorithms in Cloud Environment
A Survey on Load s in Cloud Environment M.Aruna Assistant Professor (Sr.G)/CSE Erode Sengunthar Engineering College, Thudupathi, Erode, India D.Bhanu, Ph.D Associate Professor Sri Krishna College of Engineering
More informationOptimizing the Virtual Data Center
Optimizing the Virtual Center The ideal virtual data center dynamically balances workloads across a computing cluster and redistributes hardware resources among clusters in response to changing needs.
More informationKappa: A system for Linux P2P Load Balancing and Transparent Process Migration
Kappa: A system for Linux P2P Load Balancing and Transparent Process Migration Gaurav Mogre gaurav.mogre@gmail.com Avinash Hanumanthappa avinash947@gmail.com Alwyn Roshan Pais alwyn@nitk.ac.in Abstract
More informationPerformance Metrics and Scalability Analysis. Performance Metrics and Scalability Analysis
Performance Metrics and Scalability Analysis 1 Performance Metrics and Scalability Analysis Lecture Outline Following Topics will be discussed Requirements in performance and cost Performance metrics Work
More informationEWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications
ECE6102 Dependable Distribute Systems, Fall2010 EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications Deepal Jayasinghe, Hyojun Kim, Mohammad M. Hossain, Ali Payani
More informationDistributed Operating Systems. Cluster Systems
Distributed Operating Systems Cluster Systems Ewa Niewiadomska-Szynkiewicz ens@ia.pw.edu.pl Institute of Control and Computation Engineering Warsaw University of Technology E&IT Department, WUT 1 1. Cluster
More informationLoad Balancing on a Non-dedicated Heterogeneous Network of Workstations
Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Dr. Maurice Eggen Nathan Franklin Department of Computer Science Trinity University San Antonio, Texas 78212 Dr. Roger Eggen Department
More informationProactive, Resource-Aware, Tunable Real-time Fault-tolerant Middleware
Proactive, Resource-Aware, Tunable Real-time Fault-tolerant Middleware Priya Narasimhan T. Dumitraş, A. Paulos, S. Pertet, C. Reverte, J. Slember, D. Srivastava Carnegie Mellon University Problem Description
More informationA Clustered Approach for Load Balancing in Distributed Systems
SSRG International Journal of Mobile Computing & Application (SSRG-IJMCA) volume 2 Issue 1 Jan to Feb 2015 A Clustered Approach for Load Balancing in Distributed Systems Shweta Rajani 1, Niharika Garg
More informationComparative Study of Load Balancing Algorithms
IOSR Journal of Engineering (IOSRJEN) e-issn: 2250-3021, p-issn: 2278-8719 Vol. 3, Issue 3 (Mar. 2013), V2 PP 45-50 Comparative Study of Load Balancing Algorithms Jyoti Vashistha 1, Anant Kumar Jayswal
More informationMEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS
MEASURING PERFORMANCE OF DYNAMIC LOAD BALANCING ALGORITHMS IN DISTRIBUTED COMPUTING APPLICATIONS Priyesh Kanungo 1 Professor and Senior Systems Engineer (Computer Centre), School of Computer Science and
More informationDYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM
DYNAMIC LOAD BALANCING IN A DECENTRALISED DISTRIBUTED SYSTEM 1 Introduction In parallel distributed computing system, due to the lightly loaded and overloaded nodes that cause load imbalance, could affect
More informationA Study on the Application of Existing Load Balancing Algorithms for Large, Dynamic, Heterogeneous Distributed Systems
A Study on the Application of Existing Load Balancing Algorithms for Large, Dynamic, Heterogeneous Distributed Systems RUPAM MUKHOPADHYAY, DIBYAJYOTI GHOSH AND NANDINI MUKHERJEE Department of Computer
More informationA Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique
A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique Jyoti Malhotra 1,Priya Ghyare 2 Associate Professor, Dept. of Information Technology, MIT College of
More informationEfficient Data Replication Scheme based on Hadoop Distributed File System
, pp. 177-186 http://dx.doi.org/10.14257/ijseia.2015.9.12.16 Efficient Data Replication Scheme based on Hadoop Distributed File System Jungha Lee 1, Jaehwa Chung 2 and Daewon Lee 3* 1 Division of Supercomputing,
More informationIndex Terms : Load rebalance, distributed file systems, clouds, movement cost, load imbalance, chunk.
Load Rebalancing for Distributed File Systems in Clouds. Smita Salunkhe, S. S. Sannakki Department of Computer Science and Engineering KLS Gogte Institute of Technology, Belgaum, Karnataka, India Affiliated
More informationEFFICIENT SCHEDULING STRATEGY USING COMMUNICATION AWARE SCHEDULING FOR PARALLEL JOBS IN CLUSTERS
EFFICIENT SCHEDULING STRATEGY USING COMMUNICATION AWARE SCHEDULING FOR PARALLEL JOBS IN CLUSTERS A.Neela madheswari 1 and R.S.D.Wahida Banu 2 1 Department of Information Technology, KMEA Engineering College,
More informationFault-Tolerant Framework for Load Balancing System
Fault-Tolerant Framework for Load Balancing System Y. K. LIU, L.M. CHENG, L.L.CHENG Department of Electronic Engineering City University of Hong Kong Tat Chee Avenue, Kowloon, Hong Kong SAR HONG KONG Abstract:
More informationA Survey on Load Balancing and Scheduling in Cloud Computing
IJIRST International Journal for Innovative Research in Science & Technology Volume 1 Issue 7 December 2014 ISSN (online): 2349-6010 A Survey on Load Balancing and Scheduling in Cloud Computing Niraj Patel
More informationPortable Parallel Programming for the Dynamic Load Balancing of Unstructured Grid Applications
Portable Parallel Programming for the Dynamic Load Balancing of Unstructured Grid Applications Rupak Biswas MRJ Technology Solutions NASA Ames Research Center Moffett Field, CA 9435, USA rbiswas@nas.nasa.gov
More informationAn Optimistic Parallel Simulation Protocol for Cloud Computing Environments
An Optimistic Parallel Simulation Protocol for Cloud Computing Environments 3 Asad Waqar Malik 1, Alfred J. Park 2, Richard M. Fujimoto 3 1 National University of Science and Technology, Pakistan 2 IBM
More informationCSAR: Cluster Storage with Adaptive Redundancy
CSAR: Cluster Storage with Adaptive Redundancy Manoj Pillai, Mario Lauria Department of Computer and Information Science The Ohio State University Columbus, OH, 4321 Email: pillai,lauria@cis.ohio-state.edu
More informationChapter 1 - Web Server Management and Cluster Topology
Objectives At the end of this chapter, participants will be able to understand: Web server management options provided by Network Deployment Clustered Application Servers Cluster creation and management
More informationScheduling in Distributed Systems
Scheduling in Distributed Systems Dongning Liang, Pei-Jung Ho, Bao Liu Department of Computer Science and Engineering University of California, San Diego Abstract This paper presents several scheduling/coscheduling
More informationMining Association Rules on Grid Platforms
UNIVERSITY OF TUNIS EL MANAR FACULTY OF SCIENCES OF TUNISIA Mining Association Rules on Grid Platforms Raja Tlili raja_tlili@yahoo.fr Yahya Slimani yahya.slimani@fst.rnu.tn CoreGrid 11 Plan Introduction
More informationIntroduction to Cloud Computing
Introduction to Cloud Computing Parallel Processing I 15 319, spring 2010 7 th Lecture, Feb 2 nd Majd F. Sakr Lecture Motivation Concurrency and why? Different flavors of parallel computing Get the basic
More informationPARALLEL PROCESSING AND THE DATA WAREHOUSE
PARALLEL PROCESSING AND THE DATA WAREHOUSE BY W. H. Inmon One of the essences of the data warehouse environment is the accumulation of and the management of large amounts of data. Indeed, it is said that
More information- An Essential Building Block for Stable and Reliable Compute Clusters
Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative
More informationFault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems
Fault Tolerance & Reliability CDA 5140 Chapter 3 RAID & Sample Commercial FT Systems - basic concept in these, as with codes, is redundancy to allow system to continue operation even if some components
More informationStudy of Various Load Balancing Techniques in Cloud Environment- A Review
International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-04 E-ISSN: 2347-2693 Study of Various Load Balancing Techniques in Cloud Environment- A Review Rajdeep
More informationImproved Dynamic Load Balance Model on Gametheory for the Public Cloud
ISSN (Online): 2349-7084 GLOBAL IMPACT FACTOR 0.238 DIIF 0.876 Improved Dynamic Load Balance Model on Gametheory for the Public Cloud 1 Rayapu Swathi, 2 N.Parashuram, 3 Dr S.Prem Kumar 1 (M.Tech), CSE,
More informationThe Optimistic Total Order Protocol
From Spontaneous Total Order to Uniform Total Order: different degrees of optimistic delivery Luís Rodrigues Universidade de Lisboa ler@di.fc.ul.pt José Mocito Universidade de Lisboa jmocito@lasige.di.fc.ul.pt
More informationLoad Balancing. Load Balancing 1 / 24
Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait
More informationMaginatics Cloud Storage Platform for Elastic NAS Workloads
Maginatics Cloud Storage Platform for Elastic NAS Workloads Optimized for Cloud Maginatics Cloud Storage Platform () is the first solution optimized for the cloud. It provides lower cost, easier administration,
More informationParallel Scalable Algorithms- Performance Parameters
www.bsc.es Parallel Scalable Algorithms- Performance Parameters Vassil Alexandrov, ICREA - Barcelona Supercomputing Center, Spain Overview Sources of Overhead in Parallel Programs Performance Metrics for
More informationLoad Balancing in cloud computing
Load Balancing in cloud computing 1 Foram F Kherani, 2 Prof.Jignesh Vania Department of computer engineering, Lok Jagruti Kendra Institute of Technology, India 1 kheraniforam@gmail.com, 2 jigumy@gmail.com
More information