Comparing Load Balancing for Server Selection Using Cloud Services
|
|
|
- Madeline Morgan
- 10 years ago
- Views:
Transcription
1 Comparing Load Balancing for Server Selection Using Cloud Services By Daksh Gupta A Project Report Submitted In Partial Fulfillment of the Requirements of the Degree of Masters of Science In Computer Science Supervised By Professor Xumin Liu Department Of Computer Science B. Thomas Golisano College of Computing and Information Sciences Rochester Institute Of Technology Rochester, New York November 11,
2 The project Comparing Load Balancing for Server Selection Using Cloud Services by Daksh Gupta has been examined and approved by the following Examining Committee. Dr. Xumin Liu Professor Project Committee Chair Dr. Rajendra Raj Professor Project Reader Dr. Stanislaw Radziszowski Professor Project Observer 2
3 Dedication To my family for their continuous love and support; and to all my professors for their continuous guidance. 3
4 Acknowledgements I am grateful to all the Professors who have guided me throughout my pursue for Master s Degree in Computer Science here at RIT, but a special mention for Professor Xumin Liu for providing her guidance, and help throughout the progress of my project work. 4
5 ABSTRACT Comparing Load Balancing For Server Selection Using Cloud Services Daksh Gupta Supervised By: Professor Xumin Liu Load Balancing is a technique in which the requests send to the web servers are distributed in such a way that maximum resource utilization is achieved with minimum response time and servers/instances are less overloaded too. The rate at which applications at the enterprise level are expanding, it challenges the infrastructure how to balance the load among the web servers present within the enterprise. These enterprise applications faces a roadblock of determining the least loaded and good performing servers from the pool of available servers for the request sent from the client server. So in order to remain competitive in the present scenario, enterprises are required to have effective load balancing within their environment architecture. On top of it, the main criterion of how to choose architecture for the environment depends on the cost, reliability, and management in a short time frame. To survive in the competitive market, the wisest solution is to use cloud based service for developing infrastructure. This project will help in evaluating which load balancing algorithm should be used, using cloud based service in the industry with a thought on overall performance of infrastructure/system. This report focus on improving and comparing the proposed load- balancing algorithm against load balancing algorithm already implemented, and determining proposed algorithm is better suited. The algorithms will be implemented and compared based on the load, response time, utilization etc. Performance will be evaluated against load, response time, utilization, bps, tps etc. Virtual instances of web servers will be created on cloud services, which will act as web servers, and web services will be developed which will access them. Load balancing algorithms will be developed, which will also be deployed on cloud, and would route the request. The advantages of proposed load balancing algorithm will be highlighted. 5
6 CONTENTS Chapter Introduction Overview Background and Definitions Goals and Motivation Related Work Hypothesis Problem and Proposed Solution Problem Statement Proposed Solution Road Map Chapter Algorithms Dynamic Load Balancer with Least Connections and Fastest Response Time Algorithm Description Dynamic Load Balancer with Least Connections and Fastest Response Time with weight table Algorithm Description Chapter Design Architecture Diagram Integration Diagram of Load Balancing Algorithms Service Level Diag. of Load balancing Algorithm Evaluation Analysis Chapter Implementation and Requirements
7 4.1 Implementation Implementation Strategy Service Detailed Implementation Objective SOAP Implementation Requirements Hardware Software Metrics Round Trip Time Memory Utilization Cost Scalability TPS BPS Simple Strategy Variance Strategy Thread Strategy Chapter Analysis Basic Set Up Performance Test Simple Strategy Thread Strategy Variance Strategy (variance 0.5) Testing Using HP Load Runner Tool Chapter Conclusion Current Status Future Work Lessons Learned BIBLIOGRAPHY
8 Index for Tables Table 1: Simple Strategy results comparison Table 2: Thread Strategy results comparison Table 3: Variance Strategy results comparison Table 4: HP Load Runner Results Comparison. 41 Index for Figures Figure 1: Sample SOAP Request Figure 2: Low Level Architecture Figure 3: High Level Architecture Figure 4: Integration Diagram of Load Balancing Figure 5: Service Diagram of Load Balancing Figure 6: Soap Request Figure 7: Dynamic Load Balancer with Least Connections and Fastest Response Time WSDL Figure 8: Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table WSDL Figure 9: Average Time (Strategy Mode) Figure 10: Maximum Time (Strategy Mode) Figure 11: TPS (Strategy Mode) Figure 12: Average Time (Thread Mode) Figure 13: Maximum Time (Thread Mode) Figure 14: TPS (Thread Mode) Figure 15: Average Time (Variance Mode) Figure 16: Maximum Times (Variance Mode) Figure 17: TPS (Variance Mode) Figure 18: 99%tile Response Time (HP LoadRunner) Figure 19: TPS (HP LoadRunner)
9 Chapter 1 1. Introduction 1.1 Overview Number of users using Internet is growing at an alarming rate so the need of balancing the traffic on the enterprise applications in order to provide high performance and high availability [8]. There are numbers of load balancing algorithms present, but each and every algorithm face some problem or other. The main goal of load balancing is to achieve route the requests among the web servers with a minimum response time. And another issue that enterprise faces is the high cost, managing of web servers built where load balancer can be deployed. Cloud based infrastructure/service can be used as one of solutions where web servers can be built as it has various advantages like cost effective, easily managed, virtual instances, and is scalable too. There are various load balancing techniques available in cloud- based services like Predictive load balancing, random biased, join idle queue etc. [12]. This project will help in evaluating which load balancing algorithm should be used on top of using cloud based service in the industry with a thought on overall performance of infrastructure/system. This report focus on improving and comparing the proposed load balancing algorithm against load balancing algorithm [1] already implemented, and determining proposed algorithm is more well suited. This report describes comparing load balancing techniques on the cloud services, which will act as web servers by using SOAP requests on web services as data set. Web services [13] will be implemented which will each point to one of the load balancing technique, and load will be increased on it. The main objective of load balancing algorithm will be based on distributing or route the data depending upon the closest or best performing web servers. The algorithms will be implemented and compared based on the load, response time, utilization etc. Performance will be evaluated against load, response time, utilization etc. [17, 18]. Virtual instances of web servers will be created on cloud services, which will act as web servers, and web services will be developed which will access them. Load balancing algorithms will be developed, which will also be deployed on cloud, and would route the request. The advantages of proposed load balancing algorithm will be highlighted. 1.2 Background and Definitions Number of users using Internet is growing at an alarming rate so the need of balancing the traffic on the enterprise applications is critical to provide higher tps, lower response time, and less number of outliers [5]. This is being achieved through 9
10 load balancing, which means to route the requests among web servers. There are number of load balancing algorithms present, but each of them has some issues or disadvantages related to it. Standard definitions: Load Balancing: It means balancing the increased client requests on the distributed instance deployed on the server so as to avoid increased delay in response time due to the increased load. It is mainly important for the high level applications where number of request sent by the user can range from thousands to millions i.e. it s very difficult to predict the number of requests. They need to have many web servers for load balancing. They can be categorized under either static algorithms or dynamic algorithms. Round robin, random etc. are categorized under the static algorithms, whereas least connection, observed, etc. are categorized under dynamic algorithms. [22] Random: The random scheduling, requests received from the client are being assigned to any random server from a given list. As the requests are being assigned randomly, there is no mechanism of sharing the load, which leads to overloaded servers. This algorithm leads to underutilization and overutilization of servers. [22] Round Robin: The round robin algorithm routes the new request to the instance or server, followed in queue. It distributes the requests evenly among all the instances so that load is shared evenly. This algorithm is better than the random as requests are diving evenly among the servers. But the main disadvantage of this algorithm is it works better only of all the servers have same configurations. [22] Weighted Round Robin: This algorithm is an enhancement of Round robin where each instance gets assigned with a load depending on processing capability, which is, determined on how that instance is behaving. It removes the deficiency of round robin, but it does not consider the processing time each server is taking in responding the response. [22] To remove, these deficiencies, dynamic load balancing algorithms were used. [3] Dynamic Load Balancing: Dynamic load balancing monitors the state of the server continuously and assigns request to the server, which has less load. There are various dynamic load- balancing algorithm like least connection, fastest, adaptive load balancing etc. Fastest Load Balancing: This method routes the request from the client to that server which has fastest response time [6]. The disadvantage or shortcoming of this algorithm is it s not 10
11 possible for every server to respond the response in seconds, which will lead to congestion in the network. Least Connection Load Balancing: This method derives the meaning from its name only. The server passes a new request from the client to that web server which has very least connection at that point of time [3, 22]. This technique works best where all the applications running web server have same infrastructure. The advantage of this algorithm leads to disadvantage of it. If there are two applications having different infrastructure i.e. HTML application and other uses J2EE or xml, it will lead to bottleneck of connections, as all connections will require have different round trip time as its dependent on the server from where the request is originated. [22] Observed Load Balancing: This technique makes use of fastest, least connection. Web servers are categorized based on least and response time [3, 22]. The server with less requests and higher throughput time gets the new request. But the disadvantage faced by it is it does not have any weights associated the server, due to which the server starts getting overloaded with requests. Web Service: Web service is a medium that permits communication between the applications independent of the platform as well as the language used to program it. It consists of functions, which are accessed by other applications using XML. This interface mechanism means SOA (Service Orientation Architecture). [23] SOAP: The Simple Object Access Protocol (SOAP) provides a transport layer between the applications interacting with the web service using soap wsdl in distributed network. SOAP message consists of XML, which consist of the function that application will be requesting from the web service. The web service end point and port is being provided to application for it to communicate with web service. SOAP xml is generated through the WSDL provided by the web service. WSDL consists of functions that web- service provides, and can be accessed through the network. [17] Figure 1: Sample SOAP Request 11
12 The above figure consists of sample soap request that application uses to interact with web service. JAX- WS: It s an API used in Java, which help in developing the web services, it s a segment of Java EE. It s an open source project. [16] Cloud Computing: Cloud computing is a service provided through the Internet. This service helps the enterprise to obtain software resources like apache, WAS, java, etc. as well as hardware resources i.e. getting firewall set up for the server etc. These days many enterprises or industries are going towards cloud computing, which is not only cheap, but also scalable. One more advantage of using these services is industry are developing an attitude of paying whenever the service is being requested rather than paying a lump amount for everything. Amazon Web Services (AWS) It s a pack of web services provided by Amazon.com so that they can be used for cloud computing. They have many advantages like scalable, reliable, cost effective etc. [24] Amazon Elastic Cloud Compute It is one of the web services, which comes as a part of AWS. It s a web service that helps in increasing the capacity on the server or helps in creating lot of virtual servers, which can be used by clients. [24] Billing is done on the basis of how much the virtual servers are used. 1.3 Goals and Motivation Different enterprises have different requirements, depending upon their needs and budgets. Some enterprises would like to achieve high performance from and availability of their systems without any concerns about the cost incurred, while smaller enterprises with limited budgets would like to get the most out of their systems. Number of users using Internet is growing at an alarming rate so the need of balancing the traffic on the enterprise applications in order to provide high performance and high availability. There are numbers of load balancing algorithms present, but each and every algorithm face some problem or other. The main goal of load balancing is to achieve route the requests among the web servers with a minimum response time. This report focus on improving and comparing the proposed load- balancing algorithm against load balancing algorithm [1] already implemented on cloud servers, and determining proposed algorithm is better suited. Web services will be implemented which will each point to one of the load balancing technique, and load will be increased on it. The load balancer algorithms will be developed as web service so that the client can use it as an end point. The 12
13 web service will act as global traffic manager (GTM), which will be called by the client, and algorithm will act as local traffic manager (LTM) routing the requests to the appropriate cloud servers to get the result back to client in minimum time. The algorithms will be implemented and compared based on the load, response time, utilization etc. Performance will be evaluated against load, response time, utilization etc. Virtual instances of web servers will be created on cloud services, which will act as web servers, and web services will be developed which will access them. Load balancing algorithms will be developed, which will also be deployed on cloud, and would route the request. 1.4 Related Work DONAR [2] (Decentralized Server Selection) was being developed as a distributed system that provides make sure to resolve the name conflicts and also choosing the appropriate server for the client request which has the less load. This algorithm was being developed to direct incoming request from the client to appropriate web servers in order to balance the load between the web servers. The way this algorithm works is mapping node gets the request from the client, which then sends the request to server. Mapping node has an optimizer, and is decentralized, which listens to other nodes, and collects information, and then sends the request from the client to unoccupied or least used resources/server. High Level architectures applications [1] face performance issues due to the load not getting distributed evenly among the web servers. One of the solutions designed for it was use of dynamic load balancing [1]. This paper designs an architecture where the distributed algorithm uses a load balancer which observe the distributed load centrally, and use various distributed load reallocation policies [6] which reduces the imbalance of load as the load is distributed in a distributed manner thereby it removes the single point of failure, delays, and bottlenecks [1]. This paper [3] discusses the use of adaptive load balancing in an environment. It tries to explain various load balancing algorithms. Round robin algorithm tries to route the requests equally among the instances/instances, it does not care how many connections instance already has previously. Load Connection table is used to check whether server is loaded, normal, under loaded or idle. This table helps in determining to which server the request should be routed too. It also takes into account response time that each request make, in order to find server is behaving the best, for the client to route the request too. [3] Weighted Average Load Balancing technique presented a load balancing technique using preemptive scheduling by Qin [4]. Whenever a new request is received from the client, the algorithm has to do one of the following tasks. It either has to add to the present queue or interrupting current task in execution. Whenever job gets assigned to a node, it checks whether the node has become over loaded or not. If the 13
14 node gets overloaded, algorithm tries to find a new node so that it transfers the most useful jobs from the overloaded node. Presently there are various techniques of load balancing [6] which falls under the category of either inter web servers (there is only one web servers, and algorithms for load balancing are architected to balance load within that) or intra web servers (i.e. it consists of more than one web servers, load balancing algorithms are designed to distribute load among them). This paper uses an algorithm whose aim was to reduce the imbalance in the web server. The way algorithm (VectorDot) [6] used to work is it used to find nodes in the web servers where imbalance has occurred, and then try to shift the load to available free nodes. Imbalance is found out when the threshold on particular node has increased its limit. 1.5 Hypothesis There are many papers focused on how to balance the load on high- level applications using dynamic load balancing. My aim is to compare two load- balancing algorithms on compared based on the load, response time, utilization etc. Performance will be evaluated against load, response time, utilization, bps, tps etc. I will describe the use of cloud services, Web Services, XML and improved algorithm, which will try to find server, which has less load. I will be working through my research with the initial approach of implementing the algorithms. In next phase, I will be comparing them against the parameters defined in the metrics, and in the last phase will increase the load on cloud by creating a lot of virtual instances or by using load technique in SOAP UI. 1.6 Problem and Proposed Solution Problem Statement Different enterprises have different requirements, depending upon their needs and budgets. Some enterprises would like to achieve high performance from and availability of their systems without any concerns about the cost incurred, while smaller enterprises with limited budgets would like to get the most out of their systems. Number of users using Internet is growing at an alarming rate so the need of balancing the traffic on the enterprise applications in order to provide high performance and high availability. There are numbers of load balancing algorithms present, but each and every algorithm face some problem or other. The main goal of load balancing is to achieve route the requests among the web servers with minimum response time. This report focus on improving and comparing the proposed load- balancing algorithm against load balancing algorithm [1] already implemented on cloud servers, and determining proposed algorithm is better suited. 14
15 1.6.2 Proposed Solution The main goal of load balancing is to achieve route the requests among the web servers with minimum response time. This report describes comparing load balancing techniques on the cloud services. I will be using Web services, XML and cloud to design and implement, and then will provide a comparison metrics between the algorithms in terms of query response time, bps, tps, CPU percentage etc. in against the load. The following steps will be followed in order to achieve this process: Step1: JAX- WS web service will be developed which will return response requested by client. SOAP UI tool will behave as the client for the web service, where we will send soap requests and receiving soap responses. Step2: Load balancing algorithm will be developed using Java. There are various types of Load balancing algorithms existing, but I will be using the Dynamic Load balancer [1] and will try to remove its shortcomings with by adding additional logic by introducing server weights logic to it. Dynamic Load Balancer With Least Connections and Fastest Response Time: This paper [1] describes the use of dynamic load balancing. The paper describes the importance of load balancing in system- oriented architecture. The algorithm makes use of combining the logic of least connections present, and fastest response time. Combining them helps in reducing the load. Monitoring agents are used to look at the current activities, and load. They try to classify the instances as over- loaded, balanced and under loaded. Depending upon the instance nature, that instance is being called, which returns the results. As the connections keeps on increasing, this algorithm suffers from in longer run as either new instances need to be created or all the instances are almost on the same side of balanced, or over loaded so it the efficiency reduces. Dynamic Load Balancer with Least Connections and Fastest Response Time With Weight table: This algorithm integrates the concept of dynamic load balancing [1] with weight table assigned to each instance [4]. Each instance has enhanced logic of having a dynamic number assigned for maximum connections defined in the weight table. Whenever the connections limit reaches the threshold, it resets, and starts from fresh. Step3: Amazon Web Services will be used to create EC2 instances. EC2 instances will be created where the load balancer developed in step2 and the web service will be 15
16 deployed in step1. Load balancer algorithm will determine which EC2 instance of web service should be called for sending the request, and getting the response. Step 4: Soap UI tool will be used to increase the load on the EC2 instances, and will be using soap load UI tool to compare the results from the two- load balancer depending on request/response sent. The results will be compared on the basis of performance, bps, tps, round trip time, and CPU utilization. [14]. The dynamics of creating instances, algorithm, and running tasks will be a challenge Road Map To understand the project, the project report is distributed in different sections. Chapter 1 gave a summary about introduction, background, related work, problem statement, and proposed solution. Chapter 2 will provide insight on how the algorithms work, and description. Chapter 3 and 4 will provide description about design, approach, and implementation of project. Chapter 5 describes the experimental results, and with Chapter 6 discuss about the current status, future, and conclusion. 16
17 Chapter 2 2. Algorithms Load Balancing: It means how to fine- tune the traffic among the distributed application so as to avoid the increase in response time when the load has increased on the server. It is mainly important for the high level applications where number of request sent by the user can range from thousands to millions i.e. it s very difficult to predict the number of requests. They need to have many web servers for load balancing. They can be categorized under either static algorithms or dynamic algorithms. Round robin, random etc. are categorized under the static algorithms, whereas least connection, observed, etc. are categorized under dynamic algorithms. [3, 22] 2.1 Dynamic Load Balancer with Least Connections and Fastest Response Time Algorithm LOADBALANCER (request) Input: request Output: response Declare ec2 Instance [][], url, url selectbestserverinstance(request, ec2instances) soapreplyanswer getreply (url, request) return soapreplyanswer addconnection (connection,ec2instance) Input: connection[][], ec2instance[] Output: connection[][] i 0 j 0 for ec2 1 to ec2instance.length do j 0 connection[i, j] ec2[i] j j+1 connection[i, j] 0 i i +1 return connection selectbestserver(request, ec2instances) 17
18 Input: request,ec2instances [][] Output: ec2instance Declare connection [][], responsetime[][], connection addconnection(request, connection,ec2instance) Declare leastconnectionurl, leastresponsetimeurl, secondminconnections i 0 j 1 for connection 1 to connection.length do if connection[i, j] = 0 then connection[i, j] 1 return connection[i, j] else i i +1 Sort the ec2instances of connection in increasing order of connections leastconnectionurl connection[0,0] Sort the ec2instances of responsetime in increasing order of average response time leastresponsetimeurl response[0,0] i 0 if leastconnectionurl = leastresponsetimeurl then return leastconnectionurl else for connection 1 to connection.length do if connection[i,0] = leastresponsetimeurl then numberofconnections connections[i,1 else i i +1 secondminimumconnections connection[1][1] if numberofconnections = secondminimumconnections then return numberofconnections else return responsetimeurl getreply (url, request) Input: url, request Output: reply starttime 0, i 0, j 0 endtime 0 averagetime 0, responsetime 0 starttime getcurrenttime soapreply answerfromwebservice(url,request) 18
19 endtime getcurrenttime responsetime endtime starttime for response 1 to response.length do averagetime += response[i,1] i i +1 responsetime[url][1] averagetime return soapreply ColumnComparator() Input: Output: sorted Array declare columntosort; ColumnComparator(int columntosort) this.columntosort columntosort compare(object o1, Object o2) String[] row1 (String[]) o1; String[] row2 (String[]) o2; return row1[columntosort].compareto(row2[columntosort]); Description This balancer tries to find the minimum loaded server based on the connections and response time. After finding the server, it sends that request to it. The algorithm tries to find server by looking at the number of connections and response time. There are many conditions checked in the algorithm to find out which balancer is performing best at that moment. Monitoring agents are used to look at the connections each server has and response time associated with it and after performing the logistics, best server instance is being sent the request. [1] 19
20 2.2 Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table Algorithm LOADBALANCER (request) Input: request Output: response Declare ec2 Instance [][], url, url selectbestserverinstance(connection, numberarray, responsetime) soapreplyanswer getreply (url, request) return soapreplyanswer addconnection (connection,ec2instance) Input: connection[][], ec2instance[] Output: connection[][] i 0 j 0 for ec2 1 to ec2instance.length do j 0 connection[i, j] ec2[i] j j+1 connection[i, j] 0 i i +1 return connection removeconnection (connection) Input: connection[][] Output: connection[][] connection null connection new Connection return connection addrandomnmbrtoserver (numberarray,ec2instance) Input: numberarray[][], ec2instance[] Output: numberarray[][] i 0 j 0 for ec2 1 to ec2instance.length do j 0 randomint randomgenerator.nextint(10) numberarray[i,j] ec2instance[i] i i +1 20
21 numberarray [i,j] randomint return numberarray removenumbers (numberarray) Input: numberarray[][] Output: numberarray[][] numberarray null numberarray new numberarray return numberarray checkconnectionsagainst WeightOnEachInstance(connection,numberArray) Input: numberarray[][], connection[][] Output:connection i 0, j 1,k 1 checknumberweight false for connection i to connection.length for number k to numberarray.length do if connection[i,1] = numberarray[k,1] then checknumberweight true else checknumberweight false weightnotusedurl connection[i,0] i i +1 k k +1 if checknumberweight true then removenumbers (numberarray); removeconnection(connections); addrandomnmbrtoserver (numberarray, ec2instance); addconnections(connection,ec2instance) selectbestserver(request, ec2instances) Input: request,ec2instances [][] Output: ec2instance Declare connection[][], ec2 Instance[], url, responsetime[][], numberarray[][] connection addconnection(connection,ec2instance) numberarray[] addrandomnmbrtoserver(numberarray,ec2instance) Declare leastconnectionurl, leastresponsetimeurl, secondminconnections, weightnotusedurl, lessnumberurlconnection 21
22 i 0 j 1 connection checkconnectionsagainstweightoneachinstance(connection,numberarray) for connection 1 to connection.length do if connection[i, j] = 0 then connection[i, j] 1 return connection[i, j] else i i +1 Sort the ec2instances of connection in increasing order of connections leastconnectionurl connection[0,0] Sort the ec2instances of responsetime in increasing order of avg rspnse time leastresponsetimeurl response[0,0] i 0 for connection i to connection.length for number k to numberarray.length do if connection[i,1]!= numberarray[k,1] then lessnumberurlconnection[i] = connection [i,1] if leastconnectionurl = leastresponsetimeurl && weightnotusedurl = leastresponsetimeurl then return leastresponsetimeurl else for connection 1 to lessnumberurlconnection.length do if connection[i,0] = leastresponsetimeurl then numberofconnections connections[i,1] else i i +1 secondminimumconnections connection[1][1] if numberofconnections = secondminimumconnections then return numberofconnections else return responsetimeurl getreply (url, request) Input: url, request Output: reply starttime 0, i 0, j 0 22
23 endtime 0 averagetime 0, responsetime 0 starttime getcurrenttime soapreply answerfromwebservice(url,request) endtime getcurrenttime responsetime endtime starttime for response 1 to response.length do averagetime += response[i,1] i i +1 responsetime[url][1] averagetime return soapreply ColumnComparator() Input: Output: sorted Array declare columntosort; ColumnComparator(int columntosort) this.columntosort columntosort compare(object o1, Object o2) String[] row1 (String[]) o1; String[] row2 (String[]) o2; return row1[columntosort].compareto(row2[columntosort]); Description After receiving the request from client, it refers to the connection instance to find the least loaded instance. Response time of least loaded instance is found out, and compared in response- time table and checks whether request should be sent to least loaded server or minimum response- time server. The connection instances are compared against the weight table to see whether connections have reached the number of pre- defined connection that server should have or not. The connection table, response table gets updated. The monitoring service looks to see is there a change in load level on the servers in addition to connections and response time associated. If there is change in weight level, the weight table gets updated with new values. The central server contains weight table has all the value from individual weight tables defined on respective servers. [1,4] 23
24 Chapter 3 3. Design 3.1 Architecture Diagram Client: They are the users who will request news feed from the web service. They act as requestor for the service, and only authorized users are allowed to use this application. They pass the input to the web service, and expect fast output. Web Service: A web service will be developed, which will find out which server is best available for the client to receive output. A load balancing web service will be created, which after finding the server will send the input/request from the client to another web service hosted on that server to get required output before forwarding it to the user. Web Servers created will be EC2 instances, and web service will be developed using JAX- WS, and with tomcat as the server hosted on it. Server1 Client Client Central Server Server2 Client Response Table Server3 : Describes a weight table, which has weights, assigned to each server. : Dynamic Load Balancing Diagram based on connections and response table (Existing) : Dynamic Load balancing based on connections, response time and Number of weights (Proposed) Figure 2: Low Level Architecture 24
25 Figure 3: High Level Architecture 25
26 3.2 Integration Diagram of Load Balancing Algorithms Ec2 Inst Ec2 Inst Ec2 Inst Ec2 Inst Request Request, Ec2Instances Connection, Ec2Instances S : Dynamic Load Balancing Diagram based on connections and response table (Existing) : Dynamic Load balancing based on connections, response time and Number of Weights (Proposed) Figure 4: Integration Diagram of Load Balancing 26
27 3.3 Service Level Diag. of Load balancing Algorithm All EC2 instances are already created LOAD BALANCER (request) CLIENT Client Makes a Request Central Server Receives the Request request addconnection(conctn,ec2ins) Initialize connection Array by adding all ec2instames with 0 Initialize response Array by adding all ec2instames with 0 response time initially addrandomnumbertoserve r(weightarray, ec2instnce) Initialize weight Array by adding all ec2instames with random weights initially. This will serve as load table Check the select best server function to find best server Find least connection connect server Client receives Response Find the least response time server : Load Balancer : Add Connection Function : Add Response Function : Add Weights Function : Select Best Server Function :Get Reply : checkconnectionsagainst WeightOnEachInstance response Compare least connection server with least response time server After getting server instance, it s compared against the weights to check whether the number of connections for that server has reached that weight assigned. If it has reached, we will take that server out of pool, and take next least loaded connection. Otherwise return the Number of connections is compared against weight assigned for each ec2 instance.. If they are equal, resets all connections to zero and reset the weights. Otherwise return the server, which is least, overloaded with connection with respect to weight assigned. Ec2 Instances/Servers where WS is deployed If both provide same server, return the server, otherwise compare response time of lowest 2 servers. If the difference is minimal, return one with lowest connection points, otherwise return the server with lowest response time. Server gets returned. Connection Array and response tome array are modified for that ec2 instance., And sends request to WS Get the response back from web service : Dynamic Load Balancing Diagram based on connections and response table (Existing) : Dynamic Load balancing based on connections, response time and Number of Weights (Proposed) Figure 5: Service Diagram of Load Balancing 27
28 3.4 Evaluation After setting up the project, experiments will be performed to evaluate the performance of the algorithm present, and proposed. Performance criteria will be measured on the basis of adding load to the web service by increasing number of threads, and will measure throughput per second. Soap UI tool will be used to increase the load on the web service. 3.5 Analysis The results produced from the performance will help us to analyze which algorithm is behaving better on basis of increasing the thread count, i.e. increasing number of transactions hitting against the web server. Various strategies will be used, and soap- ui tool will be used to increase the transaction per second, and response time will be monitored for it. Analysis will be based on 99%tile of all the response time, along with maximum time, minimum time and average time too. Standard deviation will also be measured for both the algorithms. Various types of strategies like variance, thread, simple etc. will be used. 28
29 Chapter 4 4. Implementation and Requirements 4.1 Implementation Implementation Strategy The project implemented will be based on three tier architecture i.e. client, web service server, and server level. Client will be implemented which will act as a web service client which access the services hosted on the server level. The middle layer will also be a web service, which will interact with the service on server level. Middle layer will be basically a mapping algorithm, which finds out which service server to ping in order to return the answer to the client in least response time. Load balancing algorithms will be implemented on middle layer i.e. they act as mapping algorithm between clients and server Service Detailed Implementation The client sends request, which is being retrieved by the central server, which decides which server it should send the request to get the response back in minimum time. It consists of 3 sections: Client: Requests the server for the response for the request sent. The request will be in form of SOAP requests, and will get SOAP response. Central Server: Dynamic Load Balancer with Least Connections and Fastest Response Time: This balancer tries to find the minimum loaded server based on the connections and response time. After finding the server, it sends that request to it. The algorithm tries to find server by looking at the number of connections and response time. Monitoring agents are used to look at the connections each server has and response time associated with it and after performing the logistics, best server instance is being sent the request. [1] Dynamic Load Balancer With Least Connections and Fastest Response Time With Weights Assigned: Each cloud tomcat server is assigned with weight through the algorithm. The weights signify total number of connections each tomcat server should handle before stopping accepting new requests of client connection. For example: if a server is assigned with a weight of 30, it can handle 30 connections. The way weights help in balancing the algorithm is it prevents the server to get overloaded with connection requests as it reset the weights on all the server after all servers have 29
30 Server: reached its assigned weight. Resetting of weights resets the load weight table, as well as the connections table i.e. it helps in overcoming the overloading of objects as they got re- initialized. The client sends the request to load balancer; it refers to the connection instance to find the least weighted instance. Response time of least connection instance is found out, and compared in response- time table and checks whether request should be sent to least connection server or minimum response- time server. The connection instances are compared against the weight table [4] to see whether connections have reached the number of pre- defined connection that server should have or not, and if it has, no new request will be connected against that particular server. The connection table, response table gets updated. If all the servers have reached respective weights assigned, the weight, and the connection get reset, and it start from the scratch, i.e. all the objects used in algorithm again initialize, thereby reducing the complexity of trying to find least connection which is used to face in algorithm [4]. The monitoring service looks to see is there a change in load level on the servers in addition to connections and response time associated. If there is change in load level, the weight table gets updated with new values. The central server weight table has all the value from individual weight tables defined on respective servers. The main objective of this section is to process soap request. Web Service will be hosted on these servers. After processing the soap request sent by the mapping algorithm through client, it sends back the response to the client, which he has requested for Objective The main objective of the project is to improve the balancing algorithm. In this project, I will try to improve the algorithm Dynamic Load Balancer with Least Connections and Fastest Response Time mentioned in [1]. The main disadvantage of the current implemented algorithm is as the connections keeps on increasing, this algorithm suffers from in longer run as either new instances need to be created or all the instances are almost on the same side of balanced, or over loaded so it the efficiency reduces. In order to remove the shortcoming discussed above, the new proposed algorithm will integrate the concept of dynamic load balancing [1] with weight table assigned to each instance [4]. Each instance has enhanced logic of having a dynamic number assigned for maximum connections defined in the weight table. Whenever the connections limit reaches the threshold, it resets, and starts from fresh, thereby increasing the efficiency. The new proposed algorithm is termed as Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table. 30
31 4.1.3 SOAP Implementation The project will be using SOAP messages for request/response data between client and the server. The request will be send to the load balancer algorithm running on tomcat, which will find out which will try to find out which server is behaving best at the moment, and will forward the request to that server, which will respond back with the output. In total 3 web services will be developed. The first two web services developed are for the load balancer algorithm, and third web service is for the program, which will return the output back to client. Both load balancer web service will be calling third program web service for the output responded back to the client. 4.2 Requirements Hardware The system hardware used for the implementation will be my local machine, which has windows and Linux both configured. The hardware details are as follow: Windows 7, i5 processor Linux Env 4Gb RAM, 64 bit Operating System Amazon Cloud Software Tomcat 6 Java 1.6 Web Service will be built using JAX- WS 3.0 API on Java1.6. SoapUI client will be used to access the web service. Eclipse IDE will be used for development purposes. Load balancing algorithms will be developed using Java 1.6. Amazon EC2 virtual instances will be created [12]. Jars will be downloaded. UTF- 8 Encoding will be used. Spring 3.0 HP Load Runner Tool Metrics The core of any comparison study is metrics. Several studies have used different metrics (such as performance, storage etc.) to evaluate and compare different services [7] [8] [9]. The project will consider metrics, which are relevant for this 31
32 comparison. The various parameters that can be considered are the performance, response time, scalability, throughput, and resource utilization. Analysis will be based on 99%tile of all the response time, along with maximum time, minimum time and average time too. Standard deviation will also be measured for both the algorithms. Various types of strategies like variance, thread, simple etc. will be used Round Trip Time The end- to- end response time will be compared between the two load balancers by sending requests through the web service. For good load balancer, end- to- end response time should be as low as possible. The Response time will be measured against the load sent on two load balancers Memory Utilization Memory utilization will be compared for the load balancers against the load. Memory utilization can be measured through JVM. To determine whether load balancer is better, it should be consuming less memory, and should do garbage collection at timely manner to remove all unused object. It should not do memory leak Cost Cost for utilizing different cloud services will be estimated as its one of the main deciding factor for the enterprises. The cost serves as one of the main deciding factor because enterprise has budget constraints, and they have to look at the best available option, which will be beneficial for the industry Scalability This is an ability of an algorithm to perform load balancing with any finite number of nodes. The scalability will be tested against the load balancer algorithm by increasing the number of nodes, or number of transactions per second to check what the upper limit is till the system starts degrading. The higher the scalable factor, better the performance is TPS It stands for transaction made per second. For example: a test case that runs for 10 seconds and handled 100 requests will get tps of 10. Algorithm, which provides higher tps, is considered better. [18] BPS It stands for bytes per second. For example: test case running for 10 seconds and handled bytes will get BPS of Algorithm, which provides higher bps, is considered better. [18] 32
33 These parameters will be compared with the data set that will be passed to the URL s where the load balancer is installed which will further send the requests to the web service and get the data back. Data set will be in form of soap requests. We will increase the number of thread count of the soap request in order to overload the server. The data set will be of the format: Figure 6: Soap Request The SOAP request will be sent through SOAPUI, and we will be using SOAP UI Load Test tool [15] to get the metrics we have to compare. The metrics will be compared against the parameters bps, tps, avg time (rtt), memory utilization. Different Load strategies will be used for each parameter for evaluating the response from the load balancer. The strategies are simple, variance, burst, thread, grid [15] Simple Strategy This strategy runs by adding definite amount of thread with a definite delay between each thread run for server. For example for a test to run with test with 5 threads with 5 seconds delay, we initialize thread array to 5, delay to 5000ms. [18] Variance Strategy This strategy fluctuate the thread amount over a course of duration in a sawtooth manor; we set interval to the desired value and the variance value which determines how much thread amount should decrease or increase during the test duration. [18] Thread Strategy This test increases the amount of thread from one level to another over the run test duration. The main purpose of this strategy is to find at which thread count the test can attain the highest transaction per second, and it can also be used to find the level of thread count where the performance starts degrading. [18] 33
34 Chapter 5 5. Analysis 5.1 Basic Set Up To check how the algorithms are behaving, and to see whether the proposed algorithm is performing better than the original algorithm, performance tests were being carried out to check the performance. The tests were performed with different parameters to find the performance. We use soap Ui tool, and add the wsdl of the load balancer algorithm to it. Every request will hit the load balancer algorithm, which has the various EC2 instances on which a web service is deployed which provide the response that client is requesting for. The main aim of this performance test is to test the middle layer. Middle layer consists of load balancer, which routes the request. The efficiency will be determined by comparing tps, bps, memory usage by JVM (will use top command) etc. We will compare efficiency of proposed algorithm, and will show that using proposed algorithm Dynamic Load Balancer with Least Connections and Fastest Response Time with weight table is better, and efficient. WSDL generated are shown below: Figure 7: Dynamic Load Balancer with Least Connections and Fastest Response Time WSDL 34
35 Figure 8: Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table WSDL The architecture design used in this project is three- tier architecture i.e. client, mapping server, and response server. The client sends request, which is being retrieved by the central server, which decides which server it should send the request to get the response back in minimum time. The main objective of this project is to develop the mapping algorithm for the central server. There will be two mapping algorithm developed, one with dynamic load balancer with least connections and fastest response time, and another with dynamic load balancer with least connection, fastest response time with weights. The way the dynamic load balancer with least connection and fastest response time works [1] is the balancer tries to find minimum loaded server based on connection and response time. After finding the server, it sends that request to it, and gets the response, and sends it back to client. The way the proposed algorithm (Dynamic Load Balancer with Least Connections and Fastest Response Time with Weights Assigned) [1,4] works is each cloud tomcat server in the algorithm are assigned with equal weights. The weights signify total number of connections each tomcat server should handle before stopping accepting new requests of client connection. After receiving the request from client, it refers to the connection instance to find the least weighted instance. Response time of least connection instance is found out, and compared in response- time table and checks whether request should be sent to least connection server or minimum response- time server. The connection instances are compared against the weight table to see whether connections have reached the number of pre- defined connection that server should have or not, and if it has, no new request will be connected against that particular server. The connection table, response table gets updated. If all the servers have reached respective weights assigned, the weight, and the connection 35
36 gets reset, and it start from the scratch, i.e. all the objects used in algorithm again initialize. 5.2 Performance Test The load balancer web service is deployed on tomcat 6.0 using java 1.6. Soap Ui tool is used to send the request to the load balancer web service, and load strategy is used to carry out the performance test. Various strategies were used to test against the load balancers (original and proposed), to find which load balancing algorithm was behaving better. We will be running three different load tests to compare our performance test results. The load tests that will be run will be simple, thread and variance mode. The thread level will be adjusted in each run of the load test along with the time too. The Parameters that will be compared will be minimum time, average time, and tps (transactions per second). The main focus of each test will be to see whether the new proposed load balancing algorithm have better tps (transactions per second) rate than original. Higher the tps better will be the algorithm performance Simple Strategy This strategy runs by adding definite amount of thread with a definite delay between each thread run for server. For example: for a test to run with test with 5 threads with 5 seconds delay, we initialize thread array to 5, delay to 5000ms. [18] Below are the results obtained after running tests against algorithm. Average Time (ms) Maximum Time (ms) TPS Thread Count Time (Sec) Original Algorithm Proposed Algorithm Original Algorithm Proposed Algorithm Original Algorithm Proposed Algorithm Table 1: Simple Strategy results comparison The table was plotted with graphs with left axis contains average time (ms) v/s right axis as thread count. 36
37 Original Algorithm Proposed Algorithm Figure 9: Average Time (Strategy Mode) By running the tests for Strategy mode against both algorithms, it was seen average response time of proposed algorithm performs better than the algorithm [3]. Response time described here is from client end. Lower response time is one of the features that are most wanted, so by analyzing the result from the graph, it can be said that proposed algorithm behaves better Figure 10: Maximum Time (Strategy Mode) Original Algorithm Proposed Algorithm This test is to find the outliers and to find whether the outlier s time is reduced or not. From the graph, we can analyze that by increasing the thread count; the maximum time keeps on decreasing, thereby concluding the maximum time reduces with increase in thread when compared proposed v/s original algorithm [3] Figure 11: TPS (Strategy Mode) Original Algorithm Proposed Algorithm The above test compares tps for the proposed algorithm against original algorithm. For this test, there is not much change in tps, but if we it in graph proposed algorithm performs marginally better, and it is linearly increasing with increase in number of threads passed against each algorithm. The above graphs compares performance of algorithms ran against simple strategy mode. And proposed algorithm has performed better in average time, max time and little better for transaction response time (it can be checked from table entries). 37
38 Thread Count The next set of performance test ran with thread strategy against both algorithms Thread Strategy Time (Sec) Original Algorithm Average Time (ms) Proposed Algorithm Original Algorithm Maximum Time (ms) Proposed Algorithm Original Algorithm TPS Proposed Algorithm Table 2: Thread Strategy results comparison The table was plotted with graphs with left axis contains average time (ms) v/s right axis as thread count. This test increases the amount of thread from one level to another over the run test duration. The main purpose of this strategy is to find at which thread count the test can attain the highest transaction per second. I ran the test against thread count till Figure 12: Average Time (Thread Mode) Original Algorithm Proposed Algorithm Average time is termed as response time for the client, and from this test it is concluded original algorithm performs marginally better than proposed algorithm, but in real time response time almost looks equal if removed one outlier Figure 13: Maximum Time (Thread Mode) Original Algorithm Proposed Algorithm 38
39 This test is to find the outliers and to find whether the outlier s time is reduced or not. From the graph, we can analyze that by increasing the thread count; the maximum time keeps on decreasing, thereby concluding the maximum time reduces with increase in thread when compared with original algorithm [3] Figure 14: TPS (Thread Mode) Original Algorithm Proposed Algorithm Thread Count The above test shows proposed algorithm behaves much better than the original algorithm [3]. It can be seen with the increase in number of threads the tps of both algorithms is increasing considerably but the rate at which tps proposed algorithm increased is more than if compared against original algorithm [3]. Higher the tps better is the performance of algorithm, and also with increase in number of threads tps is increasing which also make it more scalable. This tests helps in summarizing that proposed algorithm is much more scalable than original algorithm [3]. The next set of performance test ran with variance strategy against both algorithms Variance Strategy (variance 0.5) Time (Sec) Original Algorithm Average Time (ms) Proposed Algorithm Original Algorithm Maximum Time (ms) Proposed Algorithm Original Algorithm TPS Proposed Algorithm Table 3: Variance Strategy results comparison The table was plotted with graphs with left axis contains average time (ms) v/s right axis as thread count. The test was performed with a variance of 0.5 for all cases. 39
40 Figure 15: Average Time (Variance Mode) Original Algorithm Proposed Algorithm By running the tests for variance mode against both algorithms, it was seen average response time performs much better for proposed algorithm than the original algorithm [3] with increase in thread count. Response time described here is from client end. Lower response time is one of the features that are most wanted, so by analyzing the result from the graph, it can be said that proposed algorithm behaves better when the number of thread increases which is a nice feature Figure 16: Maximum Times (Variance Mode) Original Algorithm Proposed Algorithm This test is to find the outliers and to find whether the outlier s time is reduced or not. From the graph, we can analyze that by increasing the thread count; the maximum time keeps on decreasing, thereby concluding the maximum time reduces with increase in thread when compared with original algorithm [3]. This test show proposed algorithm behaved better than original [3] Figure 17: TPS (Variance Mode) Original Algorithm Proposed Algorithm Transactions per second (tps) for proposed algorithm are much better than the original algorithm. Higher the tps better is the performance. By increasing the number of threads, tps increases at a much higher rate for proposed algorithm. Performance degrades for both algorithms at thread count of
41 5.2.4 Testing Using HP Load Runner Tool I download HP Load runner trial version [21] to test the performance of algorithms in terms of 99%tile of response time and TPS. 99%tile response time will indicate which algorithm performs always better in terms of response time. VUsers Time (min) 99%tile Response Time (ms) Original Algorithm Proposed Algorithm Original Algorithm TPS Proposed Algorithm Table 4: HP Load Runner Results Comparison The table was plotted with graphs with left axis contains average time (ms) v/s right axis as VUsers Original Algorithm Proposed Algorithm Figure 18: 99%tile Response Time (HP LoadRunner) The graph was plotted for the 99%tile response time for both algorithms, and it is evident from the graph that proposed algorithm performed much better as compared. 99% of the time, it responded faster by 5-8 ms. Lower the response time, better is the performance considered. 41
42 Figure 19: TPS (HP LoadRunner) Original Algorithm Proposed Algorithm The graph indicates that tps rate is better for proposed algorithm against original algorithm [3]. The better tps rate means algorithm can process more transactions per second, and is scalable. From the above two graphs, its concluded that proposed algorithm has better tps, and better 99% tile response time, which states proposed algorithm is performing better compared to original algorithm for the same load and criteria. From all the performance tests (simple, thread, variance and HP Load Runner) conducted against both original algorithm [3] and proposed algorithm [3,4], it is seen that proposed algorithm has always been better in respect to increase in tps, lower response time, and low maximum value compared to original algorithm. All the three tests (average time, 99%tile, maximum time, and tps) performed better for proposed algorithm. 42
43 Chapter 6 6. Conclusion The main objective is this report was to improve the load- balancing algorithm [1] as it used to suffer from performance issues because servers used to get overloaded as it had to maintain the objects in memory in order to determine which the best available server for the client. In this report, concepts of server- weights [4] was added to algorithm [1] so that it can perform better in terms of scalability, tps, response time etc. By adding this concept of static server- weights [4], the objective of the project i.e. find the most effective load- balancing algorithm, based on performance, measured in terms of higher TPS and low response time, for the industry as well as more scalable was evaluated. After analyzing the performance results, its seen that Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table (proposed Algorithm) [1, 4] perform better than the Dynamic Load Balancer with Least Connections and Fastest Response Time (original algorithm) [1]. The better the performance, more is the transaction being processed in the same time. Load Balancing means routing the requests sent by the client, and getting response in fast manner. The proposed algorithm not only processes more transactions, it also reduces or does not change the average time (round trip time). Not only the proposed algorithm has a better tps, it has better scalability too as seen from the performance results. Performance results also confirmed that 99%tile of all the responses received, the new algorithm performs better Current Status The main objective of this report was to determine which load balancing algorithm should be used which can provide better scalability, higher tps, and better response time. I tried to improve the load balancing technique [1] by adding the concept of weighted servers [4]. Both Load Balancing Techniques were implemented as web service, and comparison of their performance was done. The web service will act as global traffic manager (GTM), which will be called by the client, and algorithm will act as local traffic manager (LTM) routing the requests to the appropriate cloud servers to get the result back to client in minimum time. Algorithms were evaluated based on the performance tests executed on them. Performance tests like simple strategy, thread strategy, variance strategy, and hp load runner tests were executed against load balancing web service for tps, avg response time, maximum time, and 99%tile response time. All the tools that were used for performing performance tests were open source, and free. HP Load runner tool was based on free trial. Based on the analysis of performance results, it was seen that proposed algorithm 43
44 (Dynamic Load Balancer with Least Connections and Fastest Response Time with Weight table) has performed better than the original algorithm (Dynamic Load Balancer with Least Connections and Fastest Response Time) for criteria of 99%tile response time, tps, max time etc. It has better tps, higher scalability, and response time improves too. 6.2 Future Work The future work for this master s project can be extended to various areas. Some of the few areas are: The proposed load balancing technique uses the concept of static weights. It can be extended to dynamic weights where weights are assigned to the server dynamically based on the server performance. The server, which is performing, better is assigned higher weight as compared to less performing server. This can be achieved by adding monitoring agents on the each server, which monitors the performance. Concept of threads can be used too. Threads will be used in load balancing algorithms where requests sent by client are added in a service pool queue, and thread fetches requests from it, performs the requested operation. It follows the lifecycle of thread. It should be made sure that all the threads are mutually exclusive of each other so that the balancing algorithm does not go in deadlock state. JMS can be used to. Instead of using web service for the load- balancing algorithm, they can be deployed as standalone applications on the servers/instances, which have the concept of JMS, coded in it. JMS is termed as Java Messaging Service. The framework will use JMS as mechanism for the clients to send the messages. Load balancing techniques can be compared against web service v/s JMS. 6.3 Lessons Learned Many things were learnt during implementation of the master s project. Algorithms were developed using JAX- WS, and use SOAP tool for load testing was beneficial. Web service was created from top to bottom approach. I learnt how to write script in order to use HP Load runner tool to call the web service. 44
45 BIBLIOGRAPHY [1] Azzedine Boukerche, Robson Eduardo De Grande, Dynamic Load Balancing Using Grid Services for HLA- Based Simulations on Large- Scale Distributed Systems, Proceedings of the th IEEE/ACM International Symposium on Distributed Simulation and Real Time Applications, [2] Patrick Wendell, Joe Wenjie Jiang, Micheal J. Freedman, and Jennifer Rexford. DONAR : Decentralized Server Selection for Cloud Services. In Proceedings of the ACM SIGCOMM 2010 conference on SIGCOMM 10. ACM, New Yrk, NY, USA, [3] Gowtham Kanagaraj, Naveen Shanmugasundaram And Sathish Prakash Adaptive Load Balancing Algorithm Using Service Queue 2nd International Conference on Computer Science and Information Technology (ICCSIT'2012) Singapore April 28-29, 2012 [4] Xiao Qin, Hong Jiang, Yifeng Zhu, David R. Swanson, Boosting Performance for I/OIntensive Workload by Preemptive Job Migrations in a Cluster System, Proceedings of the 15 th Symposium on Computer Architecture and High Performance Computing, [5] Aameek Singh, Madhukar Korupolu, Dushmanta Mohapatra, ServerStorage Virtualization: Integration and Load Balancing in Web server, Proceedings of the 2008 ACM/IEEE conference on Supercomputing, [6] Hyotaek Lim Dynamic Load Balancing and Network Monitoring in iata Protocol for Mobile Appliances, Multimedia and Ubiquitous Engineering (MUE), th International Conference on Aug [7] Cardellini, V. Dynamic load balancing on Web- server systems, Internet Computing, IEEE [8] M.Pathan, C. Vecchiola and R.Buyya, Load and proximity aware request- redirection for dynamic load distribution in peering CDNs, in OTM, Nov [9] Klaithem Al Nuaimi, Nader Mohamed, Mariam Al Nuaimi and Jameela Al- Jaroodi A Survey of Load Balancing in Cloud Computing: Challenges and Algorithm 2012 Second Symposium on Network Cloud Computing and Applications [10] A. Iosup, S. Ostermann, M. Yigitbasi, R. Prodan, T. Fahringer, and D. Epema. Performance analysis of cloud computing services for many- tasks scientific computing. In Parallel and Distributed Systems, IEEE Transactions on, [11] Allan, D. Farkas, J.; Mansfield, S. Intelligent load balancing for shortest path 45
46 bridging, 2012 [12] M. Armbrust, A. Fox, R. Griffith, A. D. Joseph, R. Katz, A. Konwinski, G. Lee, D. A.Patterson, A. Rabkin, I. Stoica, and M. Zaharia. Above the clouds: A berkeley viewof cloud computing. In EECS Department, University of California, Berkeley, http: // html. [13] T. Harris. Cloud computing services a comparison. In., com/pdfs/comparison%20of%20cloud%20computing%20services.pdf. [14]- Computing/Amazon- Web- Services- Delivers- High- Performance- Computing- Solution / [15] How to create Amazon EC2 Instance : bin/view/blog/blogentry201105x2 [16]Building Web services with JAX- WS [17] SOAP UI Tool: [18] Execute Load Test: Testing/loadtest- execution.html [19] M.Pathan, C. Vecchiola and R.Buyya, Load and proximity aware request- redirection for dynamic load distribution in peering CDNs, in OTM, Nov [20] Amazon Web Service Deliver High Performance: Computing/Amazon- Web- Services- Delivers- High- Performance- Computing- Solution / [21] HP Load Runner Tool: solutions/software.html?compuri= &action=downloadsoftware&jumpid=p s_r11374_us/en/large/eb/loadrnr_googlesemdg&tcmid=tcm: &originid=8-2ldremr§ionid=alm_ito&k_clickid=ams cb b928- f3d d159d7#! [22] Intro to Load Balancing for Developers: to- load- balancing- for- developers- ndash- the- algorithms#.un0jshb29no [23] New To SOA and Web Services: [24] Amazon Web Services: 46
A Middleware Strategy to Survive Compute Peak Loads in Cloud
A Middleware Strategy to Survive Compute Peak Loads in Cloud Sasko Ristov Ss. Cyril and Methodius University Faculty of Information Sciences and Computer Engineering Skopje, Macedonia Email: [email protected]
MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM?
MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? Ashutosh Shinde Performance Architect [email protected] Validating if the workload generated by the load generating tools is applied
Liferay Portal Performance. Benchmark Study of Liferay Portal Enterprise Edition
Liferay Portal Performance Benchmark Study of Liferay Portal Enterprise Edition Table of Contents Executive Summary... 3 Test Scenarios... 4 Benchmark Configuration and Methodology... 5 Environment Configuration...
Performance Testing of Java Enterprise Systems
Performance Testing of Java Enterprise Systems Katerina Antonova, Plamen Koychev Musala Soft Why Performance Testing? Recent studies by leading USA consultancy companies showed that over 80% of large corporations
Performance Analysis of Web based Applications on Single and Multi Core Servers
Performance Analysis of Web based Applications on Single and Multi Core Servers Gitika Khare, Diptikant Pathy, Alpana Rajan, Alok Jain, Anil Rawat Raja Ramanna Centre for Advanced Technology Department
Application Performance Testing Basics
Application Performance Testing Basics ABSTRACT Todays the web is playing a critical role in all the business domains such as entertainment, finance, healthcare etc. It is much important to ensure hassle-free
Investigations on Hierarchical Web service based on Java Technique
Investigations on Hierarchical Web service based on Java Technique A. Bora, M. K. Bhuyan and T. Bezboruah, Member, IAENG Abstract We have designed, developed and implemented a hierarchical web service
Architecting ColdFusion For Scalability And High Availability. Ryan Stewart Platform Evangelist
Architecting ColdFusion For Scalability And High Availability Ryan Stewart Platform Evangelist Introduction Architecture & Clustering Options Design an architecture and develop applications that scale
EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications
ECE6102 Dependable Distribute Systems, Fall2010 EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications Deepal Jayasinghe, Hyojun Kim, Mohammad M. Hossain, Ali Payani
How To Test A Web Server
Performance and Load Testing Part 1 Performance & Load Testing Basics Performance & Load Testing Basics Introduction to Performance Testing Difference between Performance, Load and Stress Testing Why Performance
Throughput Capacity Planning and Application Saturation
Throughput Capacity Planning and Application Saturation Alfred J. Barchi [email protected] http://www.ajbinc.net/ Introduction Applications have a tendency to be used more heavily by users over time, as the
LOAD BALANCING TECHNIQUES FOR RELEASE 11i AND RELEASE 12 E-BUSINESS ENVIRONMENTS
LOAD BALANCING TECHNIQUES FOR RELEASE 11i AND RELEASE 12 E-BUSINESS ENVIRONMENTS Venkat Perumal IT Convergence Introduction Any application server based on a certain CPU, memory and other configurations
Holistic Performance Analysis of J2EE Applications
Holistic Performance Analysis of J2EE Applications By Madhu Tanikella In order to identify and resolve performance problems of enterprise Java Applications and reduce the time-to-market, performance analysis
International Journal of Engineering Research & Management Technology
International Journal of Engineering Research & Management Technology March- 2015 Volume 2, Issue-2 Survey paper on cloud computing with load balancing policy Anant Gaur, Kush Garg Department of CSE SRM
An Oracle White Paper July 2011. Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide
Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide An Oracle White Paper July 2011 1 Disclaimer The following is intended to outline our general product direction.
OpenFlow Based Load Balancing
OpenFlow Based Load Balancing Hardeep Uppal and Dane Brandon University of Washington CSE561: Networking Project Report Abstract: In today s high-traffic internet, it is often desirable to have multiple
An objective comparison test of workload management systems
An objective comparison test of workload management systems Igor Sfiligoi 1 and Burt Holzman 1 1 Fermi National Accelerator Laboratory, Batavia, IL 60510, USA E-mail: [email protected] Abstract. The Grid
Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing
www.ijcsi.org 227 Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing Dhuha Basheer Abdullah 1, Zeena Abdulgafar Thanoon 2, 1 Computer Science Department, Mosul University,
Task Scheduling for Efficient Resource Utilization in Cloud
Summer 2014 Task Scheduling for Efficient Resource Utilization in Cloud A Project Report for course COEN 241 Under the guidance of, Dr.Ming Hwa Wang Submitted by : Najuka Sankhe Nikitha Karkala Nimisha
Enterprise Edition Scalability. ecommerce Framework Built to Scale Reading Time: 10 minutes
Enterprise Edition Scalability ecommerce Framework Built to Scale Reading Time: 10 minutes Broadleaf Commerce Scalability About the Broadleaf Commerce Framework Test Methodology Test Results Test 1: High
JBoss Seam Performance and Scalability on Dell PowerEdge 1855 Blade Servers
JBoss Seam Performance and Scalability on Dell PowerEdge 1855 Blade Servers Dave Jaffe, PhD, Dell Inc. Michael Yuan, PhD, JBoss / RedHat June 14th, 2006 JBoss Inc. 2006 About us Dave Jaffe Works for Dell
ADAPTIVE LOAD BALANCING FOR CLUSTER USING CONTENT AWARENESS WITH TRAFFIC MONITORING Archana Nigam, Tejprakash Singh, Anuj Tiwari, Ankita Singhal
ADAPTIVE LOAD BALANCING FOR CLUSTER USING CONTENT AWARENESS WITH TRAFFIC MONITORING Archana Nigam, Tejprakash Singh, Anuj Tiwari, Ankita Singhal Abstract With the rapid growth of both information and users
19.10.11. Amazon Elastic Beanstalk
19.10.11 Amazon Elastic Beanstalk A Short History of AWS Amazon started as an ECommerce startup Original architecture was restructured to be more scalable and easier to maintain Competitive pressure for
3/21/2011. Topics. What is load balancing? Load Balancing
Load Balancing Topics 1. What is load balancing? 2. Load balancing techniques 3. Load balancing strategies 4. Sessions 5. Elastic load balancing What is load balancing? load balancing is a technique to
Performance Testing and Optimization in Web-Service Based Applications
Performance Testing and Optimization in Web-Service Based Applications Mesfin Mulugeta [email protected] Sr. Software Performance Engineer Goals of the Presentation Brief introduction to software
CHAPTER 1 - JAVA EE OVERVIEW FOR ADMINISTRATORS
CHAPTER 1 - JAVA EE OVERVIEW FOR ADMINISTRATORS Java EE Components Java EE Vendor Specifications Containers Java EE Blueprint Services JDBC Data Sources Java Naming and Directory Interface Java Message
Summer Internship 2013 Group No.4-Enhancement of JMeter Week 1-Report-1 27/5/2013 Naman Choudhary
Summer Internship 2013 Group No.4-Enhancement of JMeter Week 1-Report-1 27/5/2013 Naman Choudhary For the first week I was given two papers to study. The first one was Web Service Testing Tools: A Comparative
Deploying the BIG-IP LTM with the Cacti Open Source Network Monitoring System
DEPLOYMENT GUIDE Deploying the BIG-IP LTM with the Cacti Open Source Network Monitoring System Version 1.0 Deploying F5 with Cacti Open Source Network Monitoring System Welcome to the F5 and Cacti deployment
The International Journal Of Science & Technoledge (ISSN 2321 919X) www.theijst.com
THE INTERNATIONAL JOURNAL OF SCIENCE & TECHNOLEDGE Efficient Parallel Processing on Public Cloud Servers using Load Balancing Manjunath K. C. M.Tech IV Sem, Department of CSE, SEA College of Engineering
Delivering Quality in Software Performance and Scalability Testing
Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,
Cloud computing - Architecting in the cloud
Cloud computing - Architecting in the cloud [email protected] 1 Outline Cloud computing What is? Levels of cloud computing: IaaS, PaaS, SaaS Moving to the cloud? Architecting in the cloud Best practices
LBPerf: An Open Toolkit to Empirically Evaluate the Quality of Service of Middleware Load Balancing Services
LBPerf: An Open Toolkit to Empirically Evaluate the Quality of Service of Middleware Load Balancing Services Ossama Othman Jaiganesh Balasubramanian Dr. Douglas C. Schmidt {jai, ossama, schmidt}@dre.vanderbilt.edu
ELIXIR LOAD BALANCER 2
ELIXIR LOAD BALANCER 2 Overview Elixir Load Balancer for Elixir Repertoire Server 7.2.2 or greater provides software solution for load balancing of Elixir Repertoire Servers. As a pure Java based software
Mohammed Khan SUMMARY
Mohammed Khan E-mail: [email protected] Phone: 347-878-1170 SUMMARY Over 5 years of diversified experience as a. Experience includes requirement analysis, manual testing and automation and quality
Informatica Data Director Performance
Informatica Data Director Performance 2011 Informatica Abstract A variety of performance and stress tests are run on the Informatica Data Director to ensure performance and scalability for a wide variety
An Approach to Load Balancing In Cloud Computing
An Approach to Load Balancing In Cloud Computing Radha Ramani Malladi Visiting Faculty, Martins Academy, Bangalore, India ABSTRACT: Cloud computing is a structured model that defines computing services,
McAfee Enterprise Mobility Management 12.0. Performance and Scalability Guide
McAfee Enterprise Mobility Management 12.0 Performance and Scalability Guide Contents Purpose... 1 Executive Summary... 1 Testing Process... 1 Test Scenarios... 2 Scenario 1 Basic Provisioning and Email
Project Proposal. Data Storage / Retrieval with Access Control, Security and Pre-Fetching
1 Project Proposal Data Storage / Retrieval with Access Control, Security and Pre- Presented By: Shashank Newadkar Aditya Dev Sarvesh Sharma Advisor: Prof. Ming-Hwa Wang COEN 241 - Cloud Computing Page
Amazon Web Services Primer. William Strickland COP 6938 Fall 2012 University of Central Florida
Amazon Web Services Primer William Strickland COP 6938 Fall 2012 University of Central Florida AWS Overview Amazon Web Services (AWS) is a collection of varying remote computing provided by Amazon.com.
Technology and Cost Considerations for Cloud Deployment: Amazon Elastic Compute Cloud (EC2) Case Study
Creating Value Delivering Solutions Technology and Cost Considerations for Cloud Deployment: Amazon Elastic Compute Cloud (EC2) Case Study Chris Zajac, NJDOT Bud Luo, Ph.D., Michael Baker Jr., Inc. Overview
Oracle WebLogic Server 11g Administration
Oracle WebLogic Server 11g Administration This course is designed to provide instruction and hands-on practice in installing and configuring Oracle WebLogic Server 11g. These tasks include starting and
Mark Bennett. Search and the Virtual Machine
Mark Bennett Search and the Virtual Machine Agenda Intro / Business Drivers What to do with Search + Virtual What Makes Search Fast (or Slow!) Virtual Platforms Test Results Trends / Wrap Up / Q & A Business
MID-TIER DEPLOYMENT KB
MID-TIER DEPLOYMENT KB Author: BMC Software, Inc. Date: 23 Dec 2011 PAGE 1 OF 16 23/12/2011 Table of Contents 1. Overview 3 2. Sizing guidelines 3 3. Virtual Environment Notes 4 4. Physical Environment
MAGENTO HOSTING Progressive Server Performance Improvements
MAGENTO HOSTING Progressive Server Performance Improvements Simple Helix, LLC 4092 Memorial Parkway Ste 202 Huntsville, AL 35802 [email protected] 1.866.963.0424 www.simplehelix.com 2 Table of Contents
www.progress.com DEPLOYMENT ARCHITECTURE FOR JAVA ENVIRONMENTS
DEPLOYMENT ARCHITECTURE FOR JAVA ENVIRONMENTS TABLE OF CONTENTS Introduction 1 Progress Corticon Product Architecture 1 Deployment Options 2 Invoking Corticon Decision Services 4 Corticon Rule Engine 5
BASICS OF SCALING: LOAD BALANCERS
BASICS OF SCALING: LOAD BALANCERS Lately, I ve been doing a lot of work on systems that require a high degree of scalability to handle large traffic spikes. This has led to a lot of questions from friends
Dynamic Resource allocation in Cloud
Dynamic Resource allocation in Cloud ABSTRACT: Cloud computing allows business customers to scale up and down their resource usage based on needs. Many of the touted gains in the cloud model come from
Various Schemes of Load Balancing in Distributed Systems- A Review
741 Various Schemes of Load Balancing in Distributed Systems- A Review Monika Kushwaha Pranveer Singh Institute of Technology Kanpur, U.P. (208020) U.P.T.U., Lucknow Saurabh Gupta Pranveer Singh Institute
Availability and Load Balancing in Cloud Computing
2011 International Conference on Computer and Software Modeling IPCSIT vol.14 (2011) (2011) IACSIT Press, Singapore Availability and Load Balancing in Cloud Computing Zenon Chaczko 1, Venkatesh Mahadevan
FAQ: Data Services Real Time Set Up
FAQ: Data Services Real Time Set Up Assumptions How to load balancer Real-Time job? How does a Web Service Real-Time job utilize multiple job servers? Setup with at least two Web Servers and two Job Servers
SOLUTION BRIEF: SLCM R12.7 PERFORMANCE TEST RESULTS JANUARY, 2012. Load Test Results for Submit and Approval Phases of Request Life Cycle
SOLUTION BRIEF: SLCM R12.7 PERFORMANCE TEST RESULTS JANUARY, 2012 Load Test Results for Submit and Approval Phases of Request Life Cycle Table of Contents Executive Summary 3 Test Environment 4 Server
my forecasted needs. The constraint of asymmetrical processing was offset two ways. The first was by configuring the SAN and all hosts to utilize
1) Disk performance When factoring in disk performance, one of the larger impacts on a VM is determined by the type of disk you opt to use for your VMs in Hyper-v manager/scvmm such as fixed vs dynamic.
Public Cloud Partition Balancing and the Game Theory
Statistics Analysis for Cloud Partitioning using Load Balancing Model in Public Cloud V. DIVYASRI 1, M.THANIGAVEL 2, T. SUJILATHA 3 1, 2 M. Tech (CSE) GKCE, SULLURPETA, INDIA [email protected] [email protected]
Topics. 1. What is load balancing? 2. Load balancing techniques 3. Load balancing strategies 4. Sessions 5. Elastic load balancing
Load Balancing Topics 1. What is load balancing? 2. Load balancing techniques 3. Load balancing strategies 4. Sessions 5. Elastic load balancing What is load balancing? load balancing is a technique to
Cognos8 Deployment Best Practices for Performance/Scalability. Barnaby Cole Practice Lead, Technical Services
Cognos8 Deployment Best Practices for Performance/Scalability Barnaby Cole Practice Lead, Technical Services Agenda > Cognos 8 Architecture Overview > Cognos 8 Components > Load Balancing > Deployment
bbc Adobe LiveCycle Data Services Using the F5 BIG-IP LTM Introduction APPLIES TO CONTENTS
TECHNICAL ARTICLE Adobe LiveCycle Data Services Using the F5 BIG-IP LTM Introduction APPLIES TO Adobe LiveCycle Enterprise Suite CONTENTS Introduction................................. 1 Edge server architecture......................
AN EFFICIENT LOAD BALANCING APPROACH IN CLOUD SERVER USING ANT COLONY OPTIMIZATION
AN EFFICIENT LOAD BALANCING APPROACH IN CLOUD SERVER USING ANT COLONY OPTIMIZATION Shanmuga Priya.J 1, Sridevi.A 2 1 PG Scholar, Department of Information Technology, J.J College of Engineering and Technology
Load balancing in SOAJA (Service Oriented Java Adaptive Applications)
Load balancing in SOAJA (Service Oriented Java Adaptive Applications) Richard Olejnik Université des Sciences et Technologies de Lille Laboratoire d Informatique Fondamentale de Lille (LIFL UMR CNRS 8022)
CiteSeer x in the Cloud
Published in the 2nd USENIX Workshop on Hot Topics in Cloud Computing 2010 CiteSeer x in the Cloud Pradeep B. Teregowda Pennsylvania State University C. Lee Giles Pennsylvania State University Bhuvan Urgaonkar
Performance And Scalability In Oracle9i And SQL Server 2000
Performance And Scalability In Oracle9i And SQL Server 2000 Presented By : Phathisile Sibanda Supervisor : John Ebden 1 Presentation Overview Project Objectives Motivation -Why performance & Scalability
Proposal of Dynamic Load Balancing Algorithm in Grid System
www.ijcsi.org 186 Proposal of Dynamic Load Balancing Algorithm in Grid System Sherihan Abu Elenin Faculty of Computers and Information Mansoura University, Egypt Abstract This paper proposed dynamic load
SOLUTION BRIEF: SLCM R12.8 PERFORMANCE TEST RESULTS JANUARY, 2013. Submit and Approval Phase Results
SOLUTION BRIEF: SLCM R12.8 PERFORMANCE TEST RESULTS JANUARY, 2013 Submit and Approval Phase Results Table of Contents Executive Summary 3 Test Environment 4 Server Topology 4 CA Service Catalog Settings
Business Application Services Testing
Business Application Services Testing Curriculum Structure Course name Duration(days) Express 2 Testing Concept and methodologies 3 Introduction to Performance Testing 3 Web Testing 2 QTP 5 SQL 5 Load
Application Performance Management for Enterprise Applications
Application Performance Management for Enterprise Applications White Paper from ManageEngine Web: Email: [email protected] Table of Contents 1. Introduction 2. Types of applications used
Frequently Asked Questions
Frequently Asked Questions 1. Q: What is the Network Data Tunnel? A: Network Data Tunnel (NDT) is a software-based solution that accelerates data transfer in point-to-point or point-to-multipoint network
Building Scalable Applications Using Microsoft Technologies
Building Scalable Applications Using Microsoft Technologies Padma Krishnan Senior Manager Introduction CIOs lay great emphasis on application scalability and performance and rightly so. As business grows,
Web Application Hosting Cloud Architecture
Web Application Hosting Cloud Architecture Executive Overview This paper describes vendor neutral best practices for hosting web applications using cloud computing. The architectural elements described
PERFORMANCE COMPARISON OF COMMON OBJECT REQUEST BROKER ARCHITECTURE(CORBA) VS JAVA MESSAGING SERVICE(JMS) BY TEAM SCALABLE
PERFORMANCE COMPARISON OF COMMON OBJECT REQUEST BROKER ARCHITECTURE(CORBA) VS JAVA MESSAGING SERVICE(JMS) BY TEAM SCALABLE TIGRAN HAKOBYAN SUJAL PATEL VANDANA MURALI INTRODUCTION Common Object Request
Managing your Red Hat Enterprise Linux guests with RHN Satellite
Managing your Red Hat Enterprise Linux guests with RHN Satellite Matthew Davis, Level 1 Production Support Manager, Red Hat Brad Hinson, Sr. Support Engineer Lead System z, Red Hat Mark Spencer, Sr. Solutions
XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12
XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines A.Zydroń 18 April 2009 Page 1 of 12 1. Introduction...3 2. XTM Database...4 3. JVM and Tomcat considerations...5 4. XTM Engine...5
Cloud Performance Benchmark Series
Cloud Performance Benchmark Series Amazon Elastic Load Balancing (ELB) Md. Borhan Uddin Bo He Radu Sion ver. 0.5b 1. Overview Experiments were performed to benchmark the Amazon Elastic Load Balancing (ELB)
Clustering with Tomcat. Introduction. O'Reilly Network: Clustering with Tomcat. by Shyam Kumar Doddavula 07/17/2002
Page 1 of 9 Published on The O'Reilly Network (http://www.oreillynet.com/) http://www.oreillynet.com/pub/a/onjava/2002/07/17/tomcluster.html See this if you're having trouble printing code examples Clustering
Performance Testing. Slow data transfer rate may be inherent in hardware but can also result from software-related problems, such as:
Performance Testing Definition: Performance Testing Performance testing is the process of determining the speed or effectiveness of a computer, network, software program or device. This process can involve
Test Run Analysis Interpretation (AI) Made Easy with OpenLoad
Test Run Analysis Interpretation (AI) Made Easy with OpenLoad OpenDemand Systems, Inc. Abstract / Executive Summary As Web applications and services become more complex, it becomes increasingly difficult
Efficient Parallel Processing on Public Cloud Servers Using Load Balancing
Efficient Parallel Processing on Public Cloud Servers Using Load Balancing Valluripalli Srinath 1, Sudheer Shetty 2 1 M.Tech IV Sem CSE, Sahyadri College of Engineering & Management, Mangalore. 2 Asso.
SIDN Server Measurements
SIDN Server Measurements Yuri Schaeffer 1, NLnet Labs NLnet Labs document 2010-003 July 19, 2010 1 Introduction For future capacity planning SIDN would like to have an insight on the required resources
Exploring Oracle E-Business Suite Load Balancing Options. Venkat Perumal IT Convergence
Exploring Oracle E-Business Suite Load Balancing Options Venkat Perumal IT Convergence Objectives Overview of 11i load balancing techniques Load balancing architecture Scenarios to implement Load Balancing
Architecting For Failure Why Cloud Architecture is Different! Michael Stiefel www.reliablesoftware.com development@reliablesoftware.
Architecting For Failure Why Cloud Architecture is Different! Michael Stiefel www.reliablesoftware.com [email protected] Outsource Infrastructure? Traditional Web Application Web Site Virtual
CLOUD PERFORMANCE TESTING - KEY CONSIDERATIONS (COMPLETE ANALYSIS USING RETAIL APPLICATION TEST DATA)
CLOUD PERFORMANCE TESTING - KEY CONSIDERATIONS (COMPLETE ANALYSIS USING RETAIL APPLICATION TEST DATA) Abhijeet Padwal Performance engineering group Persistent Systems, Pune email: [email protected]
Tuning Tableau Server for High Performance
Tuning Tableau Server for High Performance I wanna go fast PRESENT ED BY Francois Ajenstat Alan Doerhoefer Daniel Meyer Agenda What are the things that can impact performance? Tips and tricks to improve
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 4, July-Aug 2014
RESEARCH ARTICLE An Efficient Service Broker Policy for Cloud Computing Environment Kunal Kishor 1, Vivek Thapar 2 Research Scholar 1, Assistant Professor 2 Department of Computer Science and Engineering,
Performance brief for IBM WebSphere Application Server 7.0 with VMware ESX 4.0 on HP ProLiant DL380 G6 server
Performance brief for IBM WebSphere Application Server.0 with VMware ESX.0 on HP ProLiant DL0 G server Table of contents Executive summary... WebSphere test configuration... Server information... WebSphere
Tableau Server 7.0 scalability
Tableau Server 7.0 scalability February 2012 p2 Executive summary In January 2012, we performed scalability tests on Tableau Server to help our customers plan for large deployments. We tested three different
Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago
Globus Striped GridFTP Framework and Server Raj Kettimuthu, ANL and U. Chicago Outline Introduction Features Motivation Architecture Globus XIO Experimental Results 3 August 2005 The Ohio State University
A Case Based Tool for Monitoring of Web Services Behaviors
COPYRIGHT 2010 JCIT, ISSN 2078-5828 (PRINT), ISSN 2218-5224 (ONLINE), VOLUME 01, ISSUE 01, MANUSCRIPT CODE: 100714 A Case Based Tool for Monitoring of Web Services Behaviors Sazedul Alam Abstract Monitoring
DYNAMIC LOAD BALANCING IN CLIENT SERVER ARCHITECTURE
DYNAMIC LOAD BALANCING IN CLIENT SERVER ARCHITECTURE PROJECT OF COEN233 SUBMITTED BY Aparna R Lalita V Sanjeev C 12/10/2013 INSTRUCTOR Dr. Prof Ming-Hwa Wang Santa Clara University 1 TABLE OF CONTENTS
Enterprise Service Bus
We tested: Talend ESB 5.2.1 Enterprise Service Bus Dr. Götz Güttich Talend Enterprise Service Bus 5.2.1 is an open source, modular solution that allows enterprises to integrate existing or new applications
Top 10 Tips for z/os Network Performance Monitoring with OMEGAMON Session 11899
Top 10 Tips for z/os Network Performance Monitoring with OMEGAMON Session 11899 Dean Butler [email protected] 2012 IBM Corporation Agenda IBM Software Group Tivoli software Best Practices in Monitoring
Informatica Master Data Management Multi Domain Hub API: Performance and Scalability Diagnostics Checklist
Informatica Master Data Management Multi Domain Hub API: Performance and Scalability Diagnostics Checklist 2012 Informatica Corporation. No part of this document may be reproduced or transmitted in any
WEBAPP PATTERN FOR APACHE TOMCAT - USER GUIDE
WEBAPP PATTERN FOR APACHE TOMCAT - USER GUIDE Contents 1. Pattern Overview... 3 Features 3 Getting started with the Web Application Pattern... 3 Accepting the Web Application Pattern license agreement...
Achieving Zero Downtime and Accelerating Performance for WordPress
Application Note Achieving Zero Downtime and Accelerating Performance for WordPress Executive Summary WordPress is the world s most popular open source website content management system (CMS). As usage
Group Based Load Balancing Algorithm in Cloud Computing Virtualization
Group Based Load Balancing Algorithm in Cloud Computing Virtualization Rishi Bhardwaj, 2 Sangeeta Mittal, Student, 2 Assistant Professor, Department of Computer Science, Jaypee Institute of Information
WebSphere Application Server V6.1 Extended Deployment: Overview and Architecture
Chapter 32 WebSphere Application Server V6.1 Extended Deployment: Overview and Architecture The WebSphere Application Server Extended Deployment (WAS XD) package provides many extensions to existing functionality
http://support.oracle.com/
Oracle Primavera Contract Management 14.0 Sizing Guide October 2012 Legal Notices Oracle Primavera Oracle Primavera Contract Management 14.0 Sizing Guide Copyright 1997, 2012, Oracle and/or its affiliates.
