Lightweight Resource Scaling for Cloud Applications
|
|
|
- Muriel Gibson
- 10 years ago
- Views:
Transcription
1 Lightweight Resource Scaling for Cloud Applications Rui Han, Li Guo, Moustafa M. Ghanem, Yike Guo* Department of Computing Imperial College London London, UK {r.han10, liguo, mmg, Abstract Elastic resource provisioning is a key feature of cloud computing, allowing users to scale up or down resource allocation for their applications at run-time. To date, most practical approaches to managing elasticity are based on allocation/de-allocation of the virtual machine (VM) instances to the application. This VM-level elasticity typically incurs both considerable overhead and extra costs, especially for applications with rapidly fluctuating demands. In this paper, we propose a lightweight approach to enable cost-effective elasticity for cloud applications. Our approach operates finegrained scaling at the resource level itself (CPUs, memory, I/O, etc) in addition to VM-level scaling. We also present the design and implementation of an intelligent platform for light-weight resource management of cloud applications. We describe our algorithms for light-weight scaling and VM-level scaling and show their interaction. We then use an industry standard benchmark to evaluate the effectiveness of our approach and compare its performance against traditional approaches. Keywords- cloud computing; resource allocation algorithms; lightweight scaling I. INTRODUCTION Cloud computing has gained unquestionable commercial success in recent years. Key value propositions promoted by cloud IaaS (Infrastructure-as-a-Service) providers, namely infrastructure owners or resellers such as Amazon AWS [1] and GoGrid [2], include the user s ability to scale up or down resources used based on their computational demand. The approach is typically based on charging the application owners for using one or more Virtual Machines (VMs) per unit time. These VMs are then hosted for execution on different Physical Machines (PMs) at the infrastructure provider s cloud. In addition, many applications hosted in a cloud environment may have varying resource demands as they are expected to serve wide range of workloads, some of which could be predictable while others fluctuate based on using the application. For example, an e-commerce website may have a workload fluctuation during the release of some promotion activities or other events. Within this context, ondemand scaling (also known as dynamic or online resource provisioning) of applications becomes one of the most important features of cloud services. This feature enables real-time acquisition/release of computing resources to adjust the applications performance in order to meet QoS (Quality of Service) requirements, while also minimising the infrastructure provider s operational cost to remain competitive. * Please direct your enquiries to the communication author Professor Yike Guo. Most approaches for on-demand resource scaling, whether used in practice or described in the literature [1, 3, 4, 5, 6, 7], are typically based on controlling (increasing or decreasing) the number of VM instances that host the applications server components. The approach is appealing for a wide class of applications especially those that are based on multitier architectures, or server-side software platforms. For such applications, scaling up an application typically involves adding an extra software server, and hence an extra (server) VM in the cloud context. Using VMs as basic units for dynamic resource provisioning for cloud applications incurs considerable overhead and costs. Allocation of new VMs incurs considerable waste of computing resources as each VM consumes resources which are not directly used by the applications. As the number of allocated VMs increases, the total amount of resources used for housing keeping also increases. In addition, creating/shutting down/removing VMs dynamically at run time incurs extra overheads. Based on our experience, however, we noticed that in many real-world scenarios, scaling an application s resource up and down does not always require its underlying computing infrastructure to be changed much by adding/removing VMs constantly. More subtle changes, such as modifying VMs capacity, can be conducted to achieve the desired effects with less running cost and, usually, in less time. In this paper, we propose a lightweight approach for achieving more cost-efficient scaling of cloud resources at the IaaS cloud provider s side. This approach offers benefits for multi-tier applications that are already implemented using multiple VMs by improving the resource utilisation between them as application demands vary. In particular, we present: Fine-grained scaling approach: a novel lightweight scaling (LS) algorithm is introduced to enable fine-grained scaling of an application at the level of underlying resources, namely CPU, Memory and I/O. Improving resource utilisation: the LS algorithm increases the resources utilisation of PMs by using idle resources to release overloaded resources. Implementation and experimental evaluation: an intelligent platform based on the LS algorithm is implemented to automate the scaling process of cloud applications. The proposed lightweight approach is tested using an industry standard benchmark [8] on the IC-Cloud infrastructure [9] and the test results show its effectiveness in scaling applications to meet their QoS requirements while reducing infrastructure providers cost. The remainder of the paper is organized as follows: Section II provides the motivation and background to our
2 approach; Section III presents an overview of the related work on scaling techniques used for cloud computing; Section IV presents our LS algorithm; Section V introduces the architecture of the Imperial Smart Scaling engine (isse) implemented to support our approach; Section VI presents an experimental evaluation of our work, and compares its effectiveness against traditional approaches; and Section VII presents a summary and discussion of the framework and describes avenues for future work. II. MOTIVATION This section illustrates the idea of lightweight scaling for cost-effective resource management in a cloud environment. The approach is generic, and we present it using the example of an e-commerce application, which abstracts the typical behaviour of a wide class of applications running on cloud services. Figure 1(a) shows the basic infrastructure of the application that is composed of five tiers of components (servers): the HAProxy and Amoeba load balancing servers, the Apache HTTP server, the Tomcat application server and the MySQL database server. These servers work together to handle end-users requests. overall utilisation of resources also varies across different PMs used by the application (Figure 2(c)). The aim of lightweight scaling is to operate at resource level to address such imbalances and improve resource utilisation rather than adding new resources to scale the application up. Conceptually, lightweight scaling can be based on two techniques. The first, Self-healing scaling is applicable in cases when two servers of an application are deployed on one PM, and when the idle resources of one VM can be used to release the overloaded resources in the other. For example, one or more CPUs allocated to a VM with low CPU utilisation can be removed and to another VM which has already saturated existing CPU resources. The second technique, Resource-level scaling, is based on using unallocated resources available at a particular PM to scale up a VM executing on it. For example, a PM with low CPU and memory utilisation can allocate these types of resources to one of the VMs executing on it thus scaling it up. Figure 1. The example of an e-commerce website. Once deployed, the servers of this application are hosted across different VMs. An interesting point here is that the amount of resources that are consumed by these servers depends on the behaviour of the application itself. For example, if webpage browsing and product previewing are the application users main actions over a period of time, Apache and Tomcat web servers will be stressed and their resources become saturated. By contrast, the MySQL database component would remain performing well. However, from the application owner s point of view (ecommerce website), the overall performance of the application is affected. Ideally, they may choose to allocate more resources to the bottleneck component of the application, namely the webs servers, while keeping the MySQL database component in idle status. We call such cases imbalanced resource provisioning for an application. Typically, for an application in the cloud environment, different types of imbalanced resource provisioning could exist. To illustrate, consider Figure 2(a), where a single server VM executes on its own PM. In this case, any of the three types of resources (CPU, memory and I/O), could suffer from high or low utilisation based on some metric. For example, one can consider resources with high utilisation (e.g., a memory s utilisation is 0.8 means 80% of the memory is used) to be typically overloaded, and those with low utilisation (e.g., 0.3) to be mostly idle. Such imbalance becomes significant for applications using more than one VM executing on the same PM. Such VMs typically can compete for the same resources (Figure 2(b)). Note that the Figure 2. Resource shown as cubes and three types of unbalanced resource provisioning. Our motivation in this paper is to develop a system and algorithms based on a light-weight scaling approach to support this conceptual framework. The system would be installed at an IaaS cloud provider s side enabling them to support QoS requirements of the application owner without incurring unnecessary costs for either the provider or application owner. The approach is generic and can be applied to any cloud provider for a wide range of applications. However, our focus in this paper is on transaction-based multitier applications where the performance of the application, and hence QoS, can be assessed based on the application s response time to each incoming request. III. RELATED WORK A. Scaling Techniques before Clouds Application scaling has long been studied before clouds. Early work considers single-tier applications and focused on transforming performance targets into underlying resources such as CPU numbers and allocated bandwidth [10]. Further investigations classify an application into multiple tiers and provision each tier with enough resources to meet the application s peak workload [11, 12]. Although scaling of traditional applications on physical servers shares many commonalities with that of cloud applications, scaling in these two environments has different emphases. Conventional techniques mainly focused on how to schedule compute nodes to meet QoS requirement of
3 applications by predicting their long-term demand changes. In contrast, clouds focus on providing metered resources ondemand and on quickly scaling applications up and down whenever the user demand changes. Further investigations, therefore, are needed to address challenges brought by this requirement for high elasticity. B. Traditional Scaling Techniques in Clouds Most IaaS cloud providers (e.g., Amazon AWS [1]) and vendors (e.g., RightScale [3]) assist application owners to manage cloud infrastructure and employ pre-defined policies (rules) to guide application scaling. For example, vendors like RightScale require application owners to manually specify scaling rules after an application is deployed. These rules specify the upper and lower bounds of the number of servers, the conditions to trigger scaling and the number of changed servers in each scaling. The policy-based mechanism assumes that the application owners have particular knowledge of the application being executed so that they can define proper policies. However, this is not always the case. In addition, a variety of scaling approaches are proposed based on analytical models of the application. In [4], Bacigalupo et al. model an application by a three-tier queueing model, namely application, database and storage disk tiers. Each tier is solved to analyse the servers mean response time and throughput. A scaling algorithm is then proposed using these analysis results. Similar to [11, 12], researchers in [5] break down an application s end-to-end response time by tier. They then calculate the number of servers to be allocated to each tier in order to meet the response time target. In addition, a method is presented to support the scaling up of a two-tier web application by actively profiling the VMs CPU usage. The method also realises scaling down using a regression-model-based predictive mechanism [6]. In [7], a network flow model is introduced to analyse applications to assist application owners in making a trade-off between cost and QoS. In [13], Gong et al. propose a scaling approach that conducts accurate resource allocation using two complementary techniques: patterns-driven (they call repeating patterns as signatures) and state-driven (they use discrete-time Markov chain) demand predictions. In [14], Shen et al. further improve the prediction mechanism by introducing two handling schemas, namely online adaptive padding and reactive error correction. C. Lightweight Cloud Resource Management At present, some preliminary research has conducted on lightweight management of cloud resources. In [15], He et al. proposed a novel concept called Elastic Application Container (EAC) which is a lightweight resource management approach. It is considered as a substitution of VM virtual infrastructure for hosting cloud applications. Their experiments demonstrate that EAC is more resourceefficient and feasible than VM infrastructure in terms of application migrations and resource utilisation. In addition, a lightweight mechanism is presented to reduce VM migration time by pre-caching memory pages [16]. Researchers in [17] focus on live migration of shared storage databases. They introduce an end-to-end technique to ensure the minimal impact on database performance during migration. IV. THE LIGHTWEIGHT SCALING ALGORITHM In this section, we provide an overview of the LS algorithm, and the associated algorithms for scaling up and down. A. The LS Algorithm Overview In order to help readers understand the algorithm, Table 1 lists the main parameters used throughout the LS algorithm. Their values are obtained in a variety of ways. TABLE I. PARAMETERS OF THE LS ALGORITHM Parameter Description Source Three resource dimensions of a VM: {"#, "#$%, /}. A server hosted on a VM. PM() Server s PM. () Server s running cost. (, ) One unit of resource s running cost. " () " () (, ) The threshold of utilisation for scaling up resource. The threshold of utilisation for scaling down resource. Lower bound and upper bound of the required response time. The observed end-to-end response time (, ) Resource s utilisation in server. Cloud providers Application owners Online measurement First of all, the IaaS cloud provider needs to specify its running cost of a single server (for convenience, each server is installed on one dedicated VM) or the server s underlying resources, namely CPU, memory and I/O. This cost represents resource consumption and operating cost of cloud providers. Such a cost varies for different cloud providers and the cost function is a research topic in itself [18]. Note that resources are typically charged based on their basic unit. For example, the basic unit of memory is 1GB and it is charged 2 cents/hour. In addition, cloud providers also give two thresholds of utilisation for triggering scaling up (down) of resource. For example, an application may define the CPU s threshold for scaling up (i.e., " ("#)) as 0.8. This threshold represents high CPU usage and indicates that an additional CPU should be added. Similarly, the memory s threshold for scaling down (i.e., " ("#$% )) is 0.3, which denotes the low memory usage and so some redundant memory can be removed. In addition, application owners specify the QoS demand as the required response time in a service level agreement (SLA). Without losing generality, this response time includes a lower bound and an upper bound. These two bounds define a range of response times such as 1.8 to 2.0 seconds that are acceptable to service users. Finally, the observed
4 Accepted, to appear in CCGrid 2012 end-to-end response time of the application and each VM s resource utilisation is collected online by monitors. Given an -tier application to be scaled with the LS algorithm, this application has server components, i.e., its sever set consists of servers where and = 1,,. Pseudo code for the LS algorithm is given below. The algorithm is started after the application is initially deployed and keeps running until the application is completed (lines 2 to 7). The algorithm triggers a lightweight scaling up (LSU) whenever the observed response time is larger than the upper bound of the required response time (lines 4 and 5), or a lightweight scaling down (LSD) if the observed time is smaller than the lower bound (lines 6 and 7). Algorithm 1: LS Input:, (, ) 1. Begin 2. while (the application is not completed) Monitor once every few minutes; if >, then LSU (, (, ), ); else if <, then 7. LSD (, (, ), ). 8. End The lightweight scaling here has a twofold meaning. First, the algorithm conducts fine-grained scaling by modifying each VM s resource configuration. This ensures that the algorithm is light weight and consumes insignificant computing resources for resource adjustment. The scaling can be completed in as quick as few milliseconds, which guarantees the period of violating response time acceptable by end users. Second, the algorithm applies an automatic reactive scaling mechanism similar to mechanisms applied by Amazon EC2 and Rightscale. This measurement-based scaling needs no prior knowledge about the applications. Each time an application is scaled, its load balancing servers automatically redistribute incoming requests and the service monitor can instantaneously detect the updated end-to-end response time. The scaling process is completed when the detected response time meets the requirement (i.e., it falls between and ). This ensures the just enough resource allocation to the application and minimises cloud providers cost. B. The LSU Algorithm The LSU algorithm (Algorithm 2) aims to reduce an application s response time below the upper bound, while minimising the increase of resources or number of VM instances. Algorithm 2: LSU Input:, (, ),. Output: updated. 1. Begin 2. Conduct the self-healing scaling; if >, then Conduct the resource-level scaling up; while ( > ) Conduct the VM-level scaling up; 7. Conduct the resource-level scaling up. 8. End LSU algorithm conducts three types of scaling with different levels of priorities. The self-healing and resourcelevel scalings first reduce the application s response time by increasing VMs capacity using the available resources from these servers hosting PMs (lines 2 to 4). Note that these two scalings have some constraints. For example, a VM instance with a 32-bit operating system can only be allocated 4G memory in maximum. The LSU can be accomplished if any of these two scalings meet the response time target. Otherwise, the VM-level scaling is required to add a VM instance in a new PM to the application (line 6). This new PM may have available resources and so the resource-level scaling can be executed again (line 7). The basic idea of self-healing scaling is that if two VMs of the application s servers are hosted in the same PM, these VMs resource may complement each other. This scaling uses a candidate list (line 6 in algorithm 3) to store all pairs of eligible servers for VM modification. A pair of eligible server (, ) must satisfy: and belong to the same application and PM, s VM has one resource (e.g., CPU) whose utilisation is larger than its scaling up threshold (e.g., 0.8), and resource s utilisation in s VM is smaller than its scaling down threshold (e.g., 0.3). One unit of resource is then removed from s idle VM (line 7) and it is added to s overloaded VM (line 8). The self-healing scaling iteratively executes until the required response time is met (line 2) or no eligible server can be found (i.e., is empty). Algorithm 3: Self-healing scaling 1. Measure and (, ); // 2. While ( > and ) ={}; For each pair (, ) in if ((, ) > " () and PM( )==PM( ) and (, )< " ()), then = {(, )}; 7. Remove one unit of resource from ; 8. Add one unit of resource to ; 9. Measure and (, ). Resource-level scaling up follows the observation that if the PMs that host the application s server components have available resources, these resources can be used for scaling up while other applications hosted in these PMs are not affected. One possible way to ensure available resources is to reserve some resources in advance. Note that a PM can theoretically generate an unlimited number of CPUs. Hence the availability of CPU resource could be judged according to the PM s CPU utilisation (e.g., the CPU resource is available if the CPU utilisation is lower than 0.8). In the resource-level scaling up (see algorithm 4), an eligible server s VM must have a saturated resource (i.e., its utilisation is higher than its scaling up threshold) and this VM s hosted PM must have one unit of available resource (line 5). This scaling selects all eligible servers into a candidate list (line 3). If multiple servers exist in, the criterion (, ) is used to select a server and scale up its resource (lines 7 and 8).
5 Accepted, to appear in CCGrid 2012 As listed in definition 1, the criterion (, ) consists of two variables: 1) resource s exploitable ratio (i.e., 1 (, )), which represents the remaining amount of in server s VM; 2) s running cost (, ). Using this criterion, the resource with the least remaining amount and cost is selected in the resource-level scaling up. Algorithm 4: Resource-level scaling up 1. Measure and (, ); // 2. While ( > and ) ={}; for (=1; < ; ++) if ((, )> " () and PM( ).hasavailableresource==true), then = { }; 7. Select from with the smallest (, ); 8. Add one unit of resource to ; 9. Measure and (, ). Calculate tier s ( ); //= 1,..., Select tier with the smallest ( ); if (Add a server to tier is feasible), then = { }. // add to Note that the addition of a server must satisfy some constraints (line 5). Examples are the application owner s budget and each tier s maximum number of servers. Servers own constraints should also be considered. For instance, MySQL Master s replication constraint is 1, which means an application can have one MySQL Master server at most. C. The LSD Algorithm The LSD algorithm (algorithm 6) aims to remove as many VMs and resources from an application as possible, while still trying to hold its response time target. The algorithm first performs the VM-level scaling down to identify a bottleneck tier with the largest ( ) (definition 2) to remove a server. This removal is expected to save the maximum cost per unit of increased response time. The VMlevel scaling down keeps running until it is infeasible (line Definition 1 (Resource-level scaling criterion (, )). 2): removing a VM would violate response time target. This The criterion used to select servers in resource-level scaling, feasibility can be checked using application profiling and denoted by (, ), is the product of a parameter workload predication techniques. The LSD algorithm then (positive constant), server s exploitable ratio and conducts the resource-level scaling down. This scaling down resource s running cost: selects a resource with the largest (, ) (definition 1) (, )= (1-(, ))(, ). (1) and removes one unit of from s VM. The removal implies the resource with the largest available amount and the highest running cost is removed. VM-level scaling up complements the above two scalings and it has the lowest priority to be triggered. This Algorithm 6: LSD scaling uses the criterion ( ) to detect the bottleneck tier Input:, (, ),. from the application for adding a server. As shown in Output: updated. 1. Begin definition 2, criterion ( ) consists of two variables: 1) the 2. Conduct the VM-level scaling down; mean value of servers exploitable ratios at tier. A server Conduct the resource-level scaling down. End s exploitable ratio is the product of its three resources exploitable ratios and it comprehensively represents s V. A PLATFORM TO SUPPORT LIGTWEIGHT SCALING remaining capacity; 2) the running cost of a server at tier. This section introduces a platform, called isse, to Definition 2 (VM-level scaling criterion ( )). The support lightweight scaling of cloud applications and criterion used for detecting the bottleneck tier in VM-level automate this scaling process. As shown in Figure 3, isse scaling, denoted by ( ), is the product of a parameter acts as middleware to be installed on top of an existing IaaS (positive constant) and the mean value of each server s cloud provider s infrastructure. Using the User Portal, exploitable ratio multiplied by its cost at tier : application owners can specify the required response time for their applications and get notified when these applications ( "#,"#$%,/((,)))( ) ( )=. (2) are scaled up (down) using the Lightweight Scaling Service. Where and is tier s number of servers. The VM-level scaling up uses criterion ( ) to select the tier with the least remaining resources and the smallest cost to add a server (line 4 in algorithm 5). In other words, adding a server to this tier should reduce each unit of response time while incurring the smallest running cost for cloud providers. Algorithm 5: VM-level scaling up 1. Measure and (, ); // 2. for (=1; < ; ++) Figure The architecture of isse.
6 In isse, the Monitoring Service monitors each running application in two ways. First, an entry monitor examines the incoming requests over a finite interval (e.g., 60 seconds). The monitored information includes the requests arrival rate and the end-to-end response time. This information is used to infer whether a scaling up (or down) is needed. For instance, a scaling up is triggered if the observed response time exceeds the required time. Secondly, a monitor is installed in each server. This monitor examines the resources utilisation. When a scaling is triggered, the Lightweight Scaling Service applies the LS algorithm to control the whole scaling process and the Deployment Service scales the resources of the application upon IC Cloud [9]. In isse, each single server is described by a XML based specification using the Open Virtualization Format (OVF) standard [19]. This specification accommodates all aspects of a single server, including its VM resource configuration, software user settings (e.g., Tomcat s user name and password), links to other servers and other constraints. isse can automatically executes the scaling process, namely addition, removal and resource modification of servers, by interpreting servers specifications and calling interfaces of IC Cloud [9]. The automation process and isse are detailed in [20]. VI. EXPERIMENTAL EVALUATION The experimental evaluation is designed to illustrate the effectiveness of our approach in enabling fine-grained scaling up and down of applications (section VI.B). Furthermore, the algorithm s salient feature in supporting cost-effective scaling for cloud providers is demonstrated by comparing with existing techniques (section VI.C). A. Experimental setup and Benchmark In the experiment, the isse was implemented as a full working system and was tested in one data center running the IC-Cloud platform [9]. This data center has four PMs that share the 1 Tb centralised storage and are connected through a switched gigabit Ethernet LAN. Each PM has 8 Quad-Core AMD processors and 32GB memory. Figure 1 s e-commerce website and its five types of servers were implemented. For convenience, each server is installed on one dedicated VM running Linux Centos Note that two versions of the MySQL database, namely MySQL Master and Slave, are implemented to support the Master-Slave data replication model. Two workloads of the application were tested: 1) the browsing workload represents the browsing behaviour of customers such as searching and ranking products as well as online enquiry; 2) the ordering workload denotes the ordering actions such as logging in and making orders, which make heavy database queries. We developed a client emulator to generate a sequence of session-based requests based on the TPC benchmark [8]. After setting the test period, this emulator continuously generates a sequence of interactions initiated by multiple active sessions. After an interaction is completed, the emulator waits for a random interval before initiating the next interaction to simulate customers thinking time. As shown in Figure 4, the varying number of active sessions represents the fluctuating demand. For each type of workload, a trial is performed and it lasts 3600 seconds. Each trial can be divided into six stages, in which three changes in demand pattern are simulated: when stages 2, 4 and 6 start, the number of active sessions increases to 300, 400 and 600, respectively. Each increased period of demand lasts 600 seconds and then the level falls down to 200 sessions. In the trial, the lower and upper bound of the required average response time is 1.8 and 2.0 seconds; the threshold of utilisation for scaling up is 0.8 and for scaling down is 0.3; other parameters of the application are monitored once every 60 seconds. Initially (stage 1), six servers with basic VM configurations are deployed on a PM and newly added servers are deployed on another PM. Cloud providers costs (cost unit: cents/hour) for running these servers are: 35 for HAProxy and Amoeba, 60 for MySQL Master and 16 for the other three servers. In lightweight scaling, these costs for the three types of resource are: 3 for 10Mb/second bandwidth (i.e., I/O resource), 3 for 1 GB memory and 2 for one CPU. Figure The number of active sessions during six stages of the experiment. B. Effectiveness of the LS Algorithm Figure 5 demonstrates the fluctuation of the observed end-to-end response time for two different workloads using the LS algorithm. Each time the demand (i.e., the number of sessions) changes, the response time violates the required upper or lower bound. This violation is detected when the LS algorithm monitors the application every 60 seconds and a scaling up (down) is triggered. In most of the cases, this scaling can be completed in a few milliseconds. One exception is at time=3000 seconds, new VMs are added and this incurs the scaling up time of 1 or 2 minutes. Figure The observed end-to-end response time. Figure 6 displays the variation of the assigned resources in the trial of the browsing and ordering workloads. The browsing workload mainly consumes memory and CPU resources, so the LS algorithm mainly modifies the configurations of these resources to handle the changing demand (Figure 6(a) and (c)). Stages 2, 4 and 6 show the
7 results of the self-healing, resource-level and VM-level scaling up operations. These three scalings correspond to the small, middle and large increases in demand, respectively. In particular, at stage 6, one Tomcat and one Apache servers are added. By contrast, the ordering workload primarily uses the I/O resources and the bandwidth configuration of VMs is modified (Figure 6(b)). Results: Our lightweight scaling approach can adapt to different fluctuating demands and perform the fine granularity scaling up (down) of applications to meet response time targets. Figure Total resource configuration in the trial of the two workloads. C. A Comparison with Existing Scaling Techniques We repeat the previous section s trial to compare the LS algorithm with traditional VM-level scaling algorithms. Typically, existing scaling techniques can be divided into two groups. First, the Policy-based scaling (PBS) algorithm scales applications using pre-defined rules [1, 3]. These rules are triggered to execute scaling according to monitoring results. In the experiment, if the resource utilisation of one type of server (e.g., Apache) is larger than 0.8 for 1 minute, additional servers are added. If this utilisation is less than 0.3 for 1 minute, redundant servers are removed. Secondly, in the Tier-Dividing Scaling (TDS) algorithm, the upper bound of the required response time (2.0 seconds in the trial) is broken down into three per-tier response times, which are 40%, 20%, and 40% for the tiers of Apache, Tomcat, MySQL Master (Slave), respectively. The TDS algorithm then calculates the suitable number of servers for each tier using analytical models [5, 6, 11, 12]. Figure 7 shows the scaling results of three algorithms in the trial of two workloads. The three algorithms are compared from three aspects. Figure 7. Comparing results of the three scaling algorithms. 1) Total number of servers: Figure 7a(b) display the application s total number of servers at each stage of the trial. The PBS algorithm increases the number of overloaded servers in scaling up and decreases the number of idle servers in scaling down. Depending on different types of workloads, different tiers of the application are scaled in the PBS algorithm. The TDS conducts worst-case scaling and allocates each tier the same number of servers for both types of workloads. By contrast, the LS algorithm only needs to increase the number of VMs in the fifth scaling and the application s number of servers is the smallest in this algorithm. This implies our algorithm can meet the same QoS requirement using the smallest amount of resources, thereby improving resource utilisation of the application s hosting PMs. 2) Scaling time: Five scaling up (down) times in each trial are listed in Figure 7c(d). Each scaling time represents the reaction time to an increase (decrease) of demand. For example, the scaling time of stage 1 to 2 denotes the reaction time when the number of sessions changes from 200 to 300. Each scaling lasts at least 1 minute because the monitor needs this amount of time to detect the demand change. In both workloads, our LS algorithm has the shortest scaling time. This is because our algorithm needs the least addition/removal operations of VMs. In cloud platforms, the duration of these operations (several minutes) is much longer than that of the resource modification operations (a few milliseconds). 3) IaaS Cloud providers cost: Figure 7e(f) compare cloud providers cost for running the application. Experimental results show that the LS algorithm incurs the smallest cost by deploying as few VMs as possible in the three scaling up operations (stages 2, 4 and 6). In the scaling down (stage 3 and 5), the PBS and TDS algorithms revert the application to its initial deployment. By contrast, the LS algorithm further reduces VMs capacity to save cost by removing their idle resources. Results: Our LS algorithm can perform agile scaling up and down of applications as well as save cloud providers cost by improving resource utilisation.
8 VII. CONCLUSION AND DISCUSSIONS In this paper, we argued that conventional VM-level scaling of cloud services may over-use resources and increase cloud providers operating costs to meet QoS requirements for user applications. Based on this argument, we proposed a lightweight approach that operates finegrained scaling at resource level and improves resource utilisation for cloud service operations. The approach can efficiently scale cloud application s resources up and down in order to meet the given QoS requirements while reducing cloud providers costs. An intelligent platform based on the lightweight scaling approach has been designed to achieve this cost-effective scaling. The practical effectiveness of the approach has been successfully tested using industry standard benchmarks. The overall approach presented in this paper integrates both fine-grained scaling at the resource level itself (CPUs, memory, I/O, etc.) in addition to VM-level scaling, and can be used effectively within a single application. There are multiple avenues for extending the work and we discuss two possible directions below. We have described self-healing scaling technique in the context of a single application. This technique coordinates resource usage within an application by using the idle resource of a server to release the overloaded resource in another server hosted in the same PM. The major advantage of this type of scaling is that it scales up the application without incurring any cost on either the cloud provider or the application owner. In this paper, we have considered this type of scaling when two servers belong to the same application. This single-application requirement could be relaxed and investigated in the context of multiple applications belonging to the same user. This is appealing for two reasons. First, coordinating resources within a user s applications does not affect other users applications. Second, scaling up and down of resources can be completed within milliseconds and the approach could enable real-time scheduling to allow multiple applications to meet their QoS requirements simultaneously. However, further investigations are needed to consider how resources can be scheduled between applications with different QoS requirements in this case. We note that, in our approach, when resource-level scaling is conducted, an application can only be scaled up when its hosting PMs have available resources. This may not be the case always. One possible way to provide available resources is to reserve some resources in these PMs that can only be used by resource-level scaling. Obviously, the higher proportion of the PMs resource is reserved, the larger chance that resource-level scaling can be performed and so high-cost VM-level scaling can be avoided. However, excessive resource reservation causes high cost for the IaaS cloud providers as well. Further investigations, therefore, are needed to understand the trade-offs between reservation cost and the risk of high running cost incurred by VM-level scaling. Another approach to provide available resources is to schedule the VMs among the PMs within a data center. For example, one VM in an over-loaded PM can be migrated to another light-loaded PM to release some resources. Existing VM migration/placement techniques, therefore, can be applied in lightweight scaling to help cloud providers achieve resource efficiency in the whole data center and solve conflicts between applications during scaling. Finally we also note that we have only considered transaction-based applications that execute at a single cloud provider. Considering more complex applications high performance scientific applications that can be scheduled, or coordinated, across different providers in a Grid computing fashion [21] and how to address their requirements would also be another interesting avenue for extending this research. REFERENCES [1] Amazon elastic compute cloud (EC2): ( ). [2] GoGrid: ( ). [3] RightScale: ( ). [4] D. A. Bacigalupo, J. van Hemert, et al, "Managing dynamic enterprise and urgent workloads on clouds using layered queuing and historical performance models," Simulation Modelling Practice and Theory, vol. 19, pp , [5] J. Bi, Z. Zhu, R. Tian, and Q. Wang, "Dynamic Provisioning Modeling for Virtualized Multi-tier Applications in Cloud Data Center," in CLOUD 10, Miami, Florida, 2010, pp [6] W. Iqbal, M. N. Dailey, D. Carrera, and P. Janecek, "Adaptive resource provisioning for read intensive multi-tier applications in the cloud," Future Generation Computer Systems, vol. 27, pp , [7] J. Z. Li, "Fast Optimization for Scalable Application Deployments in Large Service Centers," Doctor of Philosophy thesis, Department of Systems and Computer Engineerin, Carleton University, [8] TPC Transaction Processing Performance Council: ( ). [9] L. Guo, Y. Guo, and X. Tian, "IC Cloud: A Design Space for Composable Cloud Computing," in CLOUD'10, 2010, pp [10] R. P. Doyle, J. S. Chase, O. M. Asad, W. Jin, and A. M. Vahdat, "Model-based resource provisioning in a web service utility," in USITS'03, 2003, vol. 4, pp. 5- [11] B. Urgaonkar, P. Shenoy, A. Chandra, P. Goyal, and T. Wood, "Agile dynamic provisioning of multi-tier Internet applications," ACM Transactions on Autonomous and Adaptive Systems (TAAS), vol. 3, pp. 1-25, [12] B. Urgaonkar, P. Shenoy, A. Chandra, and P. Goyal, "Dynamic provisioning of multi-tier internet applications," in ICAC'05, Seattle, Washington, 2005, pp [13] Z. Gong, X. Gu, and J. Wilkes, "PRESS: PRedictive Elastic ReSource Scaling for cloud systems," in CNSM 10, Niagara Falls, Canada, 2010, pp. 9-1 [14] Z. Shen, S. Subbiah, X. Gu, and J. Wilkes, "CloudScale: elastic resource scaling for multi-tenant cloud systems," in SOCC'11, Cascais, Portugal, [15] S. He, L. Guo, Y. Guo, M. Ghanem, R. Han, and C. Wu, "Elastic Application Container: A Lightweight Approach for Cloud Resource Provisioning," in The 26th IEEE International Conference on Advanced Information Networking and Applications (AINA-2012), Fukuoka, Japan, 2012, in press. [16] T. Hirofuchi, H. Nakada, S. Itoh, and S. Sekiguchi, "Enabling instantaneous relocation of virtual machines with a lightweight VMM extension," in th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing (CCGrid'11), Melbourne, VIC, Australia, 2010, pp ,
9 [17] S. Das, S. Nishimura, D. Agrawal, and A. El Abbadi, "Albatross: lightweight elasticity in shared storage databases for the cloud using live data migration," Proceedings of the VLDB Endowment, vol. 4, pp , [18] M. Mihailescu and Y. M. Teo, "Dynamic Resource Pricing on Federated Clouds," in th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing (CCGrid'11), Melbourne, VIC, Australia, 2010, pp [19] VMware OVF: ( ). [20] R. Han, L. Guo, Y. Guo, and S. He, "A Deployment Platform for Dynamically Scaling Applications in The Cloud," in 3rd IEEE International Conference on Cloud Computing Technology and Science (CloudCom'11), Athens, Greece, 2011, pp [21] M. Ghanem, Y. Guo, A. Rowe, P. Wendel, "Grid-based knowledge discovery services for high throughput informatics," in Proceedings. 11th IEEE International Symposium on High Performance Distributed Computing, (HPDC 02), Edinburgh, UK, 2002, p. 41
Figure 1. The cloud scales: Amazon EC2 growth [2].
- Chung-Cheng Li and Kuochen Wang Department of Computer Science National Chiao Tung University Hsinchu, Taiwan 300 [email protected], [email protected] Abstract One of the most important issues
USING VIRTUAL MACHINE REPLICATION FOR DYNAMIC CONFIGURATION OF MULTI-TIER INTERNET SERVICES
USING VIRTUAL MACHINE REPLICATION FOR DYNAMIC CONFIGURATION OF MULTI-TIER INTERNET SERVICES Carlos Oliveira, Vinicius Petrucci, Orlando Loques Universidade Federal Fluminense Niterói, Brazil ABSTRACT In
Affinity Aware VM Colocation Mechanism for Cloud
Affinity Aware VM Colocation Mechanism for Cloud Nilesh Pachorkar 1* and Rajesh Ingle 2 Received: 24-December-2014; Revised: 12-January-2015; Accepted: 12-January-2015 2014 ACCENTS Abstract The most of
Auto-Scaling Model for Cloud Computing System
Auto-Scaling Model for Cloud Computing System Che-Lun Hung 1*, Yu-Chen Hu 2 and Kuan-Ching Li 3 1 Dept. of Computer Science & Communication Engineering, Providence University 2 Dept. of Computer Science
International Journal of Advance Research in Computer Science and Management Studies
Volume 2, Issue 11, November 2014 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online
Multi-dimensional Affinity Aware VM Placement Algorithm in Cloud Computing
Multi-dimensional Affinity Aware VM Placement Algorithm in Cloud Computing Nilesh Pachorkar 1, Rajesh Ingle 2 Abstract One of the challenging problems in cloud computing is the efficient placement of virtual
An Experimental Study of Load Balancing of OpenNebula Open-Source Cloud Computing Platform
An Experimental Study of Load Balancing of OpenNebula Open-Source Cloud Computing Platform A B M Moniruzzaman 1, Kawser Wazed Nafi 2, Prof. Syed Akhter Hossain 1 and Prof. M. M. A. Hashem 1 Department
A Dynamic Resource Management with Energy Saving Mechanism for Supporting Cloud Computing
A Dynamic Resource Management with Energy Saving Mechanism for Supporting Cloud Computing Liang-Teh Lee, Kang-Yuan Liu, Hui-Yang Huang and Chia-Ying Tseng Department of Computer Science and Engineering,
Workload Predicting-Based Automatic Scaling in Service Clouds
2013 IEEE Sixth International Conference on Cloud Computing Workload Predicting-Based Automatic Scaling in Service Clouds Jingqi Yang, Chuanchang Liu, Yanlei Shang, Zexiang Mao, Junliang Chen State Key
PERFORMANCE ANALYSIS OF PaaS CLOUD COMPUTING SYSTEM
PERFORMANCE ANALYSIS OF PaaS CLOUD COMPUTING SYSTEM Akmal Basha 1 Krishna Sagar 2 1 PG Student,Department of Computer Science and Engineering, Madanapalle Institute of Technology & Science, India. 2 Associate
Energy Constrained Resource Scheduling for Cloud Environment
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
Windows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...
International Journal of Computer & Organization Trends Volume21 Number1 June 2015 A Study on Load Balancing in Cloud Computing
A Study on Load Balancing in Cloud Computing * Parveen Kumar * Er.Mandeep Kaur Guru kashi University,Talwandi Sabo Guru kashi University,Talwandi Sabo Abstract: Load Balancing is a computer networking
RESOURCE MANAGEMENT IN CLOUD COMPUTING ENVIRONMENT
RESOURCE MANAGEMENT IN CLOUD COMPUTING ENVIRONMENT A.Chermaraj 1, Dr.P.Marikkannu 2 1 PG Scholar, 2 Assistant Professor, Department of IT, Anna University Regional Centre Coimbatore, Tamilnadu (India)
Efficient Scheduling Of On-line Services in Cloud Computing Based on Task Migration
Efficient Scheduling Of On-line Services in Cloud Computing Based on Task Migration 1 Harish H G, 2 Dr. R Girisha 1 PG Student, 2 Professor, Department of CSE, PESCE Mandya (An Autonomous Institution under
Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing
Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing Deep Mann ME (Software Engineering) Computer Science and Engineering Department Thapar University Patiala-147004
Elastic VM for Rapid and Optimum Virtualized
Elastic VM for Rapid and Optimum Virtualized Resources Allocation Wesam Dawoud PhD. Student Hasso Plattner Institute Potsdam, Germany 5th International DMTF Academic Alliance Workshop on Systems and Virtualization
Low Cost Quality Aware Multi-tier Application Hosting on the Amazon Cloud
Low Cost Quality Aware Multi-tier Application Hosting on the Amazon Cloud Waheed Iqbal, Matthew N. Dailey, David Carrera Punjab University College of Information Technology, University of the Punjab, Lahore,
An Autonomic Auto-scaling Controller for Cloud Based Applications
An Autonomic Auto-scaling Controller for Cloud Based Applications Jorge M. Londoño-Peláez Escuela de Ingenierías Universidad Pontificia Bolivariana Medellín, Colombia Carlos A. Florez-Samur Netsac S.A.
Infrastructure as a Service (IaaS)
Infrastructure as a Service (IaaS) (ENCS 691K Chapter 4) Roch Glitho, PhD Associate Professor and Canada Research Chair My URL - http://users.encs.concordia.ca/~glitho/ References 1. R. Moreno et al.,
Dynamic Load Balancing of Virtual Machines using QEMU-KVM
Dynamic Load Balancing of Virtual Machines using QEMU-KVM Akshay Chandak Krishnakant Jaju Technology, College of Engineering, Pune. Maharashtra, India. Akshay Kanfade Pushkar Lohiya Technology, College
Two-Level Cooperation in Autonomic Cloud Resource Management
Two-Level Cooperation in Autonomic Cloud Resource Management Giang Son Tran, Laurent Broto, and Daniel Hagimont ENSEEIHT University of Toulouse, Toulouse, France Email: {giang.tran, laurent.broto, daniel.hagimont}@enseeiht.fr
EC2 Performance Analysis for Resource Provisioning of Service-Oriented Applications
EC2 Performance Analysis for Resource Provisioning of Service-Oriented Applications Jiang Dejun 1,2 Guillaume Pierre 1 Chi-Hung Chi 2 1 VU University Amsterdam 2 Tsinghua University Beijing Abstract. Cloud
Payment minimization and Error-tolerant Resource Allocation for Cloud System Using equally spread current execution load
Payment minimization and Error-tolerant Resource Allocation for Cloud System Using equally spread current execution load Pooja.B. Jewargi Prof. Jyoti.Patil Department of computer science and engineering,
International Journal Of Engineering Research & Management Technology
International Journal Of Engineering Research & Management Technology March- 2014 Volume-1, Issue-2 PRIORITY BASED ENHANCED HTV DYNAMIC LOAD BALANCING ALGORITHM IN CLOUD COMPUTING Srishti Agarwal, Research
Sla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing
Sla Aware Load Balancing Using Join-Idle Queue for Virtual Machines in Cloud Computing Mehak Choudhary M.Tech Student [CSE], Dept. of CSE, SKIET, Kurukshetra University, Haryana, India ABSTRACT: Cloud
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT 1 SARIKA K B, 2 S SUBASREE 1 Department of Computer Science, Nehru College of Engineering and Research Centre, Thrissur, Kerala 2 Professor and Head,
Energetic Resource Allocation Framework Using Virtualization in Cloud
Energetic Resource Allocation Framework Using Virtualization in Ms.K.Guna *1, Ms.P.Saranya M.E *2 1 (II M.E(CSE)) Student Department of Computer Science and Engineering, 2 Assistant Professor Department
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
Run-time Resource Management in SOA Virtualized Environments. Danilo Ardagna, Raffaela Mirandola, Marco Trubian, Li Zhang
Run-time Resource Management in SOA Virtualized Environments Danilo Ardagna, Raffaela Mirandola, Marco Trubian, Li Zhang Amsterdam, August 25 2009 SOI Run-time Management 2 SOI=SOA + virtualization Goal:
IMPROVEMENT OF RESPONSE TIME OF LOAD BALANCING ALGORITHM IN CLOUD ENVIROMENT
IMPROVEMENT OF RESPONSE TIME OF LOAD BALANCING ALGORITHM IN CLOUD ENVIROMENT Muhammad Muhammad Bala 1, Miss Preety Kaushik 2, Mr Vivec Demri 3 1, 2, 3 Department of Engineering and Computer Science, Sharda
IaaS Cloud Architectures: Virtualized Data Centers to Federated Cloud Infrastructures
IaaS Cloud Architectures: Virtualized Data Centers to Federated Cloud Infrastructures Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Introduction
This is an author-deposited version published in : http://oatao.univ-toulouse.fr/ Eprints ID : 12902
Open Archive TOULOUSE Archive Ouverte (OATAO) OATAO is an open access repository that collects the work of Toulouse researchers and makes it freely available over the web where possible. This is an author-deposited
Keywords Distributed Computing, On Demand Resources, Cloud Computing, Virtualization, Server Consolidation, Load Balancing
Volume 5, Issue 1, January 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Survey on Load
An Approach to Load Balancing In Cloud Computing
An Approach to Load Balancing In Cloud Computing Radha Ramani Malladi Visiting Faculty, Martins Academy, Bangalore, India ABSTRACT: Cloud computing is a structured model that defines computing services,
Elasticity in Multitenant Databases Through Virtual Tenants
Elasticity in Multitenant Databases Through Virtual Tenants 1 Monika Jain, 2 Iti Sharma Career Point University, Kota, Rajasthan, India 1 [email protected], 2 [email protected] Abstract -
Benchmarking the Performance of XenDesktop Virtual DeskTop Infrastructure (VDI) Platform
Benchmarking the Performance of XenDesktop Virtual DeskTop Infrastructure (VDI) Platform Shie-Yuan Wang Department of Computer Science National Chiao Tung University, Taiwan Email: [email protected]
MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM?
MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? Ashutosh Shinde Performance Architect [email protected] Validating if the workload generated by the load generating tools is applied
Resource Allocation Avoiding SLA Violations in Cloud Framework for SaaS
Resource Allocation Avoiding SLA Violations in Cloud Framework for SaaS Shantanu Sasane Abhilash Bari Kaustubh Memane Aniket Pathak Prof. A. A.Deshmukh University of Pune University of Pune University
SLA-aware Resource Scheduling for Cloud Storage
SLA-aware Resource Scheduling for Cloud Storage Zhihao Yao Computer and Information Technology Purdue University West Lafayette, Indiana 47906 Email: [email protected] Ioannis Papapanagiotou Computer and
Optimizing the Cost for Resource Subscription Policy in IaaS Cloud
Optimizing the Cost for Resource Subscription Policy in IaaS Cloud Ms.M.Uthaya Banu #1, Mr.K.Saravanan *2 # Student, * Assistant Professor Department of Computer Science and Engineering Regional Centre
Windows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described
Permanent Link: http://espace.library.curtin.edu.au/r?func=dbin-jump-full&local_base=gen01-era02&object_id=154091
Citation: Alhamad, Mohammed and Dillon, Tharam S. and Wu, Chen and Chang, Elizabeth. 2010. Response time for cloud computing providers, in Kotsis, G. and Taniar, D. and Pardede, E. and Saleh, I. and Khalil,
GUEST OPERATING SYSTEM BASED PERFORMANCE COMPARISON OF VMWARE AND XEN HYPERVISOR
GUEST OPERATING SYSTEM BASED PERFORMANCE COMPARISON OF VMWARE AND XEN HYPERVISOR ANKIT KUMAR, SAVITA SHIWANI 1 M. Tech Scholar, Software Engineering, Suresh Gyan Vihar University, Rajasthan, India, Email:
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads G. Suganthi (Member, IEEE), K. N. Vimal Shankar, Department of Computer Science and Engineering, V.S.B. Engineering College,
Towards an understanding of oversubscription in cloud
IBM Research Towards an understanding of oversubscription in cloud Salman A. Baset, Long Wang, Chunqiang Tang [email protected] IBM T. J. Watson Research Center Hawthorne, NY Outline Oversubscription
Resource Provisioning in Single Tier and Multi-Tier Cloud Computing: State-of-the-Art
Resource Provisioning in Single Tier and Multi-Tier Cloud Computing: State-of-the-Art Marwah Hashim Eawna Faculty of Computer and Information Sciences Salma Hamdy Mohammed Faculty of Computer and Information
Optimizing Shared Resource Contention in HPC Clusters
Optimizing Shared Resource Contention in HPC Clusters Sergey Blagodurov Simon Fraser University Alexandra Fedorova Simon Fraser University Abstract Contention for shared resources in HPC clusters occurs
Optimal Service Pricing for a Cloud Cache
Optimal Service Pricing for a Cloud Cache K.SRAVANTHI Department of Computer Science & Engineering (M.Tech.) Sindura College of Engineering and Technology Ramagundam,Telangana G.LAKSHMI Asst. Professor,
International Journal of Scientific & Engineering Research, Volume 6, Issue 4, April-2015 36 ISSN 2229-5518
International Journal of Scientific & Engineering Research, Volume 6, Issue 4, April-2015 36 An Efficient Approach for Load Balancing in Cloud Environment Balasundaram Ananthakrishnan Abstract Cloud computing
Service allocation in Cloud Environment: A Migration Approach
Service allocation in Cloud Environment: A Migration Approach Pardeep Vashist 1, Arti Dhounchak 2 M.Tech Pursuing, Assistant Professor R.N.C.E.T. Panipat, B.I.T. Sonepat, Sonipat, Pin no.131001 1 [email protected],
Group Based Load Balancing Algorithm in Cloud Computing Virtualization
Group Based Load Balancing Algorithm in Cloud Computing Virtualization Rishi Bhardwaj, 2 Sangeeta Mittal, Student, 2 Assistant Professor, Department of Computer Science, Jaypee Institute of Information
Multilevel Communication Aware Approach for Load Balancing
Multilevel Communication Aware Approach for Load Balancing 1 Dipti Patel, 2 Ashil Patel Department of Information Technology, L.D. College of Engineering, Gujarat Technological University, Ahmedabad 1
AN EFFICIENT LOAD BALANCING APPROACH IN CLOUD SERVER USING ANT COLONY OPTIMIZATION
AN EFFICIENT LOAD BALANCING APPROACH IN CLOUD SERVER USING ANT COLONY OPTIMIZATION Shanmuga Priya.J 1, Sridevi.A 2 1 PG Scholar, Department of Information Technology, J.J College of Engineering and Technology
CloudScale: Elastic Resource Scaling for Multi-Tenant Cloud Systems
CloudScale: Elastic Resource Scaling for Multi-Tenant Cloud Systems Zhiming Shen, Sethuraman Subbiah, Xiaohui Gu, Department of Computer Science North Carolina State University {zshen5,ssubbia2}@ncsu.edu,
Dynamic Resource allocation in Cloud
Dynamic Resource allocation in Cloud ABSTRACT: Cloud computing allows business customers to scale up and down their resource usage based on needs. Many of the touted gains in the cloud model come from
A Middleware Strategy to Survive Compute Peak Loads in Cloud
A Middleware Strategy to Survive Compute Peak Loads in Cloud Sasko Ristov Ss. Cyril and Methodius University Faculty of Information Sciences and Computer Engineering Skopje, Macedonia Email: [email protected]
Variations in Performance and Scalability when Migrating n-tier Applications to Different Clouds
Variations in Performance and Scalability when Migrating n-tier Applications to Different Clouds Deepal Jayasinghe, Simon Malkowski, Qingyang Wang, Jack Li, Pengcheng Xiong, Calton Pu Outline Motivation
Elastic Load Balancing in Cloud Storage
Elastic Load Balancing in Cloud Storage Surabhi Jain, Deepak Sharma (Lecturer, Department of Computer Science, Lovely Professional University, Phagwara-144402) (Assistant Professor, Department of Computer
SLA-based Admission Control for a Software-as-a-Service Provider in Cloud Computing Environments
SLA-based Admission Control for a Software-as-a-Service Provider in Cloud Computing Environments Linlin Wu, Saurabh Kumar Garg, and Rajkumar Buyya Cloud Computing and Distributed Systems (CLOUDS) Laboratory
Network Infrastructure Services CS848 Project
Quality of Service Guarantees for Cloud Services CS848 Project presentation by Alexey Karyakin David R. Cheriton School of Computer Science University of Waterloo March 2010 Outline 1. Performance of cloud
Cost Effective Selection of Data Center in Cloud Environment
Cost Effective Selection of Data Center in Cloud Environment Manoranjan Dash 1, Amitav Mahapatra 2 & Narayan Ranjan Chakraborty 3 1 Institute of Business & Computer Studies, Siksha O Anusandhan University,
OPTIMIZED PERFORMANCE EVALUATIONS OF CLOUD COMPUTING SERVERS
OPTIMIZED PERFORMANCE EVALUATIONS OF CLOUD COMPUTING SERVERS K. Sarathkumar Computer Science Department, Saveetha School of Engineering Saveetha University, Chennai Abstract: The Cloud computing is one
BridgeWays Management Pack for VMware ESX
Bridgeways White Paper: Management Pack for VMware ESX BridgeWays Management Pack for VMware ESX Ensuring smooth virtual operations while maximizing your ROI. Published: July 2009 For the latest information,
Green Cloud Computing 班 級 : 資 管 碩 一 組 員 :710029011 黃 宗 緯 710029021 朱 雅 甜
Green Cloud Computing 班 級 : 資 管 碩 一 組 員 :710029011 黃 宗 緯 710029021 朱 雅 甜 Outline Introduction Proposed Schemes VM configuration VM Live Migration Comparison 2 Introduction (1/2) In 2006, the power consumption
Environments, Services and Network Management for Green Clouds
Environments, Services and Network Management for Green Clouds Carlos Becker Westphall Networks and Management Laboratory Federal University of Santa Catarina MARCH 3RD, REUNION ISLAND IARIA GLOBENET 2012
VIRTUAL RESOURCE MANAGEMENT FOR DATA INTENSIVE APPLICATIONS IN CLOUD INFRASTRUCTURES
U.P.B. Sci. Bull., Series C, Vol. 76, Iss. 2, 2014 ISSN 2286-3540 VIRTUAL RESOURCE MANAGEMENT FOR DATA INTENSIVE APPLICATIONS IN CLOUD INFRASTRUCTURES Elena Apostol 1, Valentin Cristea 2 Cloud computing
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis Felipe Augusto Nunes de Oliveira - GRR20112021 João Victor Tozatti Risso - GRR20120726 Abstract. The increasing
Load Balancing to Save Energy in Cloud Computing
presented at the Energy Efficient Systems Workshop at ICT4S, Stockholm, Aug. 2014 Load Balancing to Save Energy in Cloud Computing Theodore Pertsas University of Manchester United Kingdom [email protected]
TPC-W * : Benchmarking An Ecommerce Solution By Wayne D. Smith, Intel Corporation Revision 1.2
TPC-W * : Benchmarking An Ecommerce Solution By Wayne D. Smith, Intel Corporation Revision 1.2 1 INTRODUCTION How does one determine server performance and price/performance for an Internet commerce, Ecommerce,
Delivering Quality in Software Performance and Scalability Testing
Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
AN ADAPTIVE DISTRIBUTED LOAD BALANCING TECHNIQUE FOR CLOUD COMPUTING
AN ADAPTIVE DISTRIBUTED LOAD BALANCING TECHNIQUE FOR CLOUD COMPUTING Gurpreet Singh M.Phil Research Scholar, Computer Science Dept. Punjabi University, Patiala [email protected] Abstract: Cloud Computing
White Paper. Recording Server Virtualization
White Paper Recording Server Virtualization Prepared by: Mike Sherwood, Senior Solutions Engineer Milestone Systems 23 March 2011 Table of Contents Introduction... 3 Target audience and white paper purpose...
Windows Server Performance Monitoring
Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly
Scalable Architecture on Amazon AWS Cloud
Scalable Architecture on Amazon AWS Cloud Kalpak Shah Founder & CEO, Clogeny Technologies [email protected] 1 * http://www.rightscale.com/products/cloud-computing-uses/scalable-website.php 2 Architect
CiteSeer x in the Cloud
Published in the 2nd USENIX Workshop on Hot Topics in Cloud Computing 2010 CiteSeer x in the Cloud Pradeep B. Teregowda Pennsylvania State University C. Lee Giles Pennsylvania State University Bhuvan Urgaonkar
Paul Brebner, Senior Researcher, NICTA, [email protected]
Is your Cloud Elastic Enough? Part 2 Paul Brebner, Senior Researcher, NICTA, [email protected] Paul Brebner is a senior researcher in the e-government project at National ICT Australia (NICTA,
Performance Prediction, Sizing and Capacity Planning for Distributed E-Commerce Applications
Performance Prediction, Sizing and Capacity Planning for Distributed E-Commerce Applications by Samuel D. Kounev ([email protected]) Information Technology Transfer Office Abstract Modern e-commerce
GeoCloud Project Report USGS/EROS Spatial Data Warehouse Project
GeoCloud Project Report USGS/EROS Spatial Data Warehouse Project Description of Application The Spatial Data Warehouse project at the USGS/EROS distributes services and data in support of The National
Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems
Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems Applied Technology Abstract By migrating VMware virtual machines from one physical environment to another, VMware VMotion can
Online Failure Prediction in Cloud Datacenters
Online Failure Prediction in Cloud Datacenters Yukihiro Watanabe Yasuhide Matsumoto Once failures occur in a cloud datacenter accommodating a large number of virtual resources, they tend to spread rapidly
Directions for VMware Ready Testing for Application Software
Directions for VMware Ready Testing for Application Software Introduction To be awarded the VMware ready logo for your product requires a modest amount of engineering work, assuming that the pre-requisites
Dynamic resource management for energy saving in the cloud computing environment
Dynamic resource management for energy saving in the cloud computing environment Liang-Teh Lee, Kang-Yuan Liu, and Hui-Yang Huang Department of Computer Science and Engineering, Tatung University, Taiwan
can you effectively plan for the migration and management of systems and applications on Vblock Platforms?
SOLUTION BRIEF CA Capacity Management and Reporting Suite for Vblock Platforms can you effectively plan for the migration and management of systems and applications on Vblock Platforms? agility made possible
