PICS: A Public IaaS Cloud Simulator
|
|
|
- Madeleine Melton
- 10 years ago
- Views:
Transcription
1 : A Public IaaS Cloud Simulator In Kee Kim, Wei Wang, and Marty Humphrey Department of Computer Science University of Virginia {ik2sb, wwang}@virginia.edu, [email protected] Abstract Public clouds become essential for many organizations to run their applications because they provide huge financial benefits and great flexibility. However, it is very challenging to accurately evaluate the performance and cost of applications without actual deployment on the clouds. Existing cloud simulators are generally designed from the perspective of cloud service providers, thus they can be under-developed for answering questions for the perspective of cloud users. To solve this prediction and evaluation problem, we created a Public Cloud IaaS Simulator (). enables the cloud user to evaluate the cost and performance of public IaaS clouds along with such dimensions like VM and storage service, resource scaling, job scheduling, and diverse workload patterns. We extensively validated by comparing its results with the data acquired from real public IaaS cloud using real cloud-applications. We show that provides highly accurate simulation results (less than 5% of average errors) under a variety of use cases. Moreover, we evaluated sensitivity with imprecise simulation parameters. The results show that still provides very reliable simulation results with imprecise simulation parameters and performance uncertainty. Index Terms Cloud Simulator, Cloud Performance Evaluation, Public IaaS Clouds, Cloud Applications, Cloud Resource Management. I. INTRODUCTION For many organizations today, the issue is not whether to use public IaaS cloud computing (e.g. Amazon Web Services [] or Microsoft Azure [2]) but rather how best to use public IaaS cloud capabilities. The approach taken by many organizations getting started is that a few super-human users within the organization deploy a small-scale test cloud-application on the public cloud of choice. This cloud-application usually has two components: a resource manager and user-applications, where the resource manager is responsible for the effective and efficient execution of user-applications, such as web search, SNS, big data analytics, and scientific applications. Then, the next steps are to scale-up the test cloud-application in order to better assess the capabilities and viability in the context of the organization s particular goals and requirements. The key limitation these potential cloud users are facing is that there does not appear to be a viable alternative for evaluating the cloud other than to actually use the cloud. This approach is problematic for a number of reasons. First, the time-consuming and sometimes tedious learning of idiosyncratic cloud APIs can distract from the real issue, which centers around specific application logic and requirements. Second, the evaluation tends to be specific to one cloud and not readily generalizable to other clouds. Third, to evaluate at scale via this approach, the cloud-application typically requires significant changes to its architecture. Fourth, the evaluation is geared toward the present time, whereas longerterm issues/concerns are often more important than short-term issues of today s cloud there is little opportunity to ask whatif questions of performance, reliability or cost. There are a number of cloud simulators that exist (e.g. CloudSim [], icancloud [8], GreenCloud [5]). They have the potential to aid in this evaluation. However, in general, these simulators are designed to answer questions related to datacenter management (e.g., how many concurrent users can I support if I deploy a private cloud on my existing hardware?) Furthermore, typical tools [, 2, 7, 8] provided by commercial cloud vendors only address a small part of the concerns, which is an overall cost in the aggregate based on resource utilization (e.g., how much does it cost to run small VMs for one month and to store TB in long-term cloud storage for year?) These existing cloud simulators and vendor tools do not broader, end-to-end concerns such as: ) What is the average/worst response time for a particular application and a particular arrival pattern, when servicing via a specific VM type and a specific set of auto-scaling rules? 2) Which public IaaS cloud provides the best cost efficiency for a particular application, given the different VM configurations, storage services and pricing models? 3) Which resource management and job scheduling policy maximize the cost efficiency and minimize the response time for a particular application? 4) Above all, if a simulator can provide answers for above questions, another question the cloud users could have is how reliable are the simulation results? or how accurately can the simulator resemble actual clouds behavior? To enable potential public IaaS cloud users to address these and other challenging concerns without actually deploying the cloud-application, we have create the, a tracebased public IaaS cloud simulator. provides following capabilities to address the potential cloud user s concerns: Assessing a wide range of properties of cloud services and the cloud-applications, including the cloud cost, job response time, and VM utilization. Allowing the simulation users to specify different workload types, including varying job arrival patterns and SLA requirements (e.g. deadline). Source code of is publicly available on the project web site ik2sb/ [6]
2 Datacenter Issues Cloud User Issues TABLE I: Simulation Capabilities of Existing Cloud Simulator and. Simulation Capabilities CloudSim [] icancloud [8] GreenCloud [5] VM management (allocation, provisioning, scheduling, migration) Yes Yes No No Physical resource management and scaling Yes Yes Limited No Network resource management Yes Yes Yes No Power consumption management Yes Yes Yes No Federated cloud management Yes No No No Datacenter workload management Yes No Yes No Horizontal VM auto-scaling (scale-in/out) Limited No No Yes Vertical VM auto-scaling (scale-up/down) No No No Yes Storage service management No Limited No Yes Job/Application scheduling Yes Yes No Yes Billing management (cost optimization) Limited Limited No Yes Application/Job failure management No No No Yes IaaS performance uncertainty No No No Yes Simulating of a broad range of resource management policies: i.e., horizontal/vertical 2 auto-scaling, job scheduling and job failure policies. Enabling the users to evaluate the performance of different types of public IaaS cloud configurations such as a variety of resource types (VM and storage), unique billing models, and performance uncertainty [9 2]. We validated the correctness of by focusing on following capabilities: cloud cost, the number of created VMs, VM utilization, horizontal and vertical scaling of cloud resources, and job deadline satisfaction rate. We compare the simulation results of with the actual measurements from realworld cloud-applications on Amazon Web Services (AWS). The results show that provides very accurate simulation results (less than 5% of average errors) in every validation cases. Furthermore, we conduct a sensitivity test of with ±% and ±2% of imprecise simulation parameter by considering of the performance uncertainty of IaaS clouds. The results show that with imprecise simulation parameters still provides very reliable simulation results. The contributions of this paper are: ) A public IaaS cloud simulator,, which is versatile and satisfies cloud user s various needs of cloud application and cloud service evaluations without actual cloud deployment. is easy-to-configure and fast, allowing large design space exploration in short amount of time. 2) A first simulator that supports both horizontal and vertical cloud resource scaling, to the best of our knowledge. 3) A thorough validation of against the results from real public IaaS (AWS), demonstrating that the provides accurate results that faithfully simulates real cloud services. 4) An analysis to examine sensitivity to the performance uncertainty of real cloud services. This analysis shows that can still provide reliable simulation results even when user-provided parameters are imprecise due to the unstable real cloud performance. The rest of this paper is organized as follows: Section II contains the related work. Section III describes the design and implementation of. Section IV contains validation of. Section V is discussion focusing on sensitivity of simulation parameters and Section VI provides the conclusion. 2 Vertical scaling means scale-up or scale-down operations for the cloud resources. i.e., migrating the user-applications to higher (scale-up) or lower (scale-down) performance instances. [9] II. RELATED WORK CloudSim [] is a well-known simulation framework for cloud computing and is designed to support various simulation tests across three major cloud service models (e.g. SaaS, PaaS, and IaaS). CloudSim is widely used in cloud computing research because of its various capabilities for cloud simulations such as VM allocation and provisioning, energy consumption, network management, and federated clouds. It also has several extensions [, 2, 4, 22] due to its extensible simulation architecture. These extensions support simulations for large-scale cloud applications according to the geographical distribution of users and datacenters [22], a network-based application model [4], complex scientific workflows [2] and homogeneity in the performance of computational resources and resource failure model during task executions []. icancloud [8] is a holistic simulation platform for cloud computing and offers various simulation functionalities including resource provisioning, energy consumption, and userdefined resource broker model. The goal of icancloud is to predict tradeoffs between cost and performance of applications executed on virtualized environments. icancloud offers an unique feature to configure various storage systems and a pre-defined IaaS model based on Amazon EC2. Moreover, icancloud supports large-scale simulation on distributed environments by employing MPI/POSIX-based API. GreenCloud [5] is a packet-level simulator for data centers focusing on data communication and energy cost in cloud computing. This simulator is implemented based on NS-2 [5] and offers extensive network and communication capabilities (e.g. full implementation of TCP/IP reference model, various workload models) to simulate data center operations. However, due to its narrow focus of simulation for cloud system, it lacks many simulation capabilities in both cloud user- and infrastructure-oriented resource management. Cloud providers [, 2] and third-party cloud services [7, 8] provide a tool for calculating the overall cost in the aggregate based on resource utilization. These services are only addressing a small part of the cloud user s concern (cloud cost). They do not support any capability to evaluate performance such as response time and detailed VM utilizations. Moreover, they only support resource management policies offered by commercial cloud services (e.g. AWS auto-scaling) to handle the cloud resources, so it is impossible to evaluate the cost and performance of the clouds via particular resource management policies designed for the cloud user s applications. 2
3 As discussed in this section, the capabilities of all existing simulators more focus on datacenter issues such as power consumption and physical resource management. These capabilities are suitable to answer following questions: How many concurrent users can I support if I deploy a private cloud on my existing hardware? How can I effectively manage my existing hardware resources if I deploy a private clouds on my hardware? However, the cloud users have different perspective on cloud systems. The super-human users (mentioned in Section I) is responsible for build cloud applications, which schedule jobs submitted by end users and manage cloud resources (VM, storage, network, etc.) that execute these jobs. For these users, they are more interested in how best to use clouds rather than operating cloud systems. Their major concerns are cost optimization, VM usage, resource scaling, and storage service for their cloud-applications. Due to these disparities between existing cloud simulators and the cloud users, the cloud users have difficulties to have existing simulators to deal with the problem that how to evaluate the cloud other than to actually use the cloud by focusing on cloud user s concerns. This limitation motivate us to conduct this research. The purpose of our work is to provide a viable simulation solution for the normal cloud users. III. SIMULATOR DESIGN A. Simulator Design Overview Goal: The goal of is to correctly simulate the behaviors of public clouds from the cloud users perspectives as if they deploy a particular cloud-application on public IaaS cloud. From the potential cloud users perspective, the cloud cost, the job response time and the resource usages (VM utilization and size) are the most important criteria to evaluate cloud service for their cloud-applications. Key challenges to design are: How to correctly model the behavior of public clouds. More specifically, how to handle a variety of resources (e.g. VM, storage, and network). How to properly model the behavior of the cloudapplication. More specifically, how to handle varying workload patterns and performance uncertainty [9 2]. How to correctly model the behavior of cloud users resource management policy. For the first challenge, we designed a convenient configuration interface for the simulation users to define diverse types of cloud resources as an input of the simulator. For the second challenge, we collected data from real public clouds, profile performance uncertainty, and leverage these results to design the simulator. For the last challenge, we provided abundant configuration options to let user define various custom resource management policies. Input: requires five types of inputs: VM configurations, storage and network configurations, workload information, job scheduling policies, and cloud resource management policies. The VM configuration includes detailed specifications of VMs, such as cost (including public cloud s billing model), machine performance (CPU, RAM), network performance, and the range of startup delays [7] of cloud resources. This configuration is designed to simulate various VM types of current public clouds because public clouds have a diversity of VM types based on performance, cost, and the VM s purpose. The storage and network configuration has detailed information on storage and network service on public IaaS clouds. We model storage services to reflect current public clouds actual characteristics based on Amazon S3 and Azure Storage. To model network service, we collect data from actual network- I/O test by using various types of VM on real cloud services. We then reflect the data to simulator configurations. The workload information contains detailed configurations on job generation such as job arrival time, job execution time, job deadline, size of network I/O, etc. This input reflects end users requests to a particular cloud-application. The job scheduling policy defines various cloud user job scheduling policy for end users requests. includes three types of job scheduling policies, i.e., EDF (Earliest Deadline First), Round-Robin, and Greedy scheduling mechanisms. In the future, will support more complicated job scheduling policies and APIs. Furthermore, the simulation users can configure recovery policies for job failures, which enable the users to conduct realistic tests for public cloud services. The cloud resource management policy contains detailed specifications for the cloud resource management. This input supports simulation configurations for maximum number of concurrent VMs, and horizontal and vertical scaling policies. Moreover, the simulator users can configure various ways to monitor and analyze an end user s job request patterns such as linear and time-series methods. The simulator users are able to leverage this mechanism to design and test their own resource management mechanisms. Output: provides three types of output: cost, resource usage, and job processing results. Cost results provide overall cost for total cloud usage, cost trace at fine grained time interval, and usage cost per cloud resource type. Overall cost means how much cost the simulation users are expected to spend on servicing a particular cloud-application under a particular job arrival pattern. Cost trace provides fluctuation and accumulation of usage cost at fine grained time interval. For example, expected usage cost for time at t is $ and $2 for time at t +. The usage cost per cloud resource type provides a detailed cost based on resource types such as how much cost they spent on each type of resources (e.g. VM, storage, network). Resource usage results provide detailed information on resource usage such as how many VMs created, how much storage space spent, and how many network data sent and received. Moreover, these results offer fine-grained traces for both horizontal and vertical scaling. These traces help users determine the number and types of VMs running at time t and t +. The users also check when (time) and how (scalingup/down) the vertical scaling decisions are made. 3
4 Fig. : Design overview of. Job processing results provide specific information on job processing, such as the job arrival/start/finish time, as well as whether the job deadlines are satisfied (if specified). These results are basic metrics to evaluate the user job scheduling and resource management policies. Moreover, these results include the analysis of job arrival patterns (min, max, and average of job arrival time, as well as predictions for the next job arrivals) using linear and time-series methods. B. Simulator Internals is composed of three hierarchical layers: simulation configuration layer, simulation entity layer, and simulation core layer as shown in Fig.. The simulation configuration layer is responsible for accepting the user inputs which are passed on to the simulation entity layer. The simulation entity layer contains the simulation logic and is driven by the events generated from the simulation core layer. The simulation core layer is also responsible for producing simulation reports. Because the simulation configuration layer has already been covered by the previous section, we focus on the two remain layers here. ) Simulation Core Layer: The simulation core layer consists of Simulation Wall Clock, Simulation Event Processor and Simulation Report Generator. Simulation Wall Clock is working as a heart for by managing simulation clock. (Basic time unit is a second.) To manage the simulation clock, Simulation Wall Clock collaborates with Simulation Event Processor. When the simulation clock is updated, this component sends a clock update event to Simulation Event Processor to notify an update of clock. Simulation Event Processor handles every event generated in a simulation. After receiving the clock update event from Simulation Wall Clock, this component passes on this event to simulation entities, which advance their simulation of the behaviors of public cloud services and cloud-application to the new clock cycle. This component also manages simulation timer events. Simulation entities use these events to register timers within Simulation Event Processor. This component is responsible for notifying the corresponding simulation entity when a timer expires. Moreover, this component handles all simulation command events, all invoked events between simulation entities, in order to validate the authentication of the command events. Because this component can monitor all events in the simulation, this plays a main role to generate real-time traces for the simulation. Simulation Report Generator is used to generate output reports for the simulation results such as simulation trace and overall reports. Reporting simulation results at fine-grained time interval is an important capability for simulators. Simulation Report Generator is responsible for generating realtime traces called simulation trace reports. These reports contain simulation results at user-defined time interval for cost, resource scaling, and job processing information. 2) Simulation Entity Layer: Simulation entity layer are composed of three entities, which are Workload Generator Entity, Cloud Application Management Entity, and IaaS Cloud Entity. These entities reflect three main components of cloudapplications and public clouds. Workload Generator Entity generates jobs and sends them to the Cloud Application Management Entity to process, based on the workload file from the simulation user. The workload file includes job arrival time, job execution time, job deadlines, I/O data specification, and data transfer specification to support various types of possible end user s job requests. Cloud Application Management Entity is designed to resemble the cloud-application s behaviors. There are subcomponents in this entity: job management module, resource management module and workload monitoring module. The job management module is designed to simulate job management policies of a cloud-application. This module conducts three operations for the simulation; job scheduling, job response time estimation, and job failure management. Job scheduling is used to perform job scheduling policies of the cloud users and assign an arrived job from Workload Generator Entity to a selected VM from IaaS Cloud Entity. Job response time estimation predicts the job response time, which is defined as the clock time when a particular job finishes. The prediction of job response time is based on job execution time and the current load of available cloud resources, and is used for the job scheduling. Job failure management is used for job failure simulation in the case of application failure or cloud infrastructure problem (e.g. VM down). Job failure management supports four types of recovery policies for job failure simulations. The resource management module is designed to simulate the resource management policies of a cloud-application. It handles three types of cloud resources of public IaaS clouds, i.e., VM, storage and network. For the VM management, this module enables the simulation users to examine their VM selection mechanisms (e.g. cost, performance, cost/performance- 4
5 TABLE II: Validation Workloads for. Poisson pattern has 3s of average job arrival rate and of standard deviation. Bursty pattern has 5s of average job arrival rate and.5 of standard deviation. (WC: Word Count, PI: PI Calculation, TS: TeraSort) Worloads Scaling Job Arr. Pattern Job Type # of Concurr. VMs Used VM Types # of Jobs Avg. Job DL. Std. Dev WL # m3.medium WC WL #2 272s WL #3 m3.large Poisson PI WL #4 538s 266. WL #5 m3.xlarge TS 65s 53.9 WL #6 Horizontal Scaling Unlimited 2 WL #7 m3.medium WC 5s 27.6 WL #8 WL #9 m3.large Bursty PI WL # 55s WL # m3.xlarge TS WL #2 2s WL #3 WC 3 5s Poisson WL #4 PI 7 29s Vertical Scaling 5 WL #5 WC 5 523s Bursty WL #6 PI 45s balanced) and VM scaling mechanisms (e.g. horizontal, vertical scaling). For storage and network resources, this module can simulate File Read/Write operations to cloud storage and data transmissions by collaborating with IaaS Cloud Entity. The workload monitoring module is designed to analyze workload arrival patterns from Workload Generator Entity. The simulation users can leverage this module to improve their policies for job scheduling and resource management for variable workload patterns. IaaS Cloud Entity is used to simulate the public cloud s behavior. It has sub-modules to simulate public clouds, which include cost calculation, VM management, storage service and network service module. The cost calculation module calculates all cloud cost used by Cloud Application Management Entity. It generates the cost traces based on user-defined time interval and creates the final results when the simulation is completed. The VM repository module manages the configurations of all VM types defined by the users and resembles the on-demand VM service of IaaS clouds. This module stores VM information to correctly measure VM usage cost and simulate job execution on VMs. Moreover, this module generates startupdelay for new VM creations based on user input, and simulates all operations of the VMs. This module also handles workloads on VM such as job execution and failure generation. Storage service and network service module simulates file I/O and data transmission operations based on their configurations and the workload information. It generates the overall and real-time traces of the usage of storage and network services for the workloads. A. Experiment Setup IV. SIMULATOR VALIDATION In order to validate the simulation results of, we compared results with a real cloud application on AWS. We design and implement a cloud-application that executes user-applications with three different types of MapReduce [3] jobs and two job arrival patterns. The workflow of the cloud-application goes through the following five key steps: ) Job execution time prediction via recent execution history, 2) EDF job scheduling, 3) Cost-based VM selection, 4) Deadline-based horizontal and vertical auto-scaling, and 5) Job Execution. The cloud-application starts with receiving jobs from the end users. It conducts job execution time prediction via recent execution history for incoming jobs. The cloudapplication schedules them by the EDF job scheduling and sends them to the work queue in the VMs of choice. For VM selection, this cloud-application uses cost-based VM selection that selects the cheapest VM type that meets the deadline for a job. For the VM scaling, the cloud-application makes the scaling decision based on the deadline-based horizontal and vertical auto-scaling. The cloud-application first determines the fastest possible response time for the new job based on the load of the job queues of active VMs. If the fastest possible response time still misses the job deadline, horizontal or vertical scaling is engaged. In the case of horizontal scaling, a new VM is created for the new job. In the case of vertical scaling, an active VM is scaled-up to a higher performance VM for the new job. Note that vertical scaling happens when there is a limit on the number of active VMs which prevents the creation of new VMs. Additionally, after each job finishes, the cloud-application checks active VMs. If the active VMs provide more than enough computation power to satisfy deadline requirements, then a scale-down happens to improve the cost efficiency. Once the current VM type is determined, the job is assigned to that VM for job execution. For the validation of, we use three types of MapReduce jobs (Word count, PI calculation, and Terasort) [4]. Word count is an I/O- and memory-bound job, and it uses S3 storage to download input dataset and upload the final result. PI calculation is a fully-parallelized CPU-bound job. Terasort is a standard Hadoop benchmark application. These jobs were randomly generated based on two arrival patterns (Poisson and Bursty). Poisson arrival pattern has an average job arrival interval of 3 seconds with a standard deviation of Bursty pattern has an average job arrival interval of 5 seconds with a standard deviation of.5. We deployed the cloud-application on AWS. The reasons that we use AWS for the validation are as follows: ) AWS is widely used public IaaS cloud; 2) according to recent works [6, 9 2], AWS EC2 performance fluctuates a lot and AWS has less predictability (higher variance) than other public 5
6 TABLE III: Simulation errors for horizontal scaling cases. Workloads Cost # of VMs VM Util. Job DL. WL # 3.% 3.4%.%.6% WL #2 2.4%.8%.4%.% WL #3.8% 2.3%.7%.6% WL #4 5.2% 4.7% 2.2% 3.2% WL #5 2.2% 7.%.6% 2.% WL #6.8% 2.4% 2.4% 2.6% WL #7 2.6% 3.6%.5% 2.5% WL #8 3.6%.%.%.% WL #9.5%.5%.7%.6% WL #.9%.5%.9% 3.9% WL #.% % 2.2%.5% WL #2 6.%.4%.3% 3.8% Average Error 2.6% 2.4%.3%.9% IaaS clouds, therefore AWS is better than other IaaS clouds when evaluating the sensitivity of to the performance uncertainty of public clouds. Our cloud-application uses four types of on-demand EC2 VM instances, which are m3.medium, m3.large, m3.xlarge, and m3.2xlarge. These four types of EC2 on-demand instances are general purpose VMs and commonly used by the cloud users. Based on the above experimental configurations, we created 6 validation workloads as shown in Table II. These workloads are categorized based on job arrival patterns (Poisson and bursty), job types, single or multiple VM types, and scaling policies. WL # #6, #3, and #4 are for the tests under Poisson job arrival pattern, and others are for bursty arrival pattern. WL #, #2, #7, #8 and #3 process word count jobs, WL #3, #4, #9, # and #4 handles PI calculation jobs, and WL #5, #6, # and #2 deal with Terasort jobs. WL #, #3, #5, #7, #9, and # only use a single VM type (e.g. m3.medium, m3.large, or m3.xlarge) in order to validate a case when a cloud-application uses a single type of VM. The others use all four types of general purpose EC2 instances. This is to test more complicated use cases for the VM resource management. For scaling validation, WL # #2 are for horizontal scaling and WL #3 #6 are for vertical scaling use cases. We submitted these 6 workloads to and the cloud-application running on AWS, and measured the cloud cost, the total number of created VMs, VM utilization, and job deadline satisfaction rate. These metrics are expressed as equation (), (2), and (3). We then measured the simulation error by equation (4) Cost = n i= cost V M i () V M Utilization = n i= T ime JobExec.,V M i n i= T ime T otalrun.,v M i (2) Job DL Satisfaction Rate = N DeadlineSatisfiedJobs N AllJobs (3) Simulation Error = % (4) B. Horizonal Scaling Cases Simulation We use 2 workloads (WL # #2) for the validation of horizontal scaling cases. The Cost column of Table III shows simulation error of the overall cloud cost over the actual measurements on AWS. The average error of the cloud cost from is only 2.6% compared to the actual results. The highest error is only 6.% (WL #2). A more important metric is the cost trace because it shows how accurately simulates the behavior of the cloud-application and public (a) Cost Trace for WL # (b) Cost Trace for WL # (c) Cost Trace for WL # (d) Cost Trace for WL # Fig. 2: Cost trace for horizontal scaling cases. IaaS cloud at fine-grained intervals. To show these traces, we select four complicated workloads, which are WL #4, #6, #8, #. Fig. 2 shows these four cost traces. The other workloads have similar results. As shown in Fig. 2, is able to accurately calculate the cloud cost at each time interval, demonstrating that correctly resembles the behavior of the cloud-application and public IaaS cloud for each step of execution. The simulation error of created VM numbers over the actual measurements are shown in the # of VMs column of Table III. can highly accurately calculate the number of created VMs with an average error of 2.4%. The highest error is only 7.% (WL #5). For the workloads that use multiple types of VMs, a precise simulation of the number of VMs for each VM type is also critical to the cloud users to determine if can accurately resemble the cloud-application s resource management policies (VM selection and scaling) and public cloud s behaviors. Fig. 3 shows the result of the number of created VMs for each VM type for four workloads (WL #4, #6, #, #2). The other workloads have similar results. As Fig. 3 shows, can also accurately simulate the VM numbers for 6
7 Norm. VM Numbers WL#4 WL#6 WL# WL#2 Workloads M L XL 2XL Fig. 3: The VMs numbers per VM type, normalized over the actual measurement (M: m3.medium, L: m3.large, X: m3.xlarge, XL: m3.xlarge) Norm. VM Util WL#4 WL#6 WL# WL#2 Workloads M L XL 2XL Fig. 4: The VM utilizations per VM type, normalized over the actual measurement (M: m3.medium, L: m3.large, X: m3.xlarge, XL: m3.xlarge) each type of VMs. For the horizontal scaling cases, the cloudapplicatoin created an average of 42.5 VMs and the range of the created number of actual VMs is from 5 (WL #5) to 7 (WL #2). Similarly, the overall simulation error of VM utilization are shown in the VM Util. column of Table III, while the detailed per-vm type utilization results are shown in Fig. 4. Average error of overall VM utilization results between the actual measurements and the simulations is only.3%. For the detailed results, is able to accurately simulate the utilization for each VM type with.5 2.4% of simulation error for most cases. The worst case is m3.large instance of WL #4 and the simulation error is only 5.8%. In addition, we conducted a validation focusing on job deadline satisfaction rate. Deadline satisfaction rate is an important metric for a cloud-application to ensure the cloudapplication s job scheduling and resource management policy meets certain temporal requirements, such as those found in SLA. We measured the job deadline satisfaction rates of, and compared them with the results of the actual cloudapplication on AWS. Note that the deadlines in our workload are generated randomly. The overall results are shown in the Job DL. column of Table III. The average error of is only.9%. The worst case error is about 4% (WL #, #2). More importantly, we measured the traces for job deadline satisfaction because we want to determine if can precisely simulate whether a particular job satisfies its deadline or not. Fig. 5 shows the job traces for two worst workloads (WL #, #2). Fig. 5 shows that, even for the worst cases, accurately simulates the behavior of AWS at every phase of the execution. We also measured the VM scaling of the actual cloudapplication on AWS and, over the whole course of the execution. Due to space limitation, Fig. 6 shows the scaling traces for only four workloads (WL #4, #8, #, #2). The other workloads have similar behaviors. As shown in Fig. 6, traces from and the actual cloud-application closely Norm. Job Deadline Satisfaction Rate Norm. Job Deadline Satisfaction Rate Job Sequence (a) WL # Job Sequence (b) WL #2 Fig. 5: Job deadline satisfaction traces for horizontal scaling cases. Norm. VM Number Norm. VM Number Norm. VM Number Norm. VM Number (a) VM scaling trace for WL # (b) VM scaling trace for WL # (c) VM scaling trace for WL # (d) VM scaling trace for WL #2 Fig. 6: Horizontal VM scaling traces. 7
8 TABLE IV: Simulation errors for vertical scaling cases. Workloads Cost # of VMs VM Util. Job DL. WL #3 6.% 7.% 4.3%.8% WL #4 3.%.9% 2.4% 4.6% WL #5 3.2% 3.4%.7%.9% WL #6 9.7%.9% 3.3% 3.2% Average Error 5.5% 3.6% 2.9% 2.6% (a) VM Cost Trace for WL # (b) VM Cost Trace for WL #6 Fig. 7: Cost traces for vertical scaling cases. match each other, which means accurately simulates horizontal scaling at every phase of the execution. For the VM scaling traces, the average number of running VMs in parallel (by the cloud-application) is 39 and the cloud-application runs from 2 (WL #5) to 7 VMs (WL #2) in parallel. C. Vertical Scaling Cases Due to the importance of vertical scaling in the near future, also supports vertical scaling simulation. We validate the results involving vertical scaling using four workloads (WL #3 #6). Note that for these experiments, both horizontal and vertical scaling are enabled, similar to real-life usages of cloud services. Table IV shows the overall simulation errors for vertical scaling cases. The results show that can accurately simulate the overall results for vertical scaling workloads: for the cloud cost, the average error is only 5.5% ( Cost column); for the number of created VMs, the average error is only 3.6% ( # of VMs column); for the VM utilization, the average error is only 2.9% ( VM Util. column); for deadline satisfaction, the average error is only 2.6% ( Job DL. column). To demonstrate the accuracy of simulation at finegrained time intervals and resource types, we present the detailed results for these measurements, which are the cost traces, the number of created VMs per VM type, the VM utilizations for each VM type, and the deadline satisfaction traces. Fig. 7 shows the cost traces for WL #3 and #6, which are two worst cases for the vertical scaling validations. Even for these two worse cases, the cost trace results from closely resemble with the actual measurements. Fig. 8(a) shows the normalized results for the number of created VM for other two workloads (WL #4, #5). The results show that has only 8.3% simulation error in the worst case Norm. VM Num. Norm. VM Util WL#4 Workloads WL#5 (a) Vertical Scaling VM Numbers WL#4 Workloads WL#5 (b) Vertical Scaling VM Utilization M L XL 2XL M L XL 2XL Fig. 8: The VM numbers (a) and utilizations (b) per VM type, normalized over the actual measurement (M: m3.medium, L: m3.large, X: m3.xlarge, XL: m3.xlarge) (m3.medium instance of WL #5). Fig. 8(b) represents the utilization for each VM type. In the worst cases (m3.large of WL #4), has 7.3% simulation error. Fig. 9 gives the job deadline satisfaction rate traces for the two worse cases workloads (WL #4 and #6). Both the and real measurement trace curves in Fig. 9 closely match each other. Based on the results in Fig. 7 to 9, we can conclude that can also accurately simulate the detailed results of the cloud cost, the VM creation, the VM utilization and the deadline satisfaction rate for each VM type and at fine-grained time intervals. For the last validation of the vertical scaling cases, we measured the number of vertical scaling decisions for the four workloads. These results are used to show how accurately simulates the vertical scaling operations. The results are shown in Fig.. For the total number of vertical scaling decisions, has only 3.5% average error. Average error for scaling-up is 6.7% and average error for scaling down decision is 6.3%. also has less than % of simulation errors in every scaling decision for all four workloads. These results imply that can precisely simulate the vertical scaling operations of the cloud-application on real public cloud. For the best of our knowledge, is the first cloud simulator that supports the vertical resource scaling simulation. V. DISCUSSION The previous section demonstrates that accurately simulates the behavior of cloud-application and public IaaS. However, the accuracy of depends on the accuracy of user-provided parameters and configurations. Although users can provide accurate values for most parameters and configurations with ease, one parameter the job execution time may be difficult to acquire precisely. The difficulty comes from the performance uncertainty of real public clouds [9 2]. In our experiments, we used the average execution time from 8 samples (i.e., 8 executions on AWS) for each job type and 8
9 Norm. Job Deadline Satisfaction Rate Norm. Job Deadline Satisfaction Rate Job Sequence (a) WL # Job Sequence (b) WL #6 Fig. 9: Job deadline satisfaction traces for vertical scaling cases Vertical Scaling Simulation Error(%) WL#3 WL#4 WL#5 WL#5 Workloads Total Scale-Up Scale-Down Fig. : Simulation errors for the numbers of vertical scaling decisions VM type. However, in real practice, it may not be feasible for the users to acquire such high number of samples. That is, the users may provide inaccurate job execution time to the simulator. In this section, we analyze the impact of imprecise job execution time on simulation accuracy. More specifically, we seek the answers to the following questions: ) How much error should we expect from the use-provided parameter of job execution time? 2) What is the accuracy of if the parameter of job execution time has certain errors? To answer the first question, we extensively analyzed our job execution samples on AWS. The analysis shows that the majority of the samples have execution times within % of the average execution time (for one job type running on one VM type). On average, 88% of samples have at most % errors. Moreover, the maximum difference between a sample and the average execution time is 22%. These findings corroborate the results from previous work [6]. In short, userprovided job execution times are expected to have less than % error in most cases, while maximum errors not exceeding 22%. TABLE V: Simulation Errors when the job execution time parameter has ±% and ±2% errors. Err. in Params Cost # of VMs VM Util. Job DL. -2% 6.2% 3.5%.9% 5.96% -% 7.5% 6.3%.9% 4.25% +% 4.6% 4.7%.2% 3.3% +2% 3.8%.7%.9% 2.% (a) WL #6 with imprecise parameter (b) WL # with imprecise parameter +2% +% -% -2% +2% +% -% -2% Fig. : Cost traces of simulations with ±% and ±2% of imprecise job execution time parameter. To answer the second question, we simulated the 6 workloads again using job execution times with ±% and ±2% errors. These errors represent the aforementioned expected and maximum errors from user inputs. TABLE V shows the average errors for the 6 workloads with imprecise job execution time parameter in. Fig and Fig 2 show the cost and horizontal scaling traces of selected two workloads with imprecise parameters. (The other workloads have similar results.) These table and figures show that the errors of are considerably smaller than the errors in the job execution time parameter, and retains high accuracy even when user provides imprecise job execution time parameter. We observed that has lower errors than the parameter of job execution time for two reasons: The first reason is that the running times of low-load VMs are less susceptible to input errors. Because of we used large workloads with varied job arrival times, many VMs have only a small number of jobs to execute. These VMs are usually created during periods with low job arriving rates. The running times of these lowload VMs have large fluctuations due to real cloud s unstable performance. Because of this fluctuation, the running time of many low-load VMs is close to the simulation based on the imprecise job execution time. Moreover, the execution time of a low-load VM is also considerably affected by the VM startup time, which further reduces the impact of the imprecise parameter. The second reason is the horizontal scale-in policy of our resource manager. Our resource manager keeps VMs alive for some time in the anticipation of new jobs. Thus, the total running time of a VM is longer than the total execution time of its jobs, which reduces the impact of imprecise parameter of the job execution time. In summary, due to the performance uncertainty in real clouds, user-provided job execution time usually has less than % errors, with a maximum error of 22%. With these potential input errors, the simulator can provide reliable results to help users to assess their cloud application and services. 9
10 Normalized VM Number Normalized VM Number (a) WL #6 with imprecise parameter (b) WL # with imprecise parameter +2% +% -% -2% +2% +% -% -2% Fig. 2: Horizontal VM Scaling Traces of simulations with ±% and ±2% of imprecise job execution time parameter. VI. CONCLUSION In order to answer for potential cloud users questions about evaluating the public clouds without actually deploying the cloud-application, we have created, a public IaaS cloud simulator. provides the capabilities for accurately evaluating the cloud cost, resource usages (VM scaling, utilization, and the number of VMs), and job deadline satisfaction rate. In this paper, we describe the configurations and architecture of, and validate the accuracy of by comparing it with actual measurements from a real cloud-application on real public IaaS cloud. The validation results show that very accurately simulates the behaviors of the cloud-application and public IaaS clouds (with less than 5% of average errors). Moreover, we show the sensitivity of with an imprecise simulation parameter (job execution time with ±% and ±2% errors). The results show still provides very reliable simulation results with the imprecise parameter. These results demonstrated that is both versatile and reliable for cloud user to evaluate the public clouds without actually deploying the cloud-application. In the near future, we will have more comprehensive validations of with three different directions. The first direction is to show the correctness of with cloud-applications on other public cloud services such as Microsoft Azure [2] and Google Compute Engine [3]. Validating on the different cloud providers is particularly important to demonstrate the simulation correctness and generalizability of, because these cloud providers have different performance of VMs and cloud services (e.g. storage and network) as well as different cost models for cloud usages (e.g. minute-based billing model). The second direction is validating with other cloud-applications and resource management policies. We will use n-tier web and scientific/big-data analytics applications with various management configurations because these are the common cloud-application deployment models for industry and research community. Furthermore, we will validate based on the other metrics such as storage and network I/O usage. VII. ACKNOWLEDGMENT The authors would like to thank the anonymous reviewers for their valuable comments and suggestions to improve the quality of the paper. REFERENCES [] Amazon Web Services. [2] Microsoft Azure. [3] Compute Engine - Google Cloud Platform. [4] Apache Hadoop. [5] The Network Simulator ns-2. [6] Project Web Site. [7] RightScale. [8] SCALR. [9] R. Appuswamy and et al. Scale-up vs Scale-out for Hadoop: Time to rethink? In ACM Sym. on Cloud Computing (SoCC), 23. [] M. Bux and U. Leser. DynamicCloudSim: Simulating Heterogeneity in Computational Clouds. In ACM SIGMOD Workshop on Scalable Workflow Execution Engines and Technologies (SWEET), 23. [] R. N. Calheiros and et al. Cloudsim: A toolkit for modeling and simulation of cloud computing environments and evaluation of resource provisioning algorithm. Software: Practice and Experience, 4():23 5, 2. [2] W. Chen and E. Deelman. WorkflowSim: A Toolkit for Simulating Scientific Workflows in Distributed Environments. In IEEE Int l Conf. on escience (escience), 22. [3] J. Dean and S. Ghemawat. MapReduce: Simplified Data Processing on Large Clusters. In USENIX Sym. on Operating Systems Design & Implementation (OSDI), 24. [4] S. K. Garg and R. Buyya. NetworkCloudSim: Modelling Parallel Applications in Cloud Simulations. In IEEE Int l. Conf. on Utility and Cloud Computing (UCC), 2. [5] D. Kliazovich and et al. GreenCloud: a packet-level simulator of energy-aware cloud computing data centers. Journal of Supercomputing, 62(3): , 22. [6] P. Leitner and J. Cito. Patterns in the Chaos a Study of Performance Variation and Predictability in Public Iaas Clouds. CoRR, abs/4.2429, 24. [7] M. Mao and M. Humphrey. A Performance Study on the VM Startup Time in the Cloud. In IEEE international Conference on Cloud Computing (CLOUD), 22. [8] A. Nunez and et al. icancloud: A Flexible and Scalable Cloud Infrastructure Simulator. Journal of Grid Computing, (): 85 29, 22. [9] Z. Ou and et al. Exploiting Hardware Heterogeneity within the Same Instance Type of Amazon EC2. In USENIX Workshop on Hop Topics in Cloud Computing (HotCloud), 22. [2] J. Schad and et al. Runtime Measurements in the Cloud: Observing, Analyzing, and Reducing Variance. Proc. of the VLDB Endowment, 3( 2):46 47, 2. [2] M. Schwarzkopf and et al. The Seven Deadly Sins of Cloud Computing Research. In USENIX Workshop on Hop Topics in Cloud Computing (HotCloud), 22. [22] B. Wickremasinghe and et al. CloudAnalyst: A Cloudsim-based Visual Modeller for Analysing Cloud Computing Environments and Applications. In IEEE Int l. Conf. on Advanced Information Networking and Applications (AINA), 2.
Simulation-based Evaluation of an Intercloud Service Broker
Simulation-based Evaluation of an Intercloud Service Broker Foued Jrad, Jie Tao and Achim Streit Steinbuch Centre for Computing, SCC Karlsruhe Institute of Technology, KIT Karlsruhe, Germany {foued.jrad,
Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing
Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing Deep Mann ME (Software Engineering) Computer Science and Engineering Department Thapar University Patiala-147004
DynamicCloudSim: Simulating Heterogeneity in Computational Clouds
DynamicCloudSim: Simulating Heterogeneity in Computational Clouds Marc Bux, Ulf Leser {bux leser}@informatik.hu-berlin.de The 2nd international workshop on Scalable Workflow Enactment Engines and Technologies
Efficient Service Broker Policy For Large-Scale Cloud Environments
www.ijcsi.org 85 Efficient Service Broker Policy For Large-Scale Cloud Environments Mohammed Radi Computer Science Department, Faculty of Applied Science Alaqsa University, Gaza Palestine Abstract Algorithms,
How to Do/Evaluate Cloud Computing Research. Young Choon Lee
How to Do/Evaluate Cloud Computing Research Young Choon Lee Cloud Computing Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing
Optimal Service Pricing for a Cloud Cache
Optimal Service Pricing for a Cloud Cache K.SRAVANTHI Department of Computer Science & Engineering (M.Tech.) Sindura College of Engineering and Technology Ramagundam,Telangana G.LAKSHMI Asst. Professor,
NetworkCloudSim: Modelling Parallel Applications in Cloud Simulations
2011 Fourth IEEE International Conference on Utility and Cloud Computing NetworkCloudSim: Modelling Parallel Applications in Cloud Simulations Saurabh Kumar Garg and Rajkumar Buyya Cloud Computing and
2) Xen Hypervisor 3) UEC
5. Implementation Implementation of the trust model requires first preparing a test bed. It is a cloud computing environment that is required as the first step towards the implementation. Various tools
Exploring Resource Provisioning Cost Models in Cloud Computing
Exploring Resource Provisioning Cost Models in Cloud Computing P.Aradhya #1, K.Shivaranjani *2 #1 M.Tech, CSE, SR Engineering College, Warangal, Andhra Pradesh, India # Assistant Professor, Department
Payment minimization and Error-tolerant Resource Allocation for Cloud System Using equally spread current execution load
Payment minimization and Error-tolerant Resource Allocation for Cloud System Using equally spread current execution load Pooja.B. Jewargi Prof. Jyoti.Patil Department of computer science and engineering,
A Secure Strategy using Weighted Active Monitoring Load Balancing Algorithm for Maintaining Privacy in Multi-Cloud Environments
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X A Secure Strategy using Weighted Active Monitoring Load Balancing Algorithm for Maintaining
PERFORMANCE ANALYSIS OF PaaS CLOUD COMPUTING SYSTEM
PERFORMANCE ANALYSIS OF PaaS CLOUD COMPUTING SYSTEM Akmal Basha 1 Krishna Sagar 2 1 PG Student,Department of Computer Science and Engineering, Madanapalle Institute of Technology & Science, India. 2 Associate
SLA-based Admission Control for a Software-as-a-Service Provider in Cloud Computing Environments
SLA-based Admission Control for a Software-as-a-Service Provider in Cloud Computing Environments Linlin Wu, Saurabh Kumar Garg, and Rajkumar Buyya Cloud Computing and Distributed Systems (CLOUDS) Laboratory
Mobile and Cloud computing and SE
Mobile and Cloud computing and SE This week normal. Next week is the final week of the course Wed 12-14 Essay presentation and final feedback Kylmämaa Kerkelä Barthas Gratzl Reijonen??? Thu 08-10 Group
Federation of Cloud Computing Infrastructure
IJSTE International Journal of Science Technology & Engineering Vol. 1, Issue 1, July 2014 ISSN(online): 2349 784X Federation of Cloud Computing Infrastructure Riddhi Solani Kavita Singh Rathore B. Tech.
Sla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing
Sla Aware Load Balancing Using Join-Idle Queue for Virtual Machines in Cloud Computing Mehak Choudhary M.Tech Student [CSE], Dept. of CSE, SKIET, Kurukshetra University, Haryana, India ABSTRACT: Cloud
OCRP Implementation to Optimize Resource Provisioning Cost in Cloud Computing
OCRP Implementation to Optimize Resource Provisioning Cost in Cloud Computing K. Satheeshkumar PG Scholar K. Senthilkumar PG Scholar A. Selvakumar Assistant Professor Abstract- Cloud computing is a large-scale
Profit-driven Cloud Service Request Scheduling Under SLA Constraints
Journal of Information & Computational Science 9: 14 (2012) 4065 4073 Available at http://www.joics.com Profit-driven Cloud Service Request Scheduling Under SLA Constraints Zhipiao Liu, Qibo Sun, Shangguang
Environments, Services and Network Management for Green Clouds
Environments, Services and Network Management for Green Clouds Carlos Becker Westphall Networks and Management Laboratory Federal University of Santa Catarina MARCH 3RD, REUNION ISLAND IARIA GLOBENET 2012
Cloud computing - Architecting in the cloud
Cloud computing - Architecting in the cloud [email protected] 1 Outline Cloud computing What is? Levels of cloud computing: IaaS, PaaS, SaaS Moving to the cloud? Architecting in the cloud Best practices
DataCenter optimization for Cloud Computing
DataCenter optimization for Cloud Computing Benjamín Barán National University of Asuncion (UNA) [email protected] Paraguay Content Cloud Computing Commercial Offerings Basic Problem Formulation Open Research
A Proposed Service Broker Strategy in CloudAnalyst for Cost-Effective Data Center Selection
A Proposed Service Broker Strategy in CloudAnalyst for Cost-Effective Selection Dhaval Limbani*, Bhavesh Oza** *(Department of Information Technology, S. S. Engineering College, Bhavnagar) ** (Department
OTM in the Cloud. Ryan Haney
OTM in the Cloud Ryan Haney The Cloud The Cloud is a set of services and technologies that delivers real-time and ondemand computing resources Software as a Service (SaaS) delivers preconfigured applications,
Cloud Computing Simulation Using CloudSim
Cloud Computing Simulation Using CloudSim Ranjan Kumar #1, G.Sahoo *2 # Assistant Professor, Computer Science & Engineering, Ranchi University, India Professor & Head, Information Technology, Birla Institute
Cloud deployment model and cost analysis in Multicloud
IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) ISSN: 2278-2834, ISBN: 2278-8735. Volume 4, Issue 3 (Nov-Dec. 2012), PP 25-31 Cloud deployment model and cost analysis in Multicloud
Cloud Computing Architectures and Design Issues
Cloud Computing Architectures and Design Issues Ozalp Babaoglu, Stefano Ferretti, Moreno Marzolla, Fabio Panzieri {babaoglu, sferrett, marzolla, panzieri}@cs.unibo.it Outline What is Cloud Computing? A
SLA BASED SERVICE BROKERING IN INTERCLOUD ENVIRONMENTS
SLA BASED SERVICE BROKERING IN INTERCLOUD ENVIRONMENTS Foued Jrad, Jie Tao and Achim Streit Steinbuch Centre for Computing, Karlsruhe Institute of Technology, Karlsruhe, Germany {foued.jrad, jie.tao, achim.streit}@kit.edu
Multilevel Communication Aware Approach for Load Balancing
Multilevel Communication Aware Approach for Load Balancing 1 Dipti Patel, 2 Ashil Patel Department of Information Technology, L.D. College of Engineering, Gujarat Technological University, Ahmedabad 1
IAAS CLOUD EXCHANGE WHITEPAPER
IAAS CLOUD EXCHANGE WHITEPAPER Whitepaper, July 2013 TABLE OF CONTENTS Abstract... 2 Introduction... 2 Challenges... 2 Decoupled architecture... 3 Support for different consumer business models... 3 Support
Resource Allocation Avoiding SLA Violations in Cloud Framework for SaaS
Resource Allocation Avoiding SLA Violations in Cloud Framework for SaaS Shantanu Sasane Abhilash Bari Kaustubh Memane Aniket Pathak Prof. A. A.Deshmukh University of Pune University of Pune University
Performance Gathering and Implementing Portability on Cloud Storage Data
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 17 (2014), pp. 1815-1823 International Research Publications House http://www. irphouse.com Performance Gathering
Performance Management for Cloudbased STC 2012
Performance Management for Cloudbased Applications STC 2012 1 Agenda Context Problem Statement Cloud Architecture Need for Performance in Cloud Performance Challenges in Cloud Generic IaaS / PaaS / SaaS
Keywords: Cloudsim, MIPS, Gridlet, Virtual machine, Data center, Simulation, SaaS, PaaS, IaaS, VM. Introduction
Vol. 3 Issue 1, January-2014, pp: (1-5), Impact Factor: 1.252, Available online at: www.erpublications.com Performance evaluation of cloud application with constant data center configuration and variable
Figure 1. The cloud scales: Amazon EC2 growth [2].
- Chung-Cheng Li and Kuochen Wang Department of Computer Science National Chiao Tung University Hsinchu, Taiwan 300 [email protected], [email protected] Abstract One of the most important issues
CloudSim: A Toolkit for Modeling and Simulation of Cloud Computing Environments and Evaluation of Resource Provisioning Algorithms
CloudSim: A Toolkit for Modeling and Simulation of Cloud Computing Environments and Evaluation of Resource Provisioning Algorithms Rodrigo N. Calheiros, Rajiv Ranjan, Anton Beloglazov, César A. F. De Rose,
A Generic Auto-Provisioning Framework for Cloud Databases
A Generic Auto-Provisioning Framework for Cloud Databases Jennie Rogers 1, Olga Papaemmanouil 2 and Ugur Cetintemel 1 1 Brown University, 2 Brandeis University Instance Type Introduction Use Infrastructure-as-a-Service
Building Platform as a Service for Scientific Applications
Building Platform as a Service for Scientific Applications Moustafa AbdelBaky [email protected] Rutgers Discovery Informa=cs Ins=tute (RDI 2 ) The NSF Cloud and Autonomic Compu=ng Center Department
Enhancing Dataset Processing in Hadoop YARN Performance for Big Data Applications
Enhancing Dataset Processing in Hadoop YARN Performance for Big Data Applications Ahmed Abdulhakim Al-Absi, Dae-Ki Kang and Myong-Jong Kim Abstract In Hadoop MapReduce distributed file system, as the input
CLOUD PERFORMANCE TESTING - KEY CONSIDERATIONS (COMPLETE ANALYSIS USING RETAIL APPLICATION TEST DATA)
CLOUD PERFORMANCE TESTING - KEY CONSIDERATIONS (COMPLETE ANALYSIS USING RETAIL APPLICATION TEST DATA) Abhijeet Padwal Performance engineering group Persistent Systems, Pune email: [email protected]
MyCloudLab: An Interactive Web-based Management System for Cloud Computing Administration
MyCloudLab: An Interactive Web-based Management System for Cloud Computing Administration Hoi-Wan Chan 1, Min Xu 2, Chung-Pan Tang 1, Patrick P. C. Lee 1 & Tsz-Yeung Wong 1, 1 Department of Computer Science
Network Infrastructure Services CS848 Project
Quality of Service Guarantees for Cloud Services CS848 Project presentation by Alexey Karyakin David R. Cheriton School of Computer Science University of Waterloo March 2010 Outline 1. Performance of cloud
Auto-Scaling Model for Cloud Computing System
Auto-Scaling Model for Cloud Computing System Che-Lun Hung 1*, Yu-Chen Hu 2 and Kuan-Ching Li 3 1 Dept. of Computer Science & Communication Engineering, Providence University 2 Dept. of Computer Science
International Journal of Scientific & Engineering Research, Volume 6, Issue 4, April-2015 36 ISSN 2229-5518
International Journal of Scientific & Engineering Research, Volume 6, Issue 4, April-2015 36 An Efficient Approach for Load Balancing in Cloud Environment Balasundaram Ananthakrishnan Abstract Cloud computing
Virtual Machine Instance Scheduling in IaaS Clouds
Virtual Machine Instance Scheduling in IaaS Clouds Naylor G. Bachiega, Henrique P. Martins, Roberta Spolon, Marcos A. Cavenaghi Departamento de Ciência da Computação UNESP - Univ Estadual Paulista Bauru,
Reallocation and Allocation of Virtual Machines in Cloud Computing Manan D. Shah a, *, Harshad B. Prajapati b
Proceedings of International Conference on Emerging Research in Computing, Information, Communication and Applications (ERCICA-14) Reallocation and Allocation of Virtual Machines in Cloud Computing Manan
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Hsin-Wen Wei 1,2, Che-Wei Hsu 2, Tin-Yu Wu 3, Wei-Tsong Lee 1 1 Department of Electrical Engineering, Tamkang University
Cloud Computing Paradigm
Cloud Computing Paradigm Julio Guijarro Automated Infrastructure Lab HP Labs Bristol, UK 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice
CLOUD computing has emerged as a new paradigm for
IEEE TRANSACTIONS ON SERVICES COMPUTING, VOL. 7, NO. 3, JULY-SEPTEMBER 2014 465 SLA-Based Resource Provisioning for Hosted Software-as-a-Service Applications in Cloud Computing Environments Linlin Wu,
Performance Analysis: Benchmarking Public Clouds
Performance Analysis: Benchmarking Public Clouds Performance comparison of web server and database VMs on Internap AgileCLOUD and Amazon Web Services By Cloud Spectator March 215 PERFORMANCE REPORT WEB
High performance computing network for cloud environment using simulators
High performance computing network for cloud environment using simulators Ajith Singh. N 1 and M. Hemalatha 2 1 Ph.D, Research Scholar (CS), Karpagam University, Coimbatore, India 2 Prof & Head, Department
Optimizing the Cost for Resource Subscription Policy in IaaS Cloud
Optimizing the Cost for Resource Subscription Policy in IaaS Cloud Ms.M.Uthaya Banu #1, Mr.K.Saravanan *2 # Student, * Assistant Professor Department of Computer Science and Engineering Regional Centre
Hadoop in the Hybrid Cloud
Presented by Hortonworks and Microsoft Introduction An increasing number of enterprises are either currently using or are planning to use cloud deployment models to expand their IT infrastructure. Big
Automating Big Data Benchmarking for Different Architectures with ALOJA
www.bsc.es Jan 2016 Automating Big Data Benchmarking for Different Architectures with ALOJA Nicolas Poggi, Postdoc Researcher Agenda 1. Intro on Hadoop performance 1. Current scenario and problematic 2.
21/09/11. Introduction to Cloud Computing. First: do not be scared! Request for contributors. ToDO list. Revision history
Request for contributors Introduction to Cloud Computing https://portal.futuregrid.org/contrib/cloud-computing-class by various contributors (see last slide) Hi and thanks for your contribution! If you
Grid Computing Vs. Cloud Computing
International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 6 (2013), pp. 577-582 International Research Publications House http://www. irphouse.com /ijict.htm Grid
Dynamic Round Robin for Load Balancing in a Cloud Computing
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 2, Issue. 6, June 2013, pg.274
International Journal of Digital Application & Contemporary research Website: www.ijdacr.com (Volume 2, Issue 9, April 2014)
Green Cloud Computing: Greedy Algorithms for Virtual Machines Migration and Consolidation to Optimize Energy Consumption in a Data Center Rasoul Beik Islamic Azad University Khomeinishahr Branch, Isfahan,
A Study on Analysis and Implementation of a Cloud Computing Framework for Multimedia Convergence Services
A Study on Analysis and Implementation of a Cloud Computing Framework for Multimedia Convergence Services Ronnie D. Caytiles and Byungjoo Park * Department of Multimedia Engineering, Hannam University
Heterogeneity-Aware Resource Allocation and Scheduling in the Cloud
Heterogeneity-Aware Resource Allocation and Scheduling in the Cloud Gunho Lee, Byung-Gon Chun, Randy H. Katz University of California, Berkeley, Yahoo! Research Abstract Data analytics are key applications
Evaluation Methodology of Converged Cloud Environments
Krzysztof Zieliński Marcin Jarząb Sławomir Zieliński Karol Grzegorczyk Maciej Malawski Mariusz Zyśk Evaluation Methodology of Converged Cloud Environments Cloud Computing Cloud Computing enables convenient,
INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY
Karthi M,, 2013; Volume 1(8):1062-1072 INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK EFFICIENT MANAGEMENT OF RESOURCES PROVISIONING
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads
Advanced Load Balancing Mechanism on Mixed Batch and Transactional Workloads G. Suganthi (Member, IEEE), K. N. Vimal Shankar, Department of Computer Science and Engineering, V.S.B. Engineering College,
Keywords Distributed Computing, On Demand Resources, Cloud Computing, Virtualization, Server Consolidation, Load Balancing
Volume 5, Issue 1, January 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Survey on Load
Comparison of PBRR Scheduling Algorithm with Round Robin and Heuristic Priority Scheduling Algorithm in Virtual Cloud Environment
www.ijcsi.org 99 Comparison of PBRR Scheduling Algorithm with Round Robin and Heuristic Priority Scheduling Algorithm in Cloud Environment Er. Navreet Singh 1 1 Asst. Professor, Computer Science Department
Assignment # 1 (Cloud Computing Security)
Assignment # 1 (Cloud Computing Security) Group Members: Abdullah Abid Zeeshan Qaiser M. Umar Hayat Table of Contents Windows Azure Introduction... 4 Windows Azure Services... 4 1. Compute... 4 a) Virtual
By Cloud Spectator July 2013
Benchmark Report: Performance Analysis of VS. Amazon EC2 and Rackspace Cloud A standardized, side-by-side comparison of server performance, file IO, and internal network throughput. By Cloud Spectator
Comparing major cloud-service providers: virtual processor performance. A Cloud Report by Danny Gee, and Kenny Li
Comparing major cloud-service providers: virtual processor performance A Cloud Report by Danny Gee, and Kenny Li Comparing major cloud-service providers: virtual processor performance 09/03/2014 Table
CloudAnalyst: A CloudSim-based Visual Modeller for Analysing Cloud Computing Environments and Applications
CloudAnalyst: A CloudSim-based Visual Modeller for Analysing Cloud Computing Environments and Applications Bhathiya Wickremasinghe 1, Rodrigo N. Calheiros 2, and Rajkumar Buyya 1 1 The Cloud Computing
Permanent Link: http://espace.library.curtin.edu.au/r?func=dbin-jump-full&local_base=gen01-era02&object_id=154091
Citation: Alhamad, Mohammed and Dillon, Tharam S. and Wu, Chen and Chang, Elizabeth. 2010. Response time for cloud computing providers, in Kotsis, G. and Taniar, D. and Pardede, E. and Saleh, I. and Khalil,
Amazon Web Services vs. Horizon
Amazon Web Services vs. February 2016 Comparing Cloud Services Every Cloud Service Provider (CSP) and every Cloud Service is different. Some CSPs focus on being price competitive and providing self-service,
CLOUD COMPUTING An Overview
CLOUD COMPUTING An Overview Abstract Resource sharing in a pure plug and play model that dramatically simplifies infrastructure planning is the promise of cloud computing. The two key advantages of this
Benchmarking Amazon s EC2 Cloud Platform
Benchmarking Amazon s EC2 Cloud Platform Goal of the project: The goal of this project is to analyze the performance of an 8-node cluster in Amazon s Elastic Compute Cloud (EC2). The cluster will be benchmarked
Object Storage: A Growing Opportunity for Service Providers. White Paper. Prepared for: 2012 Neovise, LLC. All Rights Reserved.
Object Storage: A Growing Opportunity for Service Providers Prepared for: White Paper 2012 Neovise, LLC. All Rights Reserved. Introduction For service providers, the rise of cloud computing is both a threat
Exploiting Cloud Heterogeneity for Optimized Cost/Performance MapReduce Processing
Exploiting Cloud Heterogeneity for Optimized Cost/Performance MapReduce Processing Zhuoyao Zhang University of Pennsylvania, USA [email protected] Ludmila Cherkasova Hewlett-Packard Labs, USA [email protected]
Dynamic Resource management with VM layer and Resource prediction algorithms in Cloud Architecture
Dynamic Resource management with VM layer and Resource prediction algorithms in Cloud Architecture 1 Shaik Fayaz, 2 Dr.V.N.Srinivasu, 3 Tata Venkateswarlu #1 M.Tech (CSE) from P.N.C & Vijai Institute of
SLA-aware Resource Scheduling for Cloud Storage
SLA-aware Resource Scheduling for Cloud Storage Zhihao Yao Computer and Information Technology Purdue University West Lafayette, Indiana 47906 Email: [email protected] Ioannis Papapanagiotou Computer and
Evaluating HDFS I/O Performance on Virtualized Systems
Evaluating HDFS I/O Performance on Virtualized Systems Xin Tang [email protected] University of Wisconsin-Madison Department of Computer Sciences Abstract Hadoop as a Service (HaaS) has received increasing
Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture
Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture He Huang, Shanshan Li, Xiaodong Yi, Feng Zhang, Xiangke Liao and Pan Dong School of Computer Science National
BlobSeer: Towards efficient data storage management on large-scale, distributed systems
: Towards efficient data storage management on large-scale, distributed systems Bogdan Nicolae University of Rennes 1, France KerData Team, INRIA Rennes Bretagne-Atlantique PhD Advisors: Gabriel Antoniu
Ø Teaching Evaluations. q Open March 3 through 16. Ø Final Exam. q Thursday, March 19, 4-7PM. Ø 2 flavors: q Public Cloud, available to public
Announcements TIM 50 Teaching Evaluations Open March 3 through 16 Final Exam Thursday, March 19, 4-7PM Lecture 19 20 March 12, 2015 Cloud Computing Cloud Computing: refers to both applications delivered
Deadline Based Task Scheduling in Cloud with Effective Provisioning Cost using LBMMC Algorithm
Deadline Based Task Scheduling in Cloud with Effective Provisioning Cost using LBMMC Algorithm Ms.K.Sathya, M.E., (CSE), Jay Shriram Group of Institutions, Tirupur [email protected] Dr.S.Rajalakshmi,
CLOUD SIMULATORS: A REVIEW
CLOUD SIMULATORS: A REVIEW 1 Rahul Singh, 2 Punyaban Patel, 3 Preeti Singh Chhatrapati Shivaji Institute of Technology, Durg, India Email: 1 [email protected], 2 [email protected],
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis Felipe Augusto Nunes de Oliveira - GRR20112021 João Victor Tozatti Risso - GRR20120726 Abstract. The increasing
Accelerating Time to Market:
Accelerating Time to Market: Application Development and Test in the Cloud Paul Speciale, Savvis Symphony Product Marketing June 2010 HOS-20100608-GL-Accelerating-Time-to-Market-Dev-Test-Cloud 1 Software
