Martin Bichler, Benjamin Speitkamp
|
|
|
- Rudolf Cornelius Sherman
- 10 years ago
- Views:
Transcription
1 Martin Bichler, Benjamin Speitkamp TU München, Munich, Germany Today's data centers offer many different IT services mostly hosted on dedicated physical servers. Server virtualization provides a new technical means for server consolidation. Thus, multiple virtual servers including operating systems and applications can be reliably hosted on a single physical server sharing the same resources. Server consolidation describes the process of combining the workloads of several different servers on a set of target servers. In this paper we focus on large-scale server consolidation problems as they can be regularly found in data centers. Cost considerations, specifically energy savings, are among the key drivers for such projects. First, this article presents different decision models to optimally allocate virtual servers. Important real-world constraints with server consolidation are presented and accounted for in our models. Our central model is proven to be an NP-hard problem and, therefore, a heuristic is presented in addition to exact solution methods to address large-scale server consolidation projects. Second, adjusted to the consolidation model, a preprocessing method for server load data is introduced allowing the determination of quality of service levels for virtual servers. Third, extensive experiments were conducted based on a large set of server load data from a data center provider focusing on managerial concerns over what types of problems can be solved. The results show that, on average, 31 percent server savings can be achieved only by taking cycles in the server workload into account in the optimization. Key words: server virtualization; server consolidation; combinatorial optimization History: This paper was first submitted on February 26, Acknowledgement This work was accomplished in collaboration with Siemens IT Solutions and Services (SIS), one of the largest IT Service Providers in Europe. We thank SIS for their technical and financial support. Patents have been filed for the data preprocessing and the optimization methods described in the paper.
2 Author: Server Consolidation 2 1 Introduction Enterprises nowadays host central servers either in internal data centers or in those of commercial IT service providers. These data centers host most of the essential IT services (e.g., CRM systems, ERP modules, databases, or Web servers) on dedicated physical servers. The complex resource requirements of enterprise services and the desire to provision for peak demand are important reasons for over-provisioning. As a consequence, the average server utilization is typically very low, which incurs high investment and operational costs. The Gartner Group estimates that the utilization of servers in data centers is less than 20 percent [1]. However, efficiency in the production of IT services is key in an industry where services are becoming more and more standardized and where revenues are therefore decreasing. In particular, the bare savings of servers is of primary importance for most data centers regarding the limitation of energy and cooling, which sometimes account for % of the total data center operation costs [2] (see also Section 7). Virtualization provides technical means for server consolidation leading to increased server utilization and allows much faster adaptation of the IT infrastructure to changing demand. The term refers to the abstraction of computing resources across many aspects of computing and has been used to describe different techniques. Server virtualization describes a virtual machine which appears to a "guest" operating system as hardware, but is simulated in a contained software environment by the host system. This way, a single server can be partitioned into multiple virtual servers. Parallels, Xen, and VMware are only a few of the products on the market enabling server virtualization. Apart from higher server utilization levels, benefits are: reduced time for deployment, easier system management, and overall lower hardware and operating costs. Virtualization has been a growing trend in the past few years, and it can now be considered an
3 Author: Server Consolidation 3 established tool that is nowadays used regularly in large-scale server-consolidation projects with IT service providers [3]. Among many of the economic and organizational issues, Information Systems research addresses issues regarding the development and deployment of information technology (IT). Capacity planning, quality-of-service, and performance modeling have long been key managerial tasks in Information Systems [4-6]. Increased outsourcing of IT services [7], as well as on-demand software provisioning for thousands or millions of users, have led to automation and further industrialization of data center operations. Production efficiency has become crucial for IT service management and capacity management is therefore a key component of related standards such as ITIL [8] or ISO/IEC [9]. Target areas of capacity planning and performance modeling in IS include file and database systems, computer networks, operating systems, fault-tolerant systems, and real-time systems [10-12]. A traditional analytical approach to support capacity planning is queuing theory, allowing for the determination of response time, service time, server utilization and many other metrics essential for capacity planning with dedicated servers [13]. Server consolidation is orthogonal to this type of capacity planning. Mostly the workload and system requirements for different services are known or can be estimated with a high degree of confidence from historical workload data. The challenge is finding an allocation of source servers (services in the following) to target servers that minimizes costs, considering quality of service (QoS) requirements. This poses new and widespread capacity planning problems that will be the focus of this work 1. 1 Storage virtualization and storage consolidation related problems are not considered in this paper. Server consolidation problems can be treated separately from the storage issue.
4 Author: Server Consolidation 4 In this paper, we consider the problem of IT service providers hosting the IT services of multiple (also internal) customers. Based on the users demands or historical workloads, the IT service provider needs to determine the allocation of these services to target servers using some predefined objective function (e.g., minimizing the amount or costs of servers used). We propose a server consolidation method that combines data analysis to characterize variations in workload traces and algorithms to optimally assign source servers to physical target servers. These types of problems are computationally hard, and can be found in almost all server consolidation projects. One contribution of this paper is decision support for IT service providers following a design science approach [14]: We provide a mathematical formulation of widespread server consolidation problems, a complexity analysis showing that the fundamental problem is an NP-hard optimization problem, an LP-relaxation based heuristic to solve this problem, and a data preprocessing method characterizing workload traces and deriving parameters for the decision model. The data preprocessing method allows the exploitation of seasonalities in the decision problem on the one hand and offers the possibility to adjust the quality of each service individually on the other. The main contribution of this paper, however, is an extensive experimental evaluation based on real-world workload traces that focuses on the managerial questions for IT service providers. Given the specifics of the server consolidation problem, it is important for managers to understand which problem sizes can be solved exactly, which can be solved heuristically, and what the impact of various model parameters is on the solution quality and time. These questions need to be answered for the server consolidation problem specifically, and are beyond the type of analysis typically provided in the optimization literature. Both our IP-based decision models and our LPbased heuristic allow for the consideration of business constraints, but at the same time solve even large-scale server consolidation problems with hundreds of servers as they can be found in practice
5 Author: Server Consolidation 5 in less than twenty minutes. We provide extensive sensitivity analyses describing how the time resolution of load data considered, or the size of the source servers relative to the target servers, impacts the size of the problems that can be solved. We also show how the risk parameters and the time resolution of load data considered impacts the quality of the allocation. A key result is that only leveraging the daily seasonalities in the optimization accounts for 31 % savings on average in the number of target servers needed as compared to the optimal solution ignoring daily variations in the workload. We also show that there are significant differences between two types of applications with respect to the size of instances that could be solved. While our work follows a design science approach, it does have managerial implications that go beyond the fact that service managers can now perform server consolidation in an automated way. Server consolidation has been embraced as a possibility to cut costs in the provisioning of IT. How much savings one can expect also depends on the characteristics of the workloads at hand. Our models allow the calculation of the potential savings in terms of hardware cost for a specific server consolidation problem at hand. Clearly, virtualization does provide a number of other benefits as well, such as easy management and migration of servers to other hardware. In this paper, however, we focus on savings due to the optimal allocation of virtual to physical servers. The paper is structured as follows: In Section 2 we will provide a brief overview of virtualization techniques. In Section 3 typical server consolidation scenarios found in practice are presented. Then, in Section 4, two fundamental capacity planning problems are introduced and discussed regarding their computational complexity. These models are supplemented by additional business constraints in Section 4.2. Section 5 describes the experimental setup, while Section 6 discusses the results of computational experiments based on real-world data sets. In Section 7 further managerial implications are presented. Section 8 provides an overview of related work, and in Section 9 we draw conclusions.
6 Author: Server Consolidation 6 2 Virtualization Virtualization is performed on a given hardware platform by a control program, which creates a simulated virtual machine for its "guest" software. The "guest" software, which is often itself a complete operating system plus applications, runs as if it was installed on a dedicated server. Typically, many such virtual machines are simulated on a given physical machine. Virtualization is an umbrella term for many techniques. For example, SMP (symmetric multiprocessor) servers can be subdivided into fractions, each of which is a complete server and able to run an operating system. This is often described as physical or logical hardware partitioning. Example products are HP npar, or IBM DLPAR. Software virtualization includes approaches on or below the operating system level, or on the application level. So-called hypervisor software creates virtual servers whose physical resource use can be adjusted dynamically, enabling multiple isolated and secure virtualized servers on a single physical server. The market for software virtualization products is growing and a description of products and techniques is beyond the scope of this article. A typical hypervisor allows the creation and execution of multiple virtual servers simultaneously. Each virtual server instance can execute a guest operating system, such as Windows or Linux. As such, a hypervisor is software that allows one physical server to run numerous operating systems plus applications simultaneously. These products bridge calls to network adapters, CD-ROM readers, hard disk drives, and USB devices. Some products even support migration of virtual servers across multiple host machines at runtime. Administrators can monitor the run-time data of logical and physical software landscapes, start, stop, or relocate servers, and assign hardware resources to application services automatically or manually. There are a number of use cases demonstrating how virtualization can be used in a data center [15]. These new technical possibilities require new methods for capacity management.
7 Author: Server Consolidation 7 Virtualization causes overhead in terms of additional CPU cycles necessary for the hypervisor, every virtual machine, and also a workload-dependent component. In the subsequent decision models, we will consider overhead in a data preprocessing step with simple additive or multiplicative factors. Virtualization overhead will be further reduced with current hardware developments which support certain virtualization features. Virtualization has benefits that are beyond savings in investment and energy costs through consolidation. For example, managing and migrating software from one server to another becomes much easier. However, virtualization software also comes at a cost. Whether the benefits outweigh the (actual) cost is a topic frequently discussed in reports by market research companies. It heavily depends on the cost of introducing virtualization software and the technical progress in the field. 3 Server Consolidation Server consolidation is an approach for the efficient usage of computer server resources in order to reduce operating costs. These costs mainly originate from space- and energy-consumption (for servers and data center cooling), data center maintenance, server administration, and purchasing costs for servers. Consolidation is used in situations in which multiple, under-utilized servers take up more space and consume more resources than can be justified by their workload. We refer to server consolidation as the process of combining the workloads of several different (source) servers or services and assigning them to a set of target servers. 3.1 Managerial Decision Problems In the following we present three widespread scenarios in server consolidation tasks which our decision models apply to. The objective of the decision models is to minimize the sum of server costs, which might be purchasing, maintenance, administration or the sum of them. Depending on the scenario, the models provide
8 Author: Server Consolidation 8 the information on how many and which servers are required for the given set of services (i.e., virtual servers), given a set of potential (hardware) servers (all scenarios), an optimal allocation of services to servers with respect to the objective function (all scenarios), and possibly additional support for decisions that aim to minimize investment and/or operational costs (first and second scenario). The first managerial decision scenario refers to an investment decision, i.e. the data center operator wants to switch to a new technology of hardware, such as a new generation of multi-core processors, and therefore has to decide on how many machines to buy. There may be alternative server types he wants to evaluate against each other. In this first scenario, coefficient represents the purchasing costs of single server i from a set of potential servers I (with ). Alternatively, may represent the total cost of ownership for each single server i, including energy and administrative costs, etc. In the case of alternative server types to choose from, respective models may be configured, with each model offering servers of one specific type. The objective values of the optimization models (i.e., the total cost of the servers) can help to decide on a configuration. Of course, the different server types can also be included in a single model. Sometimes, additional strategic considerations, such as reusability of hardware with future services, customers, technologies, and planning periods do play a role. In these situations, the decision model still serves as a decision support tool complementing the above considerations. The second decision scenario refers to the case where servers have already been purchased and purchasing costs are considered as sunk costs. Nevertheless, the use of existing servers is required to be as efficient as possible, i.e. the operational costs should be kept to a minimum. Note that the costs of electrical power supply for servers and cooling accounts for a large part of
9 Author: Server Consolidation 9 operational costs in data centers (see Section 7). Therefore, in this second scenario, the costs represent the operational costs per server, e.g. comprising energy, cooling, and administrative costs. As a special case, a third decision scenario originates from the a priori assumption of identical servers in terms of both costs and capacities, as is the case with a rack of identical blade servers. Here, the costs may just represent the information concerning which subset of servers should be used if not all servers are required. This may be of particular importance when optional technical allocation constraints have to be considered. Here, the decision models returns no cost value but, for instance, the minimum number of servers required and an allocation that satisfies additional technical constraints. 3.2 Available Data in Data Centers Data centers reserve certain amounts of IT resources for each single service or customer (such as CPU in SAPS 2 or HP Computons, memory in Gigabyte, bandwidth in Megabits per second). Data centers typically log the workloads of their servers over time. These include data about CPU, memory, and bandwidth utilization based on measurements every five minutes or every hour, for example. Usually, resource demand of this sort has seasonal patterns on a daily, weekly, or monthly basis. For example, payroll accounting is performed at the end of the week, while demand for an OLAP application has a daily peak in the morning hours, when managers access their daily reports. Consolidation can leverage these cycles and attempts to assign those services on a physical server whose demand is negatively correlated, i.e. workload peaks at different times of the day or week (see Figure 1 for a real-world example of two services). An analysis of the large set of workload traces that we have received from our industry partner can be found in Appendix B. 2 SAP Application Performance Standard (SAPS) is a wide-spread hardware-independent unit that describes the performance of a system configuration in the SAP environment. It is derived from the Sales and Distribution Benchmark, where 100 SAPS is defined as 2,000 fully business processed order-line items per hour. SAPS are also
10 Author: Server Consolidation 10 Although we did find different daily and also weekly cycles, there was hardly any trend component in any of the different types of services. One reason might be that IT managers tend to pre-select services for consolidation that exhibit no or little trend. Nevertheless, workload traces can change in extended time periods and it is important for IT service managers to monitor workload developments in the data center regularly and also reallocate servers if necessary. The models in the following section provide decision support for the initial and for subsequent allocation decisions. Figure 1: Workload of two services with complementary workload during a day. 4 Problem Formulation First we introduce two optimization models that formalize the consolidation problem based on the cost and workload data as they are typically available in data centers. Then we analyze the computational complexity of this problem and algorithmic approaches to solve the problem for practical problem sizes. 4.1 Static Server Allocation Problems A basic server consolidation problem is the Static Server Allocation Problem (SSAP). Here, the IT service manager needs to consolidate servers by assigning services (i.e., virtual servers) to (physical) target servers so as to minimize the number of servers used or minimize overall server used for sizing the capacity of an SAP server infrastructure (
11 Author: Server Consolidation 11 costs. We start out with this basic model to introduce the mathematical structure of the fundamental decision problem that will be extended later on. Suppose that we are given n services j J that are to be served by m servers i I. Different types of IT resources k K may be considered (e.g., CPU, memory, bandwidth). In this model, for service j the customer orders u jk units of resource k (e.g., SAPS, HP Computons, Megabyte, etc.) and each server has a certain capacity s ik of resource k. y i are binary decision variables indicating which servers are used, c i describes the potential cost of a server, and x ij describes which service is allocated to which server. Considering multiple types of resources, such as memory and bandwidth, the problem can be formulated as follows (see equations (1)). (1) The objective function minimizes server costs, while the first set of constraints makes sure that each service is allocated exactly once, and the second set of constraints ensures that the aggregated workload of multiple services does not exceed the capacity of a single server. The SSAP represents a single service s resource demand as constant over time. In the following we want to consider variations in the workload, and time is divided into a set of intervals T indexed by t={1,, }. Cyclic workloads over time are now represented in the matrix u jkt describing how much capacity service j requires from resource type k in time interval t. Based on this matrix we can reformulate the second set of side constraints to (2) measuring/index.epx).
12 Author: Server Consolidation 12 We will call this model variation the Static Server Allocation Problem with variable workload (SSAPv). The number of servers an IT service provider can save using virtualization has been one of the main sales arguments for software vendors in this field. This obviously depends very much on the level of interference of the respective workload traces (i.e., time series). SSAPv helps to quantify how much can actually be saved compared to dedicated server hosting based on historic demand data. The coefficients u jkt depend on the load characteristics of the servers to be consolidated. Section 5.3 will describe how these coefficients can be derived. In general, we assume. The target servers in SSAPv may be the same type of physical machine as the source servers or different machines with different resource capacities (e.g., CPU architectures). If the computer architecture of source and target servers is different, utilization parameters need to be converted. Such conversion rates can easily be estimated from server tests and benchmark studies [16]. 4.2 Extensions of the Static Server Allocation Problems In the following, we present constraint types which have been elicited from practical applications. All the model extensions have been incorporated in a software tool (vplan), which is now in use with our industry partner: 1. Max no. of Services Constraint: For any target server i a maximum number of services n i may be defined which must not be exceeded by the resulting allocation. The purpose of this constraint may be to limit administrative time and effort in the event of a server failure, etc. 2. Separation Constraints
13 Author: Server Consolidation 13 A subset of services S may have to be allocated on different servers for security or other technical reasons. 3. Combination Constraints A subset of services S may have to be allocated on the same server due to enhanced interapplication communication, or the same operating system requirements. With e being an element of S: 4. Technical Constraints and Pre-Assignment Constraints A subset R I of servers may exhibit a technical feature, e.g. a particular (storage area) network access. A service j may require this particular server attribute. Thus, the service must only be allocated to one of the servers in R providing this attribute. If R = 1, this constraint will be called pre-assignment constraint.. 5. Limits on the number of reallocations In the event that SSAP or SSAPv has to be solved repeatedly, for subsequent planning periods for example, it may be required that the current allocation does not change too much in order to limit administrative costs. The number of migrations of services already allocated in the present solution may therefore be restricted. Let X be the set of x ij with x ij = 1 in the current allocation, and let n be the number of already allocated services (i.e., n = X ) and let r be the number of reallocations allowed.
14 Author: Server Consolidation Complexity Analysis and Algorithmic Solutions Complexity analysis can help us to understand whether we can hope to find exact solutions even for large instances. Unfortunately, as we can show, the problem is strongly NP-hard, even for the simplest case with only one resource and one time interval. Theorem 1: SSAP is strongly NP-hard Corollary 1: SSAP is strongly NP-hard even when only one resource is considered, all servers have the same cost and the same capacity. We provide proof in Appendix A by reducing SSAP to the multidimensional bin packing problem (MDBP). Bin packing was described by Garey, Graham, Johnson & Yao [19], who have shown that it is NP-hard. Many variations of bin packing have attracted a considerable amount of scientific interest over the last few decades, partly due to their relevance in diverse scheduling applications. For exact solutions, one can use traditional integer programming techniques such as branchand-bound algorithms. Knowing that the problem is NP-hard is important, but it does not necessarily mean that it is intractable for practical problem sizes. So for IT managers, it is important to understand which problem sizes can be solved exactly, and how far one can get with heuristic solutions, both in terms of problem size and solution quality. For the simplest version of SSAP with only one resource and no side constraints, one can apply heuristics as they have been developed for the bin packing problem. Two well-known heuristics are the best fit decreasing (BFD) and the first fit decreasing (FFD) approach [20]. It was shown that the solutions they produce require not more than 11/9 OPT + 1 bins (with OPT being the number of bins required by an optimal solution) [21][22]. For SSAP with only a single resource, this means that in the worst-case one would need around 22 % more target servers than necessary using the FFD.
15 Author: Server Consolidation 15 MDBP is harder to solve and polynomial time approximation schemes (PTAS) with worstcase guarantees on the solution quality have been published only recently. The first non-trivial result was produced by Chekuri & Khanna [23] who gave a polynomial time algorithm that, for any fixed > 0, delivers a (1 + d + O(log -1 ))-approximate solution for constant d. The approach is based on solving a linear programming relaxation for the problem. The basic feasible solution would make fractional assignments for at most dm vectors in d dimensions and m bins or servers. In a second step, the set of fractionally-assigned vectors is assigned greedily. Recently, Bansal, Caprara, and Svirdenko [24] showed a polynomial-time randomized algorithm with approximation guarantee arbitrarily close to ln d + 1 for fixed d for MDBP. For small values of d this is a notable improvement. 5 Experimental Setup Complexity results provide general worst-case results, but provide little managerial insight into when the approach is applicable for server consolidation and what problem sizes can be solved for specific inputs of this problem. As we will see, the resource demands of different types of applications or the number of time intervals heavily impact the problem sizes that can be solved. Also, it is important to understand the impact of parameters such as the required service level on the solution quality. In the following, we will provide a detailed experimental analysis focusing on these application-specific questions. 5.1 Experimental Data From our industry partner we obtained two extensive sets of workload data. The first set contains 160 traces for the resource usage of web/application/database servers (W/A/D). The second set contains 259 traces describing the load of servers exclusively hosting ERP applications. Both sets contain data of three consecutive months measured in intervals of five minutes.
16 Author: Server Consolidation 16 We analyzed the workload traces using methods from time series analysis (see Appendix B for details). We found strong diurnal seasonalities with nearly all servers and some weekly seasonalities, but almost no long-term seasonal patterns or trend components in the workload traces. One reason is that system administrators typically select those services for consolidation that do not exhibit a significant trend and are more predictable in their overall resource demand over time. Many applications that are hosted in data centers exhibit a low but, as we could see, quite regular utilization of server resources [1]. Of course, one will also find services exhibiting strong trends or long-term seasonal patterns. With or without such services in a consolidation project, it is recommended to monitor developments in the service workloads and, if necessary, reoptimize and reallocate the servers on a regular basis. CPU is well known to be the bottleneck resource for these types of applications under consideration [25]. This obviously depends on the capacity ratio of CPU power to memory size of the target servers. This ratio causes CPU to be the bottleneck resource with our applications. For the type of applications in our data set this was also the only resource that was considered in consolidation planning of our industry partner. Without loss of generality we restrict our attention to CPU workload in our experiments. The method and toolset have been applied to multiple resources (CPU and memory), in the past as well, without any changes. It is interesting to note that CPU demand of the ERP services is typically higher than that of the W/A/D services. Since in our example these sets of applications have also been managed by different groups on different hardware, these two types of services were evaluated in separate experiments. For our experiments we generated sets of time series containing different numbers of W/A/D or ERP services by sampling with replacement from the original W/A/D or ERP set. The same problem instances were used when models (SSAP vs. SSAPv) and algorithms (see Section 5.2) were compared with respect to runtime or solution quality.
17 Author: Server Consolidation Allocation Algorithms For SSAP with only one resource, we have used branch-and-bound (B&B), as well as first fit (FF) and first fit decreasing (FFD), heuristics. FFD, for example, operates by first sorting services to be assigned in decreasing order by volume, and then assigns each to the first server in the list with sufficient remaining capacity. FF does not include the sorting. SSAPv is more complicated as it involves several time intervals or dimensions. Therefore, we propose an LP-relaxation-based heuristic. As compared to the PTAS described in Chekuri and Khanna [23], we also use the results of an LP-relaxation in the first phase, but use an integer program in the second step to find an integral assignment of those services that were fractionally assigned in the LP relaxation. This has two reasons: First, in particular for W/A/D services, the number of servers used was very low compared to the number of services, leading to a low number of integer variables. As we will see, the integer program in the second phase was responsible for only a small proportion of the overall computation time. Second, an advantage of the LP-based approach and the IP in the second phase (as compared to simple generalizations of FF or FFD) is the fact that constraints (see Section 4.2) can easily be integrated. We will call this LP-relaxationbased procedure the SSAPv Heuristic. For SSAP B&B, SSAPv B&B, and the SSAPv Heuristic, the number of servers used does have a significant impact on the computation time. Each additional server increases the number of binary decision variables by 1 + n. In order to keep the number of decision variables as low as possible, we have used a specific iterative approach for all three algorithms. In the first iteration, we solved the problem with m being the lower bound number of servers (LB) first. LB can be calculated as in (8), with s being the server capacity assumed to be equal for all servers. (8)
18 Author: Server Consolidation 18 The lower bound is based on the assumption that services could be allocated in fractional quantities of the demand of the target servers. Therefore, no integral solution can be lower. If this problem turns out to be infeasible, m is incremented by 1. Infeasibilities are typically detected very quickly, while feasible solutions with a high number of servers m can take a very long time. This is repeated until a feasible solution is found. The first feasible solution found in the B&B search tree is obviously an optimal solution, minimizing the number of target servers. The computation times reported in our experiments will summarize all iterations. Note that with identical servers (i.e., equal costs and equal capacities) there are at least m! equivalent solutions for SSAP and SSAPv, which will cause huge search trees in traditional B&B algorithms. This problem is sometimes referred to as symmetry in integer programming [17]. A straightforward approach that reduces the computation time considerably is to differentiate the cost of servers c i by a small amount. In our experiments c i was set to i (i = 1,, m). 5.3 Data Preprocessing An integral part of the server consolidation method is the data preprocessing. Data preprocessing results in a discrete characterization of daily patterns in the workload traces and allows the solving of the allocation problem as a discrete optimization problem (see Section 4). The number of time intervals considered impacts the number of constraints in our models. More time intervals might better exploit complementary resource demands and thus lead to denser packing. Assuming that there is seasonality in the workload traces, one can derive estimators for the resource demands in different discrete time slots that reflect the variations of workload over time. Our experimental data represents the average CPU utilization of consecutive five-minute intervals measured over three months for 160 W/A/D and 259 ERP services. The original workload traces are denoted by u raw jkt, whereas u jkt will be an estimator for the demand of resource k of a service j in interval t that is derived from the original workload traces.
19 Author: Server Consolidation 19 In the following, we will describe a two-step process to derive the parameters u jkt for our optimization models from the original workload traces. In the first step, we will derive an estimator for individual five-minute time intervals, while in the second step we will aggregate these intervals to reduce the number of parameters for the optimization. This procedure is based on the analysis of the workload traces (see Appendix B) that suggested leveraging daily seasonalities in the time series. We will describe a day as a period of observation. Let p denote the number of periods contained in the load data (in our case, p = 92 days in load traces of three months). A single period is then described by intervals, which are individual measurements in the raw data (in our case, = 288 five-minute intervals per day). We can now derive the empirical distribution of a random variable U jkt for each interval t = 1,, 288, each of which has 92 observations. For example, U jkt might contain all measurements for the time interval from 12:00 p.m. to 12:05 p.m. across three months. In Figure 2 the CPU requirements of a service (in custom unit SPU = 50 SAPS) are plotted for all 24 hours of a day. Hence, each parallel to the y-axis captures a sample of about 92 values according to the 92 days of measurements for each five-minute interval of the day. These empirical distributions of U jkt for the various intervals enable us to derive an estimator u jkt. This estimator takes into account the risk attitude of the IT service manager. For instance, the 0.95-quantile of U jkt is an estimator for the resource requirement of service j where 95 % of all requests can be satisfied.
20 Author: Server Consolidation 20 Figure 2: Workload profile of an ERP service However, even if at one point in time the workload of a particular service exceeds the capacity of a server, this might only mean that the response time increases, not that the request is canceled. Also, depending on the correlation, adding multiple such services or random variables might lead to a reduction of variance. The experimental results (Section 6) will provide more information on how this setting impacts the quality of service. In order to reduce the number of model parameters, it is useful to consider coarser-grained intervals (e.g., hours). This can be achieved by aggregating the estimators of five-minute intervals in a second step. For example, 12 adjacent five-minute intervals may be aggregated to a one-hour interval by choosing the maximum of these 12 five-minute interval values. This ensures that the service level is maintained in all five-minute time intervals. For SSAP, where we can only consider a single utilization parameter u jk, the maximum of all five-minute intervals was considered.
21 Author: Server Consolidation Experimental Design The experiments were designed to answer managerial questions as to which problem sizes can be solved with which quality, and how these variables depend on the models, algorithms and parameter settings above. It should provide IT managers with guidance in what methods and model parameters are appropriate for a consolidation task. In our experiments we use the following treatment variables: Model (SSAP or SSAPv) Algorithm Service type (W/A/D or ERP services) Number of services Server capacity (CPU capacity in SAPS) Risk attitude (as quantiles of the workload distributions) Number of time intervals considered in SSAPv (e.g., 1-hour vs. 6-hour intervals) Sensitivity with respect to additional allocation constraints (see Section 4.2) Depending on the model, we analyzed different solution algorithms. SSAP (with only 1 resource) was solved using Branch & Bound (B&B), First Fit (FF), and First Fit Decreasing (FFD), while SSAPv was solved using a Branch & Bound (B&B) and the SSAPv Heuristic described in Section Dependent Variables For the analysis of the solution quality, the lower bound number of servers LB and the actual number of servers required were measured. In addition, we measure computation times of the algorithms employed. The open source solver lp_solve was used for the revised simplex and the B&B algorithm. To double-check the results, we also used the COIN-OR CBC branch-and-cut IP solver with the CLP LP solver. All other software, including the implementation of the FF and FFD heuristics, was
22 Author: Server Consolidation 22 implemented in Java Experiments were executed on workstations running Windows XP Professional (Service Pack 2) on an AMD Athlon XP Model 8 with 2.1 GHz. The timeout for all single lp_solve runs (both Simplex and B&B) and for the single FF and FFD runs was set to 20 minutes. First, this has already allowed us to solve very large consolidation problems of up to 700 servers, i.e. we could solve those sizes that were considered practically relevant by our industry partner. Second, respective tools are used in an iterative manner and field consultants want to explore different settings in a matter of minutes rather than waiting for hours. It also allowed us to conduct the large number of experiments that are summarized in the following. Nevertheless, it is interesting to understand what happens to larger problem sizes that have not been time-constrained. Therefore, a number of instances were analyzed without setting a timeout, or one of 24 hours only. 6 Results In the following, we will describe selected experimental results. Unless otherwise stated, we will report the results of W/A/D instances with 24 time intervals based on the 95 th percentile. We will first discuss computation time and then solution quality with respect to different treatments.
23 Author: Server Consolidation Computation Time Depending on Problem Size Figure 3: Computation times for SSAP B&B, SSAPv B&B, SSAPv Heuristic; server cap. = 5000, W/A/D Services Figure 3 shows the computation time of the different algorithms. For each of the different numbers of services (x-axis) 20 instances have been sampled (with replacement) from the set of W/A/D services. These 20 instances were then solved using the different algorithms. Here, we examine a target server capacity of 5000 SAPS which was identical to the environment of our industry partner. Furthermore, we examine 24 time intervals and determine service demand based on the 95 th percentile of five-minute intervals. Figure 3 illustrates the computation time of different instances, while Figure 4 shows the proportion of all 20 instances that could be solved in time. Beyond 500 services, the SSAPv Heuristic is capable of solving more instances within 20 minutes than the SSAPv B&B. In total, the SSAPv Heuristic could solve 70 % and SSAPv B&B could solve 65 % of all 280 instances within 20 minutes.
24 Author: Server Consolidation 24 Figure 4: Percentage of solvable W/A/D instances within 20 minutes regarding different algorithms It is interesting to note that most of the time with the SSAPv Heuristic was spent on the initial LP relaxation, while the subsequent B&B could be solved in a few seconds due to a low number of fractional assignments, which is restricted by the number of servers multiplied by the number of time intervals considered. The run times of the SSAP B&B exhibit a larger variance beyond 300 services than the SSAPv solutions. This is due to the fact that SSAP needs more target servers, and consequently has more decision variables but fewer constraints than SSAPv. Accordingly, Figure 4 shows that SSAP B&B often solves fewer instances than the SSAPv algorithms when used with W/A/D services (in total 56 % of 280 instances). We have also analyzed run times of large instances without a timeout. A few instances with 700 or even 800 services could be solved in 100 minutes, while most took much longer. Instances with 900 services or above could not even be solved within 24 hours. For ERP services, we have also used the same server capacities of 5000 SAPS. A notable
25 Author: Server Consolidation 25 difference is that the CPU demands of ERP services are significantly higher than those of W/A/D services. As we will see, this does have a strong effect on the empirical hardness of the problem. Figure 5 shows that in this case SSAP B&B solves more instances within 20 minutes than the SSAPv algorithms (71 % of all instances). SSAPv B&B solved 45 %, and the SSAPv Heuristic 43 % of all instances. The SSAPv Heuristic, however, was faster in 72 % of those cases where at least the SSAPv B&B or the SSAPv Heuristic produced a solution within 20 minutes. Compared to the W/A/D services, we could only solve much smaller instances before the solver hit the 20-minute timeout. One explanation is the number of servers and consequently decision variables needed. Due to the resource demands of ERP services, on average only 8.47 services could be assigned to a target server, as compared to W/A/D services that would fit on such a server on average. Consequently, the number of decision variables increased more than the number of constraints, compared to a similar instance with W/A/D workload traces. For example, for an instance with 150 ERP workload traces, we had 2,567 decision variables and 558 constraints; 150 W/A/D traces resulted in only 985 decision variables and 306 constraints. Consequently, we could only solve problems with much smaller numbers of services. For ERP services, we have tested larger instances. Several instances with up to 250 services could be solved within five hours, instances with 275 and more services could not be solved within 24 hours. For both W/A/D and ERP services, the SSAP First Fit and SSAP First Fit Decreasing show very low computation times of less than one minute for instances of up to 1200 services. Also here, ERP instances take longer to solve than W/A/D instances, since more target servers are required. Note that SSAP First Fit and First Fit Decreasing can both only consider a single resource and a single time dimension.
26 Author: Server Consolidation 26 Figure 5: Proportion of solvable ERP instances within 20 minutes (5000 SAPS) 6.2 Solution Quality Depending on Problem Size For the experiments we assume servers with equal capacity. In formula (8), we have already introduced the lower bound number of servers required (LB). The lower bound is calculated as if services could be fractionally assigned to servers. Note that LB depends on the number of time slots considered. Considering more time slots tends to result in a smaller lower bound of servers. In our experimental setting, the LB for W/A/D led to an upper bound pack ratio (n/lb) of services per server on average. Figure 6 shows the factor Q by which the computed number of required servers exceeds the lower bound number of servers, i.e.. We will refer to this excess ratio as solution quality that is, the closer Q is to one, the better the solution is, i.e. the closer it is to the lower bound.
27 Author: Server Consolidation 27 SSAPv B&B always achieved an allocation with the lower bound number of servers (i.e., = 1.0). The solution quality of SSAP B&B and SSAP solved with the first-fit-decreasing (FFD) heuristic is on average 1.45 times the lower bound ( achieved the same solution quality most of the time ( = 1.45). FF and FFD ). In other words, SSAPv with 24 time intervals per day required on average 31 % fewer servers than SSAP (with only one time interval per day). For example, a consolidation problem with 250 W/A/D services resulted in an allocation of 10 servers with SSAPv, but it required 15 target servers with SSAP. The average quality of SSAPv Heuristic was equal to the optimal solution ( = 1.0). Figure 6: Solution quality of Algorithms for W/A/D services and varying no. of services (5000 SAPS) These results are in line with our analysis of ERP services. As mentioned earlier, problem instances with ERP load traces are harder to solve than with W/A/D workload traces. However, we found a similar solution quality for SSAPv and SSAP for ERP services compared with W/A/D services: The SSAP B&B solution was on average 1.48 times worse than LB ( = 1.48),
28 Author: Server Consolidation 28 SSAP FFD was on average 1.49 times worse than LB ( = 1.49, = 1.50), and the average quality of SSAPv B&B solutions was equal to 1.0, and = 1.01 (with n/lb = 8.47 on average). 6.3 Impact of Risk Attitude on Solution Quality So far we have always assumed the decision maker to select the 95th percentile in the data preprocessing. We call this parameter the decision maker s risk attitude or the aggregation quantile. In other words, if each service was considered in isolation, 95 % of the historical service demand would have been satisfied without delay at this capacity. It is not clear whether such a parameter setting leads to actual overbooking of server resources or if it is a more conservative estimate due to a reduction in variance when multiple services are assigned to the same server. The level of overbooking depends on the aggregate demand of services on a server and their correlation. One way to analyze the effect of different parameter settings is the analysis of capacity violations based on the historic workload traces that are assigned to a single server. We applied different quantiles (0.4, 0.45,, 1) of the sets U jkt (see Section 5.3), solved ten different consolidation problems of 250 W/A/D services each using SSAPv B&B, and measured the capacity violations on all servers. The number of capacity violations is depicted as average, minimum, and maximum percentage of the 10 instances of the overall number of intervals (see Figure 7).
29 Author: Server Consolidation 29 Figure 7: Min, Avg, Max percentage of capacity violations per server for 10 problem instances with 250 W/A/D services depending on aggregation quantile. Server capacity = 5000 SAPS. Even using a 0.4 quantile, only in around 0.12 % of the five-minute intervals in three months did the resource demands exceed the server capacity. It depends heavily on the type of application whether such capacity violations matter. The 0.8 quantile resulted in % capacity violations, which translates on average to 6.2 five-minute intervals per server, where the CPU demand exceeded the capacity in three months. In addition to a reduction of variance through consolidating multiple services on a server, the low number of capacity violations that result from these parameter settings is mainly to be explained by the data preprocessing. When aggregating from five-minute intervals to a 60-minute interval, the maximum of 12 five-minute interval quantiles is used in order to ensure that enough capacity is available in all five-minute intervals. Figure 7 also shows the number of servers needed for a particular aggregation quantile. Interestingly, the average number of servers does not change significantly between the 0.4 and 0.8 quantile, whereas the number of capacity violations decreases. In contrast, the number of required servers increases for quantiles between 0.8 and 1, which is based on the fact that CPU utilization
30 Author: Server Consolidation 30 has a long tail with observations where the service needed a lot of CPU. To account even for these bursts requires an extra number of servers. 6.4 Influence of the Interval Size Considered The second model parameter is the granularity of time intervals, which impacts the number of constraints in SSAPv, and the solution quality. Figure 8 shows the average, minimum, and maximum number of required servers for ten different sets of 250 W/A/D services. Note that an interval size of 24 hours (i.e., a single interval) reduces SSAPv to SSAP. On average, there is almost no improvement in the number of servers needed, whether one considers 15-minute or twohour intervals. However, while we only required 10 servers on average for the 2 hour interval, the single 24 hour interval (SSAP) required between 14 and 16 servers. Figure 8: 250 W/A/D services, with 10 sampled instances per interval size, cap = 5000, SSAPv B&B 6.5 Influence of Additional Allocation Constraints In Section 4.2, we introduced a number of additional side constraints that can become relevant in server consolidation projects. It is easy to incorporate these constraints in LP-based heuristics and
31 Author: Server Consolidation 31 the IP formulation, as compared to SSAP FF. These side constraints, however, will impact the solution time and solution quality (see Appendix C for details). If managers set an upper bound on the number of services per server (1), the number of servers typically increases. As an increasing number of servers leads to more decision variables, it also has a negative impact on computation time. The combination and separation constraints (2/3) had little effect on the solution quality for W/A/D services, even for large numbers of such constraints. They had a negative impact on computation time. Similarly, the technical constraints (4) had little effect on the number of servers needed, and the computation time decreased, as long as the number of required servers did not increase. 6.6 Summary The following provides a summary of the main results and of the experimental analysis relevant to IT managers. 1. The server consolidation problem (SSAP and SSAPv) is an NP-hard optimization problem The problem can be reduced to the vector bin packing problem, which is strongly NP-hard. 2. The solution time depends on the resource demand of services relative to server capacity There was a significant difference between W/A/D and ERP services in our experiments with respect to their resource demands. As we have shown, the relation between the resource demands of services and server capacity heavily impacts the size of the problems that can be solved. While on average only 8.5 ERP services could be allocated to a server in our experiments, it was possible to install 22 to 25 W/A/D services on such a server. Out of 20 SSAPv problem instances, most could be solved within 20 minutes with both the Heuristic and the B&B for up to 350 W/A/D services. In contrast, for ERP services, we could reliably solve these instances in time in only up to around 50 services. For W/A/D services some
32 Author: Server Consolidation 32 larger instances of 600 to 800 services could be solved within 100 minutes, but 900 instances were beyond what could be solved in a day. For W/A/D services, the SSAPv Heuristic solved a higher number of instances than SSAPv B&B. This is due to the fact that there are a smaller number of fractional assignments compared to the number of variables in W/A/D problems. SSAP not only resulted in a lower solution quality, but the computation time was also higher since more servers were needed and consequently more decision variables. Although it would appear simpler, there is no reason to use SSAP for W/A/D services. When applied to ERP services, the performance of SSAPv decreases as compared to SSAP. ERP services required more resources, so a larger number of servers were required, leading to around three times as many variables. This had more impact on the SSAPv Heuristic, since the number of fractionally-assigned variables in the initial LP relaxation increased. For ERP services, the percentage of instances solved was not significantly higher for the SSAPv Heuristic than for the B&B. The instances that could be solved, however, were solved faster with the SSAPv Heuristic. As a consequence, the SSAPv Heuristic turns out to be excellent for W/A/D services, but it loses some of its advantages when consolidating ERP services. In the future, new multi-core CPU architectures will expand the capacity of servers significantly. Since we expect hardware capacity to grow faster than resource demands of software, we can expect to solve even larger ERP consolidation problems in the future. 3. Considering daily variations in the workload in SSAPv reduces the number of servers used by 31 % on average compared to SSAP Regarding W/A/D services, the average number of servers needed by SSAPv with 24 time slots was 69 % of the number of servers required by SSAP. Almost equally, the corresponding savings with ERP services is 32 % on average. Interestingly, the SSAPv Heuristic almost always resulted in the
33 Author: Server Consolidation 33 same number of servers needed, but was faster. The SSAPv Heuristic and the B&B found the lower bound number of servers in almost all cases for both W/A/D and ERP services. For SSAP with only a single time interval and a single resource, the FF and FFD heuristics found the optimal solution most of the time. However, it is difficult to consider side constraints in these heuristics. 4. A larger number of time intervals has a positive impact on solution quality Applying the SSAPv, two-hour intervals yield already 29.6 % server savings compared to SSAP, while smaller intervals added little value. 5. High aggregation quantiles have a negative impact on solution quality The risk attitude can have a significant influence on the number of servers needed. Its impact depends on the number of time slots. We have seen, however, that there is little difference between a 0.8 quantile and a 0.6 quantile in the number of servers needed. A 0.95 quantile avoids almost all capacity violations. 6. Additional side constraints are easy to incorporate, but sometimes at the expense of solution time An important advantage of LP-based heuristics is the fact that it is easy to consider technical side constraints. The impact of side constraints on solution time varies. Some can make the problem harder to solve, while others even decrease the solution time. If the number of target servers increases through additional constraints, then computation time will increase (as is naturally the case with upper bounds on the number of services per server). If the number of servers is not increased by the constraints, then the pre-assignment constraints and technical allocation constraints will shorten the computation time, while the combination and separation constraints will increase the computation time.
34 Author: Server Consolidation 34 7 Discussion of Additional Managerial Implications Cost reduction and energy considerations in data center operations are some of the main managerial motivations for this paper. The hardware cost component of most data centers is less than 15 % (for servers, storage etc.) and is falling although the volume of new equipment being installed is increasing [37]. The volume of installed servers has been growing at an approximately 11 % to 15 % compound annual growth rate for the x86 market during the past four years. More importantly, energy consumption of hardware components and for cooling is increasing [38]. Over the past decade the cost of power and cooling has increased 400 % and these costs are expected to rise. In some cases energy costs account for % of the total data center operations costs [2]. Another motivation for energy conservation in data centers is the current debate on climate change. The IT industry is estimated to contribute 2 % of the global CO 2 emissions. 23 % of these emissions are produced by running servers, including cooling [41]; this means about 0.5 % of global CO 2 emissions are caused by running servers. By means of a more efficient resource usage both energy and hardware costs can be decreased. There are different approaches to slowing down the increasing energy demand in data centers. While energy consumption for cooling can be decreased by more efficient data center design in terms of infrastructure topology, floor space, placement of server racks etc., the power consumption of hardware and thereby for cooling, too can be decreased by server consolidation, increasing hardware utilization rates [39]. According to Gartner s Server Consolidation Poll [40], the top three reasons for the interest in server consolidation are Control server sprawl (33 %), Reduce power and cooling needs (25 %), and Provide TCO savings (20 %). In summary, beyond the minimization of investment costs as achieved through scenario one or two (see Section 3.1) the consolidation of servers is of significant importance for data centers regarding the reduction of energy and cooling costs.
35 Author: Server Consolidation 35 Although energy costs as well as server capacities are constantly changing, we will provide a small example, illustrating expected savings through optimal server consolidation. We assume a server consolidation project with 250 source servers. A state-of-the-art midrange server with average investment costs of 2,500 results in energy costs of approximately per year. This estimate is based on 700 W energy consumption and 0.12 / kwh. Note that these costs do not include costs for cooling in a data center. Experts believe that for cooling one could assume the same amount of energy cost per year as for operating the server. In the optimal assignment with SSAPv, we require 10 new midrange servers. The optimal assignment in SSAP (ignoring daily variations) led to a demand for 15 new midrange servers. The energy costs for the operations of these five additional servers in four years (typical write-off time frame) amount to 14,716, with cooling around 28,000. Investment costs for the five servers are 5* 2,500= 12,500, totaling 40,500 as a lower bound on the cost savings. Note that we did not consider costs for space and administration of additional hardware. Moreover, a manual assignment will likely be suboptimal and not be able to consider multiple resources, multiple time intervals and side constraints. If IT managers do not do a thorough analysis of the workload traces, it could easily happen that they consolidate servers with positively correlated workloads, leading to further losses in production efficiency. Apart from basic savings, decision support for server consolidation reduces the error-prone and time-consuming process of manual planning. Obviously, virtualization has benefits that are beyond savings in investment and energy costs. For example, managing and migrating servers from one hardware server to another becomes much easier. However, virtualization software also comes at a cost. Whether these benefits outweigh the (actual) cost is a topic frequently discussed in reports by market research analysts. It heavily depends on the cost of introducing virtualization software. For these reasons, a total cost analysis of server consolidation projects is outside the scope of this article.
36 Author: Server Consolidation 36 8 Related Work Our paper follows a design science approach as outlined in [14]. It adds also to the discussion on business value of IS as in [27] and [28]. While there has been a lot of work on capacity planning and resource allocation in IS in general, little work has focused on server consolidation and respective planning problems. Closest in spirit to SSAP is a workshop paper by Rolia, Andrzejak and Arlitt [29], who suggest an integer program for allocation problems in a data center. Another heuristic for a similar allocation problem was presented in a talk by HP Labs [30]. Further details have not yet been published. In contrast to this work, we discuss a set of models focused on server consolidation, where we explicitly consider variations in workload over time and the server costs, suggest specific algorithms for the data aggregation and the optimization, and provide an extensive empirical evaluation focusing on managerial questions. Rolia et al. [31] and Cherkasova and Rolia [32] describe the architecture of a workload placement service that is inspired by portfolio theory and enables different QoS objectives for different services in the presence of common resource schedulers. Gmach et al. [33] also address the characterization of workloads. They propose a method to extract workload patterns, periodicities, and trends from past workload traces in order to provide a forecast of service demands. Seltzsam et al. [34] describe a system called AutoGlobe for service-oriented database applications. In contrast to SSAP(v), the static allocation heuristic in [34] determines an initial schedule trying to balance the workload across servers, while a fuzzy controller is used to handle overload situations based on fuzzy rules provided by system administrators. SSAP also has similarities with some optimization problems as can be found in the Operations Management (OM) literature. We have already discussed related literature on MDBP, but it is also useful to describe the relation to the well-known capacitated facility location problem
37 Author: Server Consolidation 37 (CFL) to highlight the specifics of server consolidation. The main structural variables in SSAPv (y i and x ij ) are binary decision variables, while in the CFL one has integer variables describing the number of facilities built in a location, and real valued variables describing whether demand in location j is assigned to a facility built at location i. Also, service costs are included in the objective function of the CFL, which is not an issue in SSAPv. The CFL [35] has three sets of side constraints; SSAPv has two. More importantly, u jk, the resource demand in SSAPv is not treated in the CFL. In particular, the treatment of workload traces (u jk ) and the respective data preprocessing is one of the main contributions of our paper (see Section 5.3) and specific to server consolidation. 9 Conclusions Efficiency in the production of IT services is crucial in today s competitive markets. Virtualization and server consolidation can lead to increased utilization of hardware, and therefore to increased production efficiency. Server consolidation poses a number of new and fundamental decision and planning problems in IT service management. In this paper, we have proposed a capacity planning method for virtualized IT infrastructures that combines a specific data preprocessing and an optimization model. 3 We characterized the computational complexity of these models, proposed algorithms and provided extensive experimental evaluations based on workload traces from an industry partner. The consideration of variations in the workload in SSAPv yielded hardware savings of around 31 % as compared to optimal allocations in SSAP. This result holds for two widespread classes of applications, namely web/application/database servers on the one hand, and ERP services on the other. The SSAPv Heuristic allowed us to solve large-scale server consolidation problems within minutes, while easily integrating additional technical side constraints for the allocation. The different resource demands 3 Patents have been filed for methods described in this paper.
38 Author: Server Consolidation 38 of the two types of services, however, had a significant impact on the problem sizes that could be solved. The approach is now in use with consultants in the field employing a custom-developed software tool. It reduces the error-prone and timely process of manual planning. Workloads can change over time. It is important to analyze the workload traces regularly and re-optimize the allocation if necessary. If workloads are less predictable or exhibit significant trend, an automated controller can perform these tasks and move applications automatically from one physical server to another, allowing for adaptive, self-organizing data center management. The basic models discussed in this paper can serve as the basis for respective controllers. 10 Literature 1. Parent, K. Server Consolidation Improves IT's Capacity Utilization. Court Square Data Group, Filani, D., He, J., Gao, S., Rajappa, M., Kumar, A., Shah, P., and Nagappan, R. Technology with the Environment in Mind. Intel Technology Journal, 12, 1 (2008). 3. IDC-Press. Increasing the Load: Virtualization Moves Beyond Proof of Concept in the Volume Server Market., Chang, J.C.-J., and King, W.R. Measuring the Performance of Information Systems: A Functional Scorecard. Journal of Management Information Systems, 22, 1 (2005), Cox, J.F., and Clark, S.J. Problems in Implementing and Operating a Manufacturing Resource Planning Information System. Journal of Management Information Systems, 1, 1 (1984), Sandeep, P., and Han, T.-D. Distributing Multimedia Content to Balance Quality of Service and Cost. Journal of Management Information Systems, 17, 1 (2000). 7. Lee, J.-N., Miranda, S.M., and Kim, Y.-M. IT Outsourcing Strategies: Universalistic, Contingency, and Configurational Explanations of Success. Information Systems Research, 15, 2 (2004), OGC. ITIL Best Practice for Service Delivery. Norwich: The Stationary Office, ISO. ISO IT Service Management Standards. ISO/IEC, Noel, E., and Tang, K.W. Performance modeling of multihop network subject to uniform and nonuniform geometric traffic. Networking, IEEE/ACM Transactions on, 8, 6 (2000), Thakkar, S.S., and Schweiger, M. Performance of an OLTP application on Symmetry multiprocessor system. 17th Annual International Symposium on Computer Architecture, Seattle, WA., 1990, pp
39 Author: Server Consolidation Menasce, D. Performance by Design. Computer Capacity Planning. Upper Saddle River, New Jersey: Prentice Hall PTR Menasce, D., Almeida, V., and Dody, L. Performance by Design: Computer Capacity Planning by Example. Prentice Hall, Peffers, K., Tuunanen, T., Rothenberger, M.A., and Chatterjee, S. A Design Science Research Methodology for Information Systems Research. Journal of Management Information Systems, 24, 3 (2008), Missbach, M., and Stelzel, J. Adaptive Hardware-Infrastrukturen für SAP. SAP Press, Zhang, Q., Cherkasova, L., Mathews, G., Greene, W., and Smirni, E. R-Capriccio: A Capacity Planning and Anomaly Detection Tool for Enterprise Services with Live Workloads. HP Laboratories, Margot, F. Symmetric ILP: Coloring and small integers. Discrete Optimization, 4, 1 (2007), Garey, M.R., and Johnson, D.S. Computers and Intractability: A Guide to the Theory of NP- Completeness. New York: W.H. Freeman and Company, Garey, M., and Graham, R. Resource constrained scheduling as generalized bin packing. Journal of Combinatorial Theory, Ser. A 21 (1976), Wikipedia. Bin packing problem Yue, M. A simple proof of the inequality FFD (L) 11/9 OPT (L) + 1, for all L, for the FFD binpacking algorithm. Acta Mathematicae Applicatae Sinica (English Series), 7, 4 (1991), Dósa, G. The Tight Bound of First Fit Decreasing Bin-Packing Algorithm Is FFD ( I ) 11/9 OPT ( I )+6/9. Combinatorics, Algorithms, Probabilistic and Experimental Methodologies, 2007, pp Chekuri, C., and Khanna, S. On Multi-dimensional Packing Problems. ACM-SIAM Symposium on Discrete algorithms, Baltimore, Maryland, United States: Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 1999, pp Bansal, N., Caprara, A., and Sviridenko, M. Improved approximation algorithms for multidimensional bin packing problems. 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06) (2006), Kemper, A., and Eickler, E. Datenbanksysteme. Eine Einführung.: Oldenbourg, Williams, H.P. Model Building in Mathematical Programming. John Wiley & Sons Ltd, Mahmood, M.A., Kohli, R., and Devaraj, S. Measuring Business Value of Information Technology in E-Business Environments. Journal of Management Information Systems, 21, 1 (2004), Chan, Y.E. IT Value: The Great Divide Between Qualitative and Quantitative and Individual and Organizational Measures. Journal of Management Information Systems, 16, 4 (2000), Rolia, J., Andrzejak, A., and Arlitt, M. Automating Enterprise Application Placement in Resource Utilities. Workshop on Distributed Systems: Operations and Management, Heidelberg, Germany: Springer, 2003, pp. 12.
40 Author: Server Consolidation Zhang, A., Safai, F., and Beyer, D. Server Consolidation: High-Dimensional Probabilistic Bin- Packing. INFORMS Conference, San Francisco, Rolia, J., Cherkasova, L., Arlitt, M., and Machiraju, V. An Automated Approach for Supporting Application QoS in Shared Resource Pools. IFIP/IEEE International Workshop on Self-Managed Systems & Services, France, Nice, Cherkasova, L., and Rolia, J. R-Opus: A Composite Framework for Application Performability and QoS in Shared Resource Pools. International Conference on Dependable Systems and Networks, Philadelphia, PA, USA, 2006, pp Gmach, D., Rolia, J., Cherkasova, L., and Kemper, A. Capacity Management and Demand Prediction for Next Generation Data Centers. IEEE Conference on Web Services, Salt Lake City, Utah, USA, Seltzsam, S., Gmach, D., Krompass, K., and Kemper, A. AutoGlobe: An Automatic Administration Concept for Service-Oriented Database Applications 22nd International Conference on Data Engineering (ICDE 2006), Atlanta, Georgia, USA, Nemhauser, G.L., and Wolsey, L.A. Integer and Combinatorial Optimization. Wiley-Interscience, Shumway, R.H., and Stoffer, D.S. Time Series Analysis and Its Applications. Springer, Kumar, R. Important Questions from Gartner's EMEA Data Center Conference 2006: Data Center Trends. Gartner Research (2006). 38. Kumar, R. Data Center Power and Cooling Scenario through Gartner Research (2007). 39. Kumar, R., and Mingay, S. How IT Management Can "Green" the Data Center. Gartner Research (2008). 40. Phelps, J.R. Data Center Conference 2007 Server Consolidation Poll Finds Projects Increasing, Reasons Changing and Outside Assistance Growing. Gartner Research (2008). 41. Kumar, R., and Mieritz, L. Conceptualizing 'Green' IT and Data Center Power and Cooling Issues. Gartner Research (2007).
41 Author: Server Consolidation 41 Appendix A Complexity Analysis of SSAP In the following, we show that SSAP is strongly NP-hard, and hence it cannot be solved optimally for large instances [18]. Theorem 1: SSAP is strongly NP-hard Proof: Let us first define SSAP' as the decision version of SSAP, where the question is whether it is possible to assign a number n of services, each characterized by u jk to a given total number m of servers, each having a capacity of s ik for resource k. c i describes the potential cost of a server. Obviously, SSAP' is in NP, since given a solution it suffices to check the constraints and the value of the solution, which can easily be done in polynomial time [18]. The reduction is from the Multi- Dimensional Bin Packing (MDBP) problem. The decision version of the MDBP or d-dim Vector Packing (VP) is described as follows: We are given a set D of d-dimensional vectors specified by a d-tuple (a i 1, a i 2,, a i d) that must be packed in a certain number z of unit size bins, i.e., with dimensions (1,,1). We say that a set C of items is feasibly packed if these items can be placed in a bin without any two overlapping with each other, this means if for each l = 1,,d. Given an arbitrary instance of MDBP, an instance of the SSAP is constructed as follows: We first set i = j, and D = J, the number of bins z = m, l = k, and the d-tuple a i l = u jk. Without loss of generality, let us assume that all s ik across all servers i in SSAP have an equal capacity s k and the cost of all servers c i to be 1. Now, u jk can be normalized by s k and s k = 1 for k = 1, 2,...,d. An MDBP problem then provides a solution to the decision problem. Conversely, a solution to the decision problem of SSAP provides a solution to the MDBP problem.
42 Author: Server Consolidation 42 In SSAPv u jk is extended to u jkt, i.e. it can be reduced to MDBP as well. For d = 1 this problem is identical to the 1-Dim Bin Packing problem, which is known to be strongly NP-hard [19], i.e. the following corollary directly follows. Corollary 1 The SSAP is strongly NP-hard even when only one resource is considered, all servers have the same cost and the same capacity. Appendix B Analysis of Server Load Data Much of our work is based on an extensive analysis of the workload traces (e.g., cyclic, seasonal, and trend components) from our industry partner. Both ERP and W/A/D services stem from a diverse set of customers in different industries. Selected results on seasonality of the workload traces, characteristics of ERP and W/A/D services, and the utilization of source servers relevant to our study are presented below. Seasonality We have used traditional forecasting methods such as multiple linear regression (MLR) models to model seasonalities. Since this is an extensive set of workload traces (419 overall), we cannot provide results of all the traces. Therefore, Table 1 shows the aggregated results of time series analyses of the workload data available. Here, the relative proportion of the time series is presented which exhibit a particular seasonality or trend, e.g., 87.5 % of the ERP load traces contain weekly seasonality (e.g., weekends with lower utilization). Table 1: Aggregated results of CPU load analysis Seasonalities Trend hourly daily weekly Negative positive W/A/D (160) 12.5% 100.0% 92.5% 0.0% 2.5% ERP (259) 20.0% 90.0% 87.5% 0.0% 2.5% One reason for the fact that few workload traces exhibit a trend might be the fact that IT service providers select those source servers for consolidation that are not characterized by strong growth
43 Author: Server Consolidation 43 rates. Server consolidation is used to consolidate those servers that have low utilization (and no significant trend). If a trend exists, the IT service provider might want to use forecasts and reallocate servers periodically, based on changes in the workload mix. We have also used autocorrelation functions (ACF) as an instrument to analyze seasonalities. ACFs show the correlation (y-axis) for all pairs of observations which have the same lag (x-axis). Given a times series x 1,, x n the sample autocovariance function with lag h is defined as (A1) Accordingly, the sample autocorrelation function (ACF) is defined as (A2) By means of the ACF the degree of a linear relationship between observations with the same time distance is measured using the Pearson correlation coefficient [36]. For instance, for time series measured in five-minute intervals, a lag of 12 means measuring the correlation between all pairs of observations, which are one hour apart. Figure 9 shows the autocorrelation function of a typical workload trace exhibiting both daily and weekly seasonality. Hence, very significant peaks can be observed both at the 288 th lag and its multiples (since a day consists of 288 five-minute intervals) and at the 2,016 th lag and its multiples (number of five-minute intervals in one week). Furthermore, it can be observed that the autocorrelation function not only shows high correlations at the lag of multiples of 288 five-minute intervals but also in the neighborhood around 288 lags (see bottom left plot in Figure 9). This indicates that the daily pattern is slightly shifted back or forth on some days. The aggregation of daily load patterns to larger time slots, such as hours or 30-minute time slots, takes these effects into account when solving SSAPv.
44 Author: Server Consolidation 44 Figure 9: Autocorrelation function of a workload trace (W/A/D, CPU) with daily and weekly seasonality A plot of the workload trace during one week (see Figure 10) shows nearly no load on weekends except for scheduled backup jobs, but very regular utilization throughout the week leading to a daily and weekly seasonality.
45 Author: Server Consolidation 45 Figure 10: Example workload trace of each day in a week (W/A/D, CPU in SPU: 1 SPU = 50 SAPS) with nearly no load at the weekend Descriptive Statistics of ERP and W/A/D workloads Figure 11 and Figure 12 show histograms of the average CPU load for the two sets of services W/A/D and ERP. Figure 11: Histogram of CPU average load of W/A/D services [1 SPU = 50 SAPS]
46 Author: Server Consolidation 46 Figure 12: Histogram of CPU average load of ERP services [1 SPU = 50 SAPS] These histograms illustrate the differences in resource demands of both services. ERP services require significantly more CPU power than the W/A/D services. Most of the W/A/D and ERP services have low average load. Tables 4 and 5 show additional summary statistics, such as average load and coefficient of variation for both types of services. Table 2: Min, max and median of average load for 160 W/A/D and 259 ERP series [1 SPU = 50 SAPS] Average load W/A/D- CPU [SPU] Min Median Max ERP-CPU [SPU] Table 3: Min, max and median of coefficient of variation for 160 W/A/D and 259 ERP series Coefficient of Variation W/A/D- ERP-CPU CPU Minimum Median Maximum
47 Author: Server Consolidation 47 Descriptive Statistics of the Utilization of Source Servers In addition, Table 6, as well as Figures 27 and 28, show the average capacity utilization of ERP source servers, which is higher than that of W/A/D servers. Table 4: Average percental capacity utilization of non-virtualized source servers W/A/D ERP avg-%-utilization [in %] (of 160 series) MAX (of 160 series) AVG (of 160 series) 3.20 MIN (of 160 series) 0.23 avg-%-utilization (of 259 series) MAX (of 259 series) AVG (of 259 series) 6.10 MIN (of 259 series) 0.31 Figure 13: Histogram of average percental CPU utilization of source servers (W/A/D services)
48 Author: Server Consolidation 48 Figure 14: Histogram of average percental CPU utilization of source servers (ERP services) Appendix C Impact of Allocation Constraints Section 6.5 summarizes the impact of side constraints on the solution time. In the following, we will provide a more detailed discussion on the various allocation constraints. Maximum number of services per server constraint The number of required servers and thereby the number of decision variables will possibly increase as the maximum number of services per server is decreased. With n i being the maximum number of services allowed to be assigned to server i, the lower bound number of servers also needs to be redefined.
49 Author: Server Consolidation 49 Figure 15: 200 W/A/D services assigned to 5000 SAPS servers. Left y-axis: no. of servers required; right y-axis: computation time in milliseconds (timeout = 1200 sec), x-axis = max. no. of services allowed per server (-1 = unlimited) The example in Figure 15 shows an increase of computation time with a decreasing limit of services per server. This increase is only present if the respective constraints have an impact on the number of servers employed. An unlimited problem (-1 in the figure) and one that is limited to 26 services per server did not have an impact on computation time, since the number of servers required remained at eight. Tighter upper bounds, however, did have an impact on the number of servers required and the solution time. Combine- and Separate Constraints For our experiments on combine- and separate constraints we used two treatments: a) the number of sets each of which contains services which are to be combined (or separated), and b) the set size that is, the number of services in each set, being the same for all sets. For generating the sets, services have been sampled without replacement and with uniformly distributed probabilities from an experimental set of 300 services which had to be consolidated. Thus, for a set of 300 services to be consolidated, for instance, at most 30 combine (or separate) sets with 10 services each were
50 Author: Server Consolidation 50 configured. Since results for both the combine and separate constraints turned out to be very similar, the results are described together. With small combine (separate) sets, the computation time only slightly increases (even if all services are included in combine constraints). With larger combine sets (5 and 10 services) the computation time increases significantly. Even instances with large numbers of services included in combine (separate) sets have been solvable (as with 60 combine sets containing five services each, for instance). However, with 10 services per combine set, the problem including all 300 services could not be solved within 20 minutes. Interestingly, none of the solutions involved more servers than instances without a combine constraint. The same observations with respect to computation time and solution quality apply to the separate constraints. Figure 16: Computation times depending on the size and number of combine constraints (W/A/D, 300 services)
51 Author: Server Consolidation 51 Pre-assign Constraints We uniformly distributed different numbers of services to be preassigned across the LB number of servers randomly. In our experiments the preassign constraint was implemented by setting upper and lower bounds of corresponding decision variables, which indicate assignments of services to servers, to one or zero when preparing the integer program for the IP solver. Representing constraints (where possible) as lower or upper bounds of decision variables is a recommended practice with current IP solvers [26]. The computation time even decreases with an increasing number of preassignments. Again, this will only hold as long as the number of resulting target servers will not significantly increase, as this would imply an increase in the number of decision variables and thereby increased computation time. Also, if many services are arbitrarily preassigned to target servers, infeasibilities may occur (see Figure 17 with 280 and 300 preassignments). Figure 17: 300 W/A/D services to be assigned to 5,000 SAPS or 7,500 SAPS servers with varying number of preassigned services (20,, 300)
52 Author: Server Consolidation 52 Technical Constraints With technical constraints (as described in Section 4.2), various scenarios and parameterizations of experiments are possible. But, even with multiple types of binary server attributes, such as the presence of particular network accesses, a minimum hard disk drive capacity, or the ability to run a particular guest operating system, the feasible assignments regarding one service are restricted to a particular subset of servers. Servers which a particular service may be assigned to will therefore be called the feasible set of servers in the following. The following two parameters were used for experiments: The number of services which will require a server from a restricted set of feasible servers. The feasible set size being equal for all services which require a particular set of feasible servers. Random feasible sets of the specified size were generated for each of the services that require a set of feasible servers by sampling with replacement from the complete set of servers. Figure 18 shows computation times of the SSAPv B&B for consolidating 300 services to 12 servers. Without additional constraints, 12 servers are sufficient to solve the problem. When adding the attribute-requirement constraints, all instances depicted in Figure 18, that is almost all except for two instances (setsize = 1, 200 services and setsize = 1, 300 services), could be solved with 12 servers. We observed that with an increasing number of services which obtain a restricted set of feasible servers, the computation time decreases. On the other hand, with increasing size of feasible sets, the computation time increases again.
53 Author: Server Consolidation 53 Figure 18: SSAPv Computation time for 300 W/A/D services with different numbers and sizes of attribute-requirement constraints.
54 Author: Server Consolidation 54 Appendix D List of Symbols Model Parameters Attributes of Services u jkt raw U jkt u jkt Set of m (target) servers Set of n services (to be assigned to servers) Set of server resources (e.g., CPU, RAM); index is left away for most of the paper since one resource (CPU) is considered Set T of time intervals t (in most experiments: = 24 intervals corresponding to one hour time slots) Measured average utilization of resource k in time interval t for service j (in our case: ) Set of empirical resource demands of service j in time interval t (of a day), for resource type k ( ) Estimator for the resource demand of service j in time interval t for resource type k ( ). Index k is left away if only one type of resource is considered Attributes of Servers c i Cost of server i (type of cost depends on decision scenario as described in Section 3.1) s ik (or s i, or s) Capacity of server i with respect to resource k, index k is left away if there is only one resource type considered; index i is left away if all servers considered have the same capacity. Data Preprocessing p Number p of periods contained in the raw load data u jkt raw Number of intervals contained by one period (in our case: = 288 five-minute intervals = 1 day) Binary Decision Variables x ij x ij = 1 service j is assigned to server i, otherwise x ij = 0 y i y i = 1 server i is required for solving the consolidation problem the solution of the consolidation problem contains at least one service being assigned to server i
55 Author: Server Consolidation 55 Variables in Model Extensions n i S (or S q ) Maximum number of services which are allowed to be allocated to target server i Set S of services to be combined or separated when being assigned to servers ( ), possibly indexed by q (S q ) if there are multiple combine or separate constraints. R Subset R of target servers ( ) which exhibit a particular attribute (e.g., a technical feature) r Evaluation of Results LB Q Maximum number of reallocations allowed with a repeated application of the SSAP or SSAPv based on an existing allocation of services to servers. Lower bound on the number of servers needed to solve the SSAPv Solution quality (as measured in number of required servers / lower bound number of servers) Indices m Number of servers n Number of services i Index for servers j Index for services k Index for resource types q Index for multiple separate or combine sets (S q ) Number of measurement intervals of one period Appendix E Abbreviations ACF B&B BFD CPU ERP FF FFD IP LB LP MDBP OPT Autocorrelation function Branch and Bound (principle of a tree search algorithm for solving integer programs) Best Fit Decreasing Central Processing Unit Enterprise Resource Planning Services First Fit First Fit Decreasing Integer Program Lower Bound (Number of Servers) Linear Program Multi Dimensional Bin Packing Problem Optimal Solution (in case of the First Fit and First Fit Decreasing algorithms the optimal
56 Author: Server Consolidation 56 solution is defined by the number of required bins) PTAS Polynomial Time Approximation Scheme RAM Random Access Memory SAPS SAP Application Performance Standard SPU Custom unit proportional to SAPS (1 SPU = 50 SAPS) SSAP Static Server Allocation Problem SSAPv Static Server Allocation Problem with variable workloads W/A/D Web-/Application-/Database Services
A Mathematical Programming Approach for Server Consolidation Problems in Virtualized Data Centers. Benjamin Speitkamp and Martin Bichler
IEEE TRANSACTIONS ON SERVICES COMPUTING, VOL. 3, NO. X, XXXXXXX 2010 1 A Mathematical Programming Approach for Server Consolidation Problems in Virtualized Data Centers Benjamin Speitkamp and Martin Bichler
Efficient and Robust Allocation Algorithms in Clouds under Memory Constraints
Efficient and Robust Allocation Algorithms in Clouds under Memory Constraints Olivier Beaumont,, Paul Renaud-Goud Inria & University of Bordeaux Bordeaux, France 9th Scheduling for Large Scale Systems
Parallels Virtuozzo Containers
Parallels Virtuozzo Containers White Paper Top Ten Considerations For Choosing A Server Virtualization Technology www.parallels.com Version 1.0 Table of Contents Introduction... 3 Technology Overview...
Virtual Machine Allocation in Cloud Computing for Minimizing Total Execution Time on Each Machine
Virtual Machine Allocation in Cloud Computing for Minimizing Total Execution Time on Each Machine Quyet Thang NGUYEN Nguyen QUANG-HUNG Nguyen HUYNH TUONG Van Hoai TRAN Nam THOAI Faculty of Computer Science
supported Application QoS in Shared Resource Pools
Supporting Application QoS in Shared Resource Pools Jerry Rolia, Ludmila Cherkasova, Martin Arlitt, Vijay Machiraju HP Laboratories Palo Alto HPL-2006-1 December 22, 2005* automation, enterprise applications,
Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1
Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System
Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows
TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents
The Truth Behind IBM AIX LPAR Performance
The Truth Behind IBM AIX LPAR Performance Yann Guernion, VP Technology EMEA HEADQUARTERS AMERICAS HEADQUARTERS Tour Franklin 92042 Paris La Défense Cedex France +33 [0] 1 47 73 12 12 [email protected] www.orsyp.com
Joint Optimization of Overlapping Phases in MapReduce
Joint Optimization of Overlapping Phases in MapReduce Minghong Lin, Li Zhang, Adam Wierman, Jian Tan Abstract MapReduce is a scalable parallel computing framework for big data processing. It exhibits multiple
Parallels Virtuozzo Containers
Parallels Virtuozzo Containers White Paper Virtual Desktop Infrastructure www.parallels.com Version 1.0 Table of Contents Table of Contents... 2 Enterprise Desktop Computing Challenges... 3 What is Virtual
Genetic Algorithms for Energy Efficient Virtualized Data Centers
Genetic Algorithms for Energy Efficient Virtualized Data Centers 6th International DMTF Academic Alliance Workshop on Systems and Virtualization Management: Standards and the Cloud Helmut Hlavacs, Thomas
W H I T E P A P E R. Reducing Server Total Cost of Ownership with VMware Virtualization Software
W H I T E P A P E R Reducing Server Total Cost of Ownership with VMware Virtualization Software Table of Contents Executive Summary............................................................ 3 Why is
Managing the Real Cost of On-Demand Enterprise Cloud Services with Chargeback Models
Managing the Real Cost of On-Demand Enterprise Cloud Services with Chargeback Models A Guide to Cloud Computing Costs, Server Costs, Pricing Plans, and Chargeback Implementation and Systems Introduction
Cloud Management: Knowing is Half The Battle
Cloud Management: Knowing is Half The Battle Raouf BOUTABA David R. Cheriton School of Computer Science University of Waterloo Joint work with Qi Zhang, Faten Zhani (University of Waterloo) and Joseph
DATA CENTER WORKLOAD CONSOLIDATION BASED ON TRUNCATED SINGULAR VALUE DECOMPOSITION OF WORKLOAD PROFILES
DATA CENTER WORKLOAD CONSOLIDATION BASED ON TRUNCATED SINGULAR VALUE DECOMPOSITION OF WORKLOAD PROFILES Thomas Setzer Technische Universität München [email protected] Abstract In today s data centers, typically
Cloud Computing Capacity Planning. Maximizing Cloud Value. Authors: Jose Vargas, Clint Sherwood. Organization: IBM Cloud Labs
Cloud Computing Capacity Planning Authors: Jose Vargas, Clint Sherwood Organization: IBM Cloud Labs Web address: ibm.com/websphere/developer/zones/hipods Date: 3 November 2010 Status: Version 1.0 Abstract:
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
Energy Constrained Resource Scheduling for Cloud Environment
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing N.F. Huysamen and A.E. Krzesinski Department of Mathematical Sciences University of Stellenbosch 7600 Stellenbosch, South
Scheduling using Optimization Decomposition in Wireless Network with Time Performance Analysis
Scheduling using Optimization Decomposition in Wireless Network with Time Performance Analysis Aparna.C 1, Kavitha.V.kakade 2 M.E Student, Department of Computer Science and Engineering, Sri Shakthi Institute
Managing Capacity Using VMware vcenter CapacityIQ TECHNICAL WHITE PAPER
Managing Capacity Using VMware vcenter CapacityIQ TECHNICAL WHITE PAPER Table of Contents Capacity Management Overview.... 3 CapacityIQ Information Collection.... 3 CapacityIQ Performance Metrics.... 4
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. !-approximation algorithm.
Approximation Algorithms Chapter Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of
White Paper. Recording Server Virtualization
White Paper Recording Server Virtualization Prepared by: Mike Sherwood, Senior Solutions Engineer Milestone Systems 23 March 2011 Table of Contents Introduction... 3 Target audience and white paper purpose...
A Constraint Programming based Column Generation Approach to Nurse Rostering Problems
Abstract A Constraint Programming based Column Generation Approach to Nurse Rostering Problems Fang He and Rong Qu The Automated Scheduling, Optimisation and Planning (ASAP) Group School of Computer Science,
Microsoft SQL Server 2012 Virtualization Licensing Guide. June 2012
Microsoft SQL Server 2012 Virtualization Licensing Guide June 2012 Contents Introduction...4 Quick Overview: Licensing SQL Server 2012 for Virtualization...4 Licensing Individual Virtual Machines...4 Licensing
Integrating Benders decomposition within Constraint Programming
Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP
Beyond the Stars: Revisiting Virtual Cluster Embeddings
Beyond the Stars: Revisiting Virtual Cluster Embeddings Matthias Rost Technische Universität Berlin September 7th, 2015, Télécom-ParisTech Joint work with Carlo Fuerst, Stefan Schmid Published in ACM SIGCOMM
Windows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...
Optimization applications in finance, securities, banking and insurance
IBM Software IBM ILOG Optimization and Analytical Decision Support Solutions White Paper Optimization applications in finance, securities, banking and insurance 2 Optimization applications in finance,
Scheduling Algorithm with Optimization of Employee Satisfaction
Washington University in St. Louis Scheduling Algorithm with Optimization of Employee Satisfaction by Philip I. Thomas Senior Design Project http : //students.cec.wustl.edu/ pit1/ Advised By Associate
Windows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described
How To Model A System
Web Applications Engineering: Performance Analysis: Operational Laws Service Oriented Computing Group, CSE, UNSW Week 11 Material in these Lecture Notes is derived from: Performance by Design: Computer
Parallel Scalable Algorithms- Performance Parameters
www.bsc.es Parallel Scalable Algorithms- Performance Parameters Vassil Alexandrov, ICREA - Barcelona Supercomputing Center, Spain Overview Sources of Overhead in Parallel Programs Performance Metrics for
Discrete Optimization
Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT 1 SARIKA K B, 2 S SUBASREE 1 Department of Computer Science, Nehru College of Engineering and Research Centre, Thrissur, Kerala 2 Professor and Head,
Applied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS. Received December May 12, 2003; revised February 5, 2004
Scientiae Mathematicae Japonicae Online, Vol. 10, (2004), 431 437 431 JUST-IN-TIME SCHEDULING WITH PERIODIC TIME SLOTS Ondřej Čepeka and Shao Chin Sung b Received December May 12, 2003; revised February
Dynamic Load Balancing of Virtual Machines using QEMU-KVM
Dynamic Load Balancing of Virtual Machines using QEMU-KVM Akshay Chandak Krishnakant Jaju Technology, College of Engineering, Pune. Maharashtra, India. Akshay Kanfade Pushkar Lohiya Technology, College
Identify and control performance and capacity risks. Introduction... 2
Application performance testing in VMware environments Identify and control performance and capacity risks Table of contents Introduction... 2 Performance and capacity planning techniques... 2 Rough sizing
IOS110. Virtualization 5/27/2014 1
IOS110 Virtualization 5/27/2014 1 Agenda What is Virtualization? Types of Virtualization. Advantages and Disadvantages. Virtualization software Hyper V What is Virtualization? Virtualization Refers to
The Benefits of POWER7+ and PowerVM over Intel and an x86 Hypervisor
The Benefits of POWER7+ and PowerVM over Intel and an x86 Hypervisor Howard Anglin [email protected] IBM Competitive Project Office May 2013 Abstract...3 Virtualization and Why It Is Important...3 Resiliency
Web Load Stress Testing
Web Load Stress Testing Overview A Web load stress test is a diagnostic tool that helps predict how a website will respond to various traffic levels. This test can answer critical questions such as: How
Delivering Quality in Software Performance and Scalability Testing
Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
Rackspace Cloud Databases and Container-based Virtualization
Rackspace Cloud Databases and Container-based Virtualization August 2012 J.R. Arredondo @jrarredondo Page 1 of 6 INTRODUCTION When Rackspace set out to build the Cloud Databases product, we asked many
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments Applied Technology Abstract This white paper introduces EMC s latest groundbreaking technologies,
Black-box Performance Models for Virtualized Web. Danilo Ardagna, Mara Tanelli, Marco Lovera, Li Zhang [email protected]
Black-box Performance Models for Virtualized Web Service Applications Danilo Ardagna, Mara Tanelli, Marco Lovera, Li Zhang [email protected] Reference scenario 2 Virtualization, proposed in early
5.1 Bipartite Matching
CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the Ford-Fulkerson
GEDAE TM - A Graphical Programming and Autocode Generation Tool for Signal Processor Applications
GEDAE TM - A Graphical Programming and Autocode Generation Tool for Signal Processor Applications Harris Z. Zebrowitz Lockheed Martin Advanced Technology Laboratories 1 Federal Street Camden, NJ 08102
The Benefits of Virtualizing Citrix XenApp with Citrix XenServer
White Paper The Benefits of Virtualizing Citrix XenApp with Citrix XenServer This white paper will discuss how customers can achieve faster deployment, higher reliability, easier management, and reduced
VMWARE WHITE PAPER 1
1 VMWARE WHITE PAPER Introduction This paper outlines the considerations that affect network throughput. The paper examines the applications deployed on top of a virtual infrastructure and discusses the
Virtual Data Warehouse Appliances
infrastructure (WX 2 and blade server Kognitio provides solutions to business problems that require acquisition, rationalization and analysis of large and/or complex data The Kognitio Technology and Data
Parallels Virtuozzo Containers
Parallels White Paper Total Cost of Ownership Analysis: Parallels Virtuozzo VS Traditional Hypervisor Virtualization www.parallels.com The ROI of just the hardware and software components, without considering
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis
Dynamic Resource Allocation in Software Defined and Virtual Networks: A Comparative Analysis Felipe Augusto Nunes de Oliveira - GRR20112021 João Victor Tozatti Risso - GRR20120726 Abstract. The increasing
Solution Guide Parallels Virtualization for Linux
Solution Guide Parallels Virtualization for Linux Overview Created in 1991, Linux was designed to be UNIX-compatible software that was composed entirely of open source or free software components. Linux
Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays
Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays Database Solutions Engineering By Murali Krishnan.K Dell Product Group October 2009
Characterizing Task Usage Shapes in Google s Compute Clusters
Characterizing Task Usage Shapes in Google s Compute Clusters Qi Zhang 1, Joseph L. Hellerstein 2, Raouf Boutaba 1 1 University of Waterloo, 2 Google Inc. Introduction Cloud computing is becoming a key
An Oracle White Paper November 2010. Oracle Real Application Clusters One Node: The Always On Single-Instance Database
An Oracle White Paper November 2010 Oracle Real Application Clusters One Node: The Always On Single-Instance Database Executive Summary... 1 Oracle Real Application Clusters One Node Overview... 1 Always
Step by Step Guide To vstorage Backup Server (Proxy) Sizing
Tivoli Storage Manager for Virtual Environments V6.3 Step by Step Guide To vstorage Backup Server (Proxy) Sizing 12 September 2012 1.1 Author: Dan Wolfe, Tivoli Software Advanced Technology Page 1 of 18
Hard Partitioning and Virtualization with Oracle Virtual Machine. An approach toward cost saving with Oracle Database licenses
Hard Partitioning and Virtualization with Oracle Virtual Machine An approach toward cost saving with Oracle Database licenses JANUARY 2013 Contents Introduction... 2 Hard Partitioning Concepts... 2 Oracle
VMware and Xen Hypervisor Performance Comparisons in Thick and Thin Provisioned Environments
VMware and Xen Hypervisor Performance Comparisons in Thick and Thin Provisioned Environments Devanathan Nandhagopal, Nithin Mohan, Saimanojkumaar Ravichandran, Shilp Malpani [email protected],
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
Red Hat Network Satellite Management and automation of your Red Hat Enterprise Linux environment
Red Hat Network Satellite Management and automation of your Red Hat Enterprise Linux environment WHAT IS IT? Red Hat Network (RHN) Satellite server is an easy-to-use, advanced systems management platform
Microsoft SQL Server 2014 Virtualization Licensing Guide
Microsoft SQL Server 2014 Virtualization Licensing Guide June 2014 Contents 1 Introduction 1 Quick Overview: Licensing SQL Server 2014 for Virtualization -Licensing Individual Virtual Machines -Licensing
A closer look at HP LoadRunner software
Technical white paper A closer look at HP LoadRunner software Table of contents Sizing up the system 2 The limits of manual testing 2 A new take on testing: the HP LoadRunner solution 3 The HP LoadRunner
Rapid Bottleneck Identification
Rapid Bottleneck Identification TM A Better Way to Load Test WHITEPAPER You re getting ready to launch or upgrade a critical Web application. Quality is crucial, but time is short. How can you make the
Everything you need to know about flash storage performance
Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices
Real-Time Scheduling 1 / 39
Real-Time Scheduling 1 / 39 Multiple Real-Time Processes A runs every 30 msec; each time it needs 10 msec of CPU time B runs 25 times/sec for 15 msec C runs 20 times/sec for 5 msec For our equation, A
Red Hat Satellite Management and automation of your Red Hat Enterprise Linux environment
Red Hat Satellite Management and automation of your Red Hat Enterprise Linux environment WHAT IS IT? Red Hat Satellite server is an easy-to-use, advanced systems management platform for your Linux infrastructure.
Data Centers and Cloud Computing
Data Centers and Cloud Computing CS377 Guest Lecture Tian Guo 1 Data Centers and Cloud Computing Intro. to Data centers Virtualization Basics Intro. to Cloud Computing Case Study: Amazon EC2 2 Data Centers
The Trip Scheduling Problem
The Trip Scheduling Problem Claudia Archetti Department of Quantitative Methods, University of Brescia Contrada Santa Chiara 50, 25122 Brescia, Italy Martin Savelsbergh School of Industrial and Systems
An Energy-aware Multi-start Local Search Metaheuristic for Scheduling VMs within the OpenNebula Cloud Distribution
An Energy-aware Multi-start Local Search Metaheuristic for Scheduling VMs within the OpenNebula Cloud Distribution Y. Kessaci, N. Melab et E-G. Talbi Dolphin Project Team, Université Lille 1, LIFL-CNRS,
Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.
Oracle BI EE Implementation on Netezza Prepared by SureShot Strategies, Inc. The goal of this paper is to give an insight to Netezza architecture and implementation experience to strategize Oracle BI EE
Performance Workload Design
Performance Workload Design The goal of this paper is to show the basic principles involved in designing a workload for performance and scalability testing. We will understand how to achieve these principles
Virtualization for Cloud Computing
Virtualization for Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF CLOUD COMPUTING On demand provision of computational resources
Estimating the Size of Software Package Implementations using Package Points. Atul Chaturvedi, Ram Prasad Vadde, Rajeev Ranjan and Mani Munikrishnan
Estimating the Size of Software Package Implementations using Package Points Atul Chaturvedi, Ram Prasad Vadde, Rajeev Ranjan and Mani Munikrishnan Feb 2008 Introduction 3 Challenges with Existing Size
VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5
Performance Study VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5 VMware VirtualCenter uses a database to store metadata on the state of a VMware Infrastructure environment.
How Customers Are Cutting Costs and Building Value with Microsoft Virtualization
How Customers Are Cutting Costs and Building Value with Microsoft Virtualization Introduction The majority of organizations are incorporating virtualization into their IT infrastructures because of the
Statistical Profiling-based Techniques for Effective Power Provisioning in Data Centers
Statistical Profiling-based Techniques for Effective Power Provisioning in Data Centers Sriram Govindan, Jeonghwan Choi, Bhuvan Urgaonkar, Anand Sivasubramaniam, Andrea Baldini Penn State, KAIST, Tata
CloudSim: A Toolkit for Modeling and Simulation of Cloud Computing Environments and Evaluation of Resource Provisioning Algorithms
CloudSim: A Toolkit for Modeling and Simulation of Cloud Computing Environments and Evaluation of Resource Provisioning Algorithms Rodrigo N. Calheiros, Rajiv Ranjan, Anton Beloglazov, César A. F. De Rose,
can you effectively plan for the migration and management of systems and applications on Vblock Platforms?
SOLUTION BRIEF CA Capacity Management and Reporting Suite for Vblock Platforms can you effectively plan for the migration and management of systems and applications on Vblock Platforms? agility made possible
whitepaper Network Traffic Analysis Using Cisco NetFlow Taking the Guesswork Out of Network Performance Management
whitepaper Network Traffic Analysis Using Cisco NetFlow Taking the Guesswork Out of Network Performance Management Taking the Guesswork Out of Network Performance Management EXECUTIVE SUMMARY Many enterprise
Directions for VMware Ready Testing for Application Software
Directions for VMware Ready Testing for Application Software Introduction To be awarded the VMware ready logo for your product requires a modest amount of engineering work, assuming that the pre-requisites
Technical Investigation of Computational Resource Interdependencies
Technical Investigation of Computational Resource Interdependencies By Lars-Eric Windhab Table of Contents 1. Introduction and Motivation... 2 2. Problem to be solved... 2 3. Discussion of design choices...
