CHAPTER 5 OPTIMIZING ENERGY IN DATACENTER USING DVFS 5.1 Introduction Energy efficiency and low carbon strategies have attracted a lot of concern. The goal for 20% energy efficiency and carbon reduction by 2020 drove the Information Communication Technologies (ICT) (Koutitas 2010) sector to strategies that incorporate modern designs for a low carbon and sustainable growth. The ICT sector is part of the 2020 goal and participates in three different ways. In the direct way, ICT are called to reduce their own energy demands (green networks, green IT), in the indirect way ICT are used for carbon displacements and in the systematic way ICT collaborate with other sectors of the economy to provide energy efficiency (smartgrids, smart buildings, intelligent transportations systems, etc.). ICT and in particular datacenters have a strong impact to the global CO2 emissions. Moreover, an important part of the operational expenditure is due to the electricity demands. This work presents the sources and challenges that have to be addressed to reduce carbon emissions and electricity expenses of the sector. The datacenter is the most active element of an ICT infrastructure that provides computations and storage resources and supports respective applications. The datacenter infrastructure is central to the ICT architecture, from which all content is sourced or passes through. Worldwide, datacenters consume around 40,000,000,000 kw/hr of electricity per year and a big portion of this consumption is wasted due to inefficiencies and non-optimized designs. According to the Gartner Report, a typical datacenter consumes per year is equivalent to 25000 households the same amount of energy, and the electricity consumption by datacenters is about 0.5% of the world production (Koutitas 2010). In terms of carbon emissions this power consumption pattern is identical to the airline industry and comparable to emissions generated by
Argentina, Malaysia and the Netherlands. Energy efficiency in ICT is defined as the ratio of data processed over the required energy (Gbps/Watt) and is different than power conservation where the target is to reduce energy demands without considering the data volume. Taking into consideration this ratio, green IT technologies have important benefits in terms of reduce electricity costs and OPEX (Operational expenditure); improve corporate image; provide sustainability; extend useful life of hardware; reduce IT maintenance activities; reduce carbon emissions and prevent climate change; Provide foundations for the penetration of renewable energy sources in IT systems. The demand for high speed data transfer and storage capacity together with the increasingly growth of broadband subscribers and services will drive the green technologies to be of vital importance for the telecommunication industry, in the near future. Already, recent research and technological papers show that energy efficiency is an important issue for the future networks. A review of energy efficient technologies for wireless and wired networks is presented by Brown et al (2007). The design of energy efficient WDM (Wavelength Division Multiplexing) ring networks is highlighted. It is shown that energy efficiency can be achieved by increasing the capital expenditure of the
network, by reducing the complexity and by utilizing management schemes. The case of thin client solutions is investigated and it is shown that employing power states in the operation of a datacenter can yield energy efficiency (Tsirogiannis et al 2010). Efforts have been cited related to agreeing and enabling standard efficiency metric, real time measurement systems, modeling energy efficiency, suggesting optimal designs, incorporating renewable energy sources in the datacenter and developing sophisticated algorithms for designing and managing the datacenters. These approaches have been published by various companies, experts in the field and organizations (Rivoire 2007; Poess 2008; Barroso 2007). Although the IT industry has begun greening major corporate datacenters, most of the cyber infrastructure on a university campus or SMEs of suboptimal energy environment and ad hoc involves a complex network, in small departmental facilities placed with clusters. 5.2 Datacenter Infrastructure and Power Consumption 5. 2. 1 Datacenter Infrastructure Datacenters incorporate critical and non-critical equipments. Critical equipments are related to devices that are responsible for data delivery and are usually named as IT equipments. Non-critical equipments are devices responsible for cooling and power delivery and are named as Non Critical Physical Infrastructure (NCPI). The Fig 5.1 presents a typical datacenter block diagram. The overall design of a datacenter can be classified in four categories Tier I IV each one presenting advantages and disadvantages related to power consumption and availability. In most cases availability and safety issues yield to redundant N +1, N +2 or 2N datacenter designs and this has a serious effect on power consumption.
Figure 5.1 Typical datacenter infrastructures. According to Fig 5.1, a datacenter has the following main units: Heat Rejection is usually placed outside the main infrastructure and incorporates chillers, dry coolers and present an N + 1 design. Pump Room they are used to pump chilled water between dry coolers and CRACs (Computer Room Air Conditioner) and present an N + 1 design (one pump in standby). Switchgear it provides direct distribution to mechanical equipment and electrical equipment via the UPS.
UPS Uninterruptible Power Supply modules provide power supply and are usually designed with multiple redundant configurations for safety. Usually 1000 kva to 800 kw per module. EG Emergency Generators supply with the necessary power supply to the datacenter in case of a breakdown. Usually diesel generators. PDU Power Distribution Units for power delivery to the IT. Usually 200 kw per unit and dual PDUs (2N) for redundancy and safety. CRAC Computer Room Air Conditioners provide cooling and air flow in the IT equipments. Usually air discharge is in up-flow or down-flow configuration. IT Room incorporates computers and servers placed in blades, cabinets or suites in a grid formation. Provide data manipulation and transfer 5.3 Power Consumption in Datacenters The overall datacenter power consumption is related to the associated power consumed by each unit. Efficiency at individual parts is an important step for greening the datacenter aims to the overall datacenter design but optimization is achieved efficiency. The power delivery in a typical datacenter is presented in Fig 5.2. The power is divided into two types: parallel path and a series path. The power enters the datacenter from the main utility (electric grid, generator), the Renewable Energy Supply (RES) (Saynor 2003) utility and feeds the switchgear in series. Within the switchgear, transformers scale down the voltage to 400 600 V. Therefore, In case of a utility failure UPS flows voltage and also fed by the EG. The UPS incorporates batteries for emergency
power supply and process the voltage with a double AC-DC-AC conversion to protect from utility failures and smooth transition to the EG system. Of course, the AC-DC-AC conversion is a process that wastes power and reduces the overall efficiency. The output of the UPS feeds the PDUs that are placed within the main datacenter room. The PDUs break the high voltage from the UPS into many 110 220 V circuits to supply the electronic equipments. Finally, power is consumed for the IT processes namely as storage, networking, CPU and in general data manipulation. Renewable Energy Supply (RES) Data Center IT Equipment Racks P G NCPI Services P IN In Series Power Path Switch Gear UPS PDU Prr Storage P M In Parallel Power Path Main Energy Supply Lights Networking Cooling Telecoms Equip Emergency energy Supply Figure 5.2 Power deliveries in a typical datacenter
Figure 5.3 Power consumption by various parts of a datacenter (system refers to motherboards, fans, etc.) The parallel path feeds the cooling system that is important for the heat protection of a datacenter. The cooling system is also connected to the EG since without cooling a typical datacenter can operate for a few minutes before getting overheated. The cooling system incorporates fans and liquid chillers. The power distribution of these processes in an inefficient datacenter is presented in Fig 5.3. It can be observed that almost 70% of the power is consumed for non-critical operations like cooling and power delivery and conversion and 30% is used by the IT equipments. Of course, a portion of this percentage is also wasted for networking, CPU, fans, storage and memory processing. In other
words, the useful work of the datacenter is associated to a percentage of power, the IT equipments are delivered at smaller than 30%. The power consumption pattern presented in Figure 5.3 is not constant with time but varies according to different parameters (Schneider et al 2012). The main are the workload of the datacenter and the outside environment. Modeling the energy efficiency and the losses of the datacenter s equipments is a complex tasks and crucial simplificative assumptions yielded great errors. Firstly, the assumption that losses associated to the power and cooling equipments can be found constant. It has been observed that the energy efficiency of these equipments is a function of the IT load and presents a nonlinear behavior. In addition, these equipments are usually operating at lower than the maximum capacity loads and this increases the losses of the system. Finally, the heat generated by the network-critical physical infrastructure (NCPI) equipments is not insignificant. In general, the losses of NCPI equipments are highly correlated to the workload of the datacenter in a complex nonlinear relationship. According to the input workload the losses of NCPI equipments can be categorized as follows: No load losses Losses that are fixed even if the datacenter has no workload. The loss percentage increases with decrease of load. Proportional losses Losses that depend linearly on workload. The loss percentage is constant with load. Square law losses Losses that depend on the square of the workload. These losses appear at high workloads (over 90%). The loss percentage decreases with a decrease of load. The IT equipment also presents non-constant losses and variable depends up on the input workload from energy efficiency. Based on these observations it is concluded that
energy efficiency of a datacenter is a complicate factor parameter, non-constant with time. For this reason, a technique for measuring and predicting the datacenter s energy efficiency is of great importance. 5.4 Sources of Losses at Datacenters The operation of datacenters suffers great inefficiencies and a great amount of power is wasted for the operation of non-critical equipments and for the produced heat by the electronic equipments. The main disadvantage of real datacenters is wasted the great amount of energy cooling or it is transformed to heat because of the inefficient operation of electronic equipments that can be NCPI or IT type. The main causes of power are summarized as Power units (UPS, Transformers, etc.) operate below their full load capacities. UPS are oversized to the actual load requirements in order to avoid operating near their capacity limit. Air conditioning equipment consumes to deliver the cool air flow at long distances for taking the extra power. Inefficient UPS equipments. Blockages between air conditioners and equipments that yield to inefficient operation. No virtualization and consolidation.
Inefficient servers. No closed coupling cooling. No efficient lighting. No energy management and monitoring. Underutilization due to N + 1 or 2N redundant designs. Over sizing of datacenter. Under-floor blockages that contribute to inefficiency by forcing cooling devices to work harder to accommodate existing load heat removal requirements. The procedure to transform a datacenter into an energy efficient one (green datacenter) is complex and it can only be achieved by targeting both individual part optimization that can be considered as operational costs and overall system performance that can be considered as planning actions. According to Fig s 5.2 and 5.3, the optimized operation of the datacenter requires the input power to be minimized without affecting the operation of the IT equipments. 5.5 Challenges for Energy Efficiency In general, efficiency can be achieved through the optimization of the operation and the optimal planning. Moreover, standards can be engineered in order to drive energy efficiency. This process incorporates the domains presented in Fig 5.4.
1. Free Cooling 2. Air-Flow Management 3. Liquid Heat Removal 4. Room Lay out 1. Alternative Energy Sources 2. Tri-Generator 1. Efficient UPS and Transformers 2. Blade-Servers 3. Dual-Core Processors 4. Energy Proportional Computing 5. Right Sizing Figure 5.4 Directions for green datacenter transformation
5.6 Optimization of Operational Costs The operational costs are associated to the optimization of individual equipments like the IT equipments and NCPI. 5.6.1 IT-Equipment Efficiency of IT equipments is an important step for the green operation of a datacenter (Kumon 2012). The DCeP metric and the factors ITEE and ITEU show s that energy efficiency is correlated to the efficient use of the IT equipment. Achieving efficiency at the IT level can be considered as the most important strategy for a green datacenter since for every Watt saved in computation, two additional Watts are saved one Watt in power conversion and one Watt for cooling. In general the following actions are necessary to achieve this goal. Retiring some datacenters have application servers which are operating but have no users. These servers add no load losses to removing the system is need. The usable lifetime for servers within the datacenter varies greatly, for some X86 servers ranging from as two years to seven years or more for large, scalable SMP server systems (Jamal 2009). IDC surveys indicate that almost 40% of deployed servers have been operating in place for four years or longer. That represents over 12 million single core-based servers still in use. To exists the servers performance of most datacenters has been designed for cost optimization and not for energy efficiency. Many servers in datacenters have power supplies that are only 70% efficient. To loss the 30% of power heat is going to the server. Therefore, having inefficient power supplies it means too excess money has been spent on power as well as with additional cooling is needed. With current servers is that they are used at only 15 25% of their capacity is the another problem. While the problem of this, from a cooling perspective
and power, to require of the power amount for run traditional servers does not vary linearly with the server utilization. Therefore, to run the ten servers at 10% of utilization will consume the power more than of one or two servers each run at 80 90% utilization. Migrating to more energy efficient platforms - Blade servers that produce a less heat in a smaller area around it is used. Non-blade systems require bulky, space inefficient components are space and hot, to across the many duplicate computers perform at capacity is may or may not. More efficient of the overall utilization by locating these services in one place the blade computers are sharing between them. The efficiency of blade centers is obvious if one considers power, cooling, networking and storage capabilities. Koutitas (2010) has presented an example where 53 blade servers consuming 21 KWatts provided 3.6 Tflops of operation when in 2002 required Intel Architecture based HPC cluster of 512 servers arraigned in 25 racks consuming 128 KWatts. This means that compared to 2002 technology today 17% of the energy is necessary for the same performance. Power of Blade Servers At higher voltage is required within the computers, these computers DC voltages are operate over a range but utilities deliver power as AC. One or more Power Supply Units (or PSUs) are required for converting. The computer operations is to be ensure that the failure of one power source does not affect, even entry level servers has been redundancy of the supplying power, to design the output and adding to the bulk and heat. Therefore, within the enclosures power source from power supply provides a single for all blades. Hence PSU supplying DC to multiple enclosures has been dedicated for separate the single power source may come as a power supply. Cooling of Blade Servers The proper functionality of the components to ensure the system while during the operation, mechanical and electrical components are producing the heat. Most blade enclosures, remove heat by using fans, like most computing systems.
A power supply does not generate the blade s shared cooling and power for heat as traditional servers. To design the newer blades enclose feature high speed, requirements of the system for adjustable fans and control logic is tune the liquid cooling system. Networking of Blade Servers To providing the blade enclosure which blade will connect the one or more network buses from single location (versus one in each computer chassis) ports are individually is present to reducing the connection cost for the individual devices. This also means that the probability of wiring blockage to air flow of cooling systems in the datacenter is minimized. More Efficient Server For energy efficiency, should focus on the vendors making to consume the processors less energy and using the energy is possible as efficiently. To designing the system coordination is closed between the server manufacture and processor. The practical strategies of power efficient computing technologies are presented based on microelectronic, off-chip interconnect, power delivery system, underlying logic device, and associated cache memory. In addition, replacement of old single core processors with dual or quad core machines is important. This combination can improve performance per Watt and efficiency per square meter. Energy Proportional Computing Many servers operate at a fraction of their maximum processing capacity. Efficiency can be achieved when the server scales down its power use when the workload is below its maximum capacity. All servers are being used heavily when the search traffic is high, but low traffic of during periods, hundreds of queries per second for server might shown, any idleness periods of meanings are not to be longer than a few seconds. In addition, it is well known that rapid transition between idle and full mode consumes high energy. Borroso et al (2007) has researched over 5000 Google server shows that the activity profile for most of the time is limited to 20 30% of the servers maximum capacity. Server s power consumption responds differently at varying input workloads.
Server Power Usage Relative to Peak (%) Figure 5.5 Comparison of power usage and energy efficiency for Energy Proportional Computing (EPC) and common server (NO EPC) (Kim, 2010). In Fig 5.5 the normalized power usage and energy efficiency of a common server is presented as a function of the utilization (% of maximum capacity). It can be observed that for typical server operation (10 50%) the energy efficiency is very low meaning that the server is oversized for the given input workload. This means that even for a no workload scenario, the power consumption of the server is high. In case of Energy Proportional Computing (Kim 2010) (EPC) the energy varies proportionally to the input work. Energy-proportional machines must exhibit a wide the range dynamic power a property in the other domains that might be rare but it is not unprecedented in computing equipment. To achieve energy proportional computing two key features are necessary such as wide dynamic power range and active low power modes.
Current processors have a wide dynamic power range of even more than 70% of peak power. On the other hand the dynamic power range of other equipment is much narrower such as DRAM 50%, disk drives 25% and network switches 15%. A processor running at a lower voltage-frequency mode can still execute instructions without requiring a performance impacting mode transition. With active low-power modes there are no other components in the system. Networking equipment rarely offers any lowpower modes, and the only low-power modes currently available in mainstream DRAM and disks are fully inactive. A technique to improve power efficiency is with dynamic Voltage Scaling and Clock Frequency (DFVS) (Yang et al 2012). The performance of CPU is adjusting dynamically (via clock rate and voltage) to match the workload for provide CFVS performance on demand. Another advantage obtained by energy proportional machines is the ability to develop power management software that can identify underutilized equipments and set to idle or sleep modes. Finally, provide the benchmarks such as SPECpower ssj2008a standard application is representative of server workloads for broad class and it can help isolate efficiency differences in the hardware platform. SPECpower_ssj2008 is the first industry-standard SPEC benchmark that evaluates the power and performance characteristics of volume server class and multi-node class computers. With SPECpower_ssj2008, SPEC is defining server power measurement standards in the same way we have done for performance. 5.7 Problem Description Majority of the prior research work done in the area of analyzing power utilization mainly concentrates on task scheduling in the center with respect to task allocation among the application servers targeted power saving or the criteria considering thermal factors. The major research gap is that there are only few implementation work being
carried out in past considering datacenters, unwanted power consumption along with task provisioning. Cloud computing has been accounted for diversified emerging issues which is yet to be resolved. Unfortunately, the current potentials and capability of cloud computing has yet not been enhanced to cater the crucial scope of the usage. There are technical and nontechnical gaps. Most prominent issues is that although vision of cloud computing is to deploy at any location irrespective of time and devices, particularly in case of distributed system and existing clouds is reported to deliver very poor performance in applications. Any significant issue is the migration of data from client s infrastructure to clouds actually cost very high and it is also a time consuming process. Various legal restrictions are yet to be created for the acceptance of this new technology in socioeconomic market. In many datacenters today, the cost of power is only second to the actual cost of the equipment investments. Today the cloud datacenter consumes 1-2 percent of world energy. If left unchecked, they will rapidly consume an ever-increasing amount of power consumption. 5.8 Proposed System The proposed technique attempts to reduce the cumulative power utilization of a specified datacenter which is done by choosing the suitable evaluating resources for task processing depending on the load quantity and transmission capability of the variants of datacenter specified. The transmission capability is represented by available bandwidth facilitated to the unique application servers or clusters of servers. The proposed technique designs an architectural framework possessing the better designed topology of datacenter. The main contribution of this proposed system is to give (i) analysis and in-depth study of real-time Cloud service with virtualization, and (ii) to probe various energyaware virtual machine optimizing schemes based on Dynamic Voltage Frequency Scaling schemes (DVFS). The weighted grouping of the server level W s, rack level W r, and module level W m is given by,
Weighted Grouping = P.Ws + Q. W r + R. W m. (5.1) where P, Q, and R are the weighted coefficients factors. The allocation of higher set of load towards the server is preferred if the value of P is very high. Figure: 5.6. Implementation Plan Priority is designed higher for higher set of task in the racks with reduced traffic events when the value of Q is higher. The selection of loaded modules are preferred if value of R is higher. For understanding the magnitude of the task allocation in datacenter, the value of R plays an important role. The implementation plan of the proposed system is as shown in Fig 5.6.
The selection of the operationally processing server will group the server load S L (l) along with its transmission capability T C (q) which is analogous to equivalent distribute of the uplink resources on the top of rack switch. W T ( q) C s ( l, q) SL( l). (5.2) r In equation 5.2, S L (l) is a parameter which is proportional to the load of any individual servers l, T C (q) which represents the load at the rack uplink by evaluating the traffic blocking intensity in the designed queue q, Δ r is a bandwidth over scheduled parameter at the rack switch and φ will represent a coefficient for share between S L (l) and T C (q). It is also known that both S L (l) and T C (q) should lie within a range (0,1) greater value φ reduces the significance of network variants T C (q). Therefore, with context to (5.2), the criteria having influencing the rack switch can be expressed as: W ( l, q) S r R Tm ( q) ( l). m j 0 Tm ( q) m 1. n n i 1 S ( l) L (5.3) k 1 Wm ( l) Sm( l) SR( l) (5.4) k In the equation (5.4), S R (l) is a load on rack which is derived from the addition of all the component loads on server in the rack, S M (l), is the module load derived as the normalized addition of all of the loads in rack in this module, n and k are quantity of the servers in a rack and the quantity of the racks in the module respectively, T C (q) is dependent to the network load at incoming networking devices and Δ m will represents channel capacity as over scheduled criteria at the module switches. It can be seen that module level criteria F M possesses only components l related to load. This phenomenon is due to the reason that all the components are actually associated to the uniform networking device and split the similar channel capacity by deploying a technique which can work on unit load balancing resulting in discretion of traffic flows by estimating a hash function on all ingress message. It was also noted that idle server in datacenter
consumes 2/3 rd of its crest utilization which concludes that if any provisioned is to be designed that it must merge all the allocated task on a minimum feasible group of computing resources. Also, it is observed that if the application servers are kept in uniform execution mode at high frequency than hardware reliability is found less with its performance effectiveness. Therefore, the proposed load criterion is designed in equation 5.5: 1 1 S (5.5) L ( l) 1 10 10( l ) ( l (1 )) 2 2 1 e 1 e The preliminary component as shown in equation (5.5) explains the data-structure of the function while the 2 nd variant denotes the fining method targeted at the junction towards the highest load of the server. The factor ε represents the dimension and the shift of the descent slope. The load of application server l should be in constraint of (0,1). In case of highly dynamic load towards server, the load on application server will be evaluated as the total of all processing load of the allocated present task. But in case of jobs where the deadline is fixed for completion, the load on the server can be represented as the least quantity of the networking resources which will be required from the application server to accomplish all the jobs at pre-defined time as per SLA. In datacenter, the application server will split the top of rack switch for meeting their transmission requirements. Conversely, representing a section of this channel capacity utilized by a specified application server or a transmission at higher frequency deployed in defined server during execution is definitely highly expensive in terms of processing the algorithm. In order to mitigate such issues, using equation 5.3 and 5.4 will use a variant based on the task allocated stage of the jobqueue at the networking device and levels the channel capacity over scheduling criteria Δ.
As an alternative of depending on the appropriate queue size, the allocation stage q will be leveled with the cumulative queue size T max within the range of (0,1). The variety is analogous to void and complete buffer allocation. By depending on the buffer allocation, the proposed system will be responsive to the increasing blocking in the racks moderately in comparison to rate of transmission differences. Therefore T (q) will be represented as in equation 5.6. T ( q ) 2 q ( T max 2 ) e (5.6) 5.9 Research Methodology Three-tier trees of hosts and switches form the most widely used datacenter architecture consists of the core tier at the root of the tree, the aggregation tier that is responsible for routing, and access the computing server s tier that hold the pool as in fig 5.7. Early datacenters used two tier architectures. Thus, such type of datacenters switches, depending on the requirements bandwidth per-host is used, could typically support not more than 5,000 hosts. Given the pool of servers in today s datacenters that are of the order of 100,000 hosts and the requirement to keep the two layer switches in the access network, a three-tiered design becomes the most appropriate option.
Core Network Aggregation Network Module Module Access Network RACK RACK RACK RACK RACK RACK 10 GE Link GE Link L3 Switch L2/L3 Rack Switch Computing Server Figure 5.7 Three-tier datacenter architecture Although 10 Gigabit Ethernet (GE) transceivers are commercially available, in a three-tiered architecture the computing servers (grouped in racks) are interconnected using 1 GE links. This is due to the fact those 10 GE transceivers: (a) too expensive and (b) needed for connecting computing servers probably offer more capacity. In current datacenters, rack connectivity is achieved with inexpensive Top-of-Rack (ToR) switches. A typical ToR switch shares two 10 GE uplinks with 48 GE links that interconnect computing servers within a rack. The difference between capacities of the uplink and the downlink a switch defines its oversubscription ratio is equal to 48/20 = 2.4:1. Therefore, remaining available individual servers out of their 1 GE links only 416 Mb/s. At the higher layers of hierarchy, the racks are arranged in modules (see Fig. 5.7) with a pair of aggregation switches servicing the module connectivity. Typical oversubscription ratios around 1.5:1 for these aggregation switches, which further reduces the available bandwidth for the individual computing servers to 277 Mb/s. The bandwidth between the aggregation and core networks is distributed a multi-path routing technology is used for Equal Cost Multi-Path (ECMP) routing (Kantola 2011). The ECMP technique performs a
per-flow load balancing that differentiates the flows by computing a hash function on the incoming packet headers. For a three-tiered architecture, the maximum number of allowable ECMP paths bounds the total number of core switches to eight. Such a bound also limits the deliverable bandwidth to the aggregation switches. This limitation is overcome with the (commercial) availability of 100 GE links, standardized in June 2010. The important research topic is datacenter is an extremely designed. Fat-tree successors are constantly proposed for datacenters large-scale. Thus, the fact that not even single datacenter has been built (to this date) based on such proposals, we constrict the scope of this thesis to the three-tiered architecture. Nevertheless, it is taken into account that all types of datacenter topologies findings which remain valid researches. 5.10 Result Analysis For performance evaluation purposes, the proposed system was implemented in the Java. The simulator is designed to capture datacenter communication processes at the packet level. It implements a set of energy-efficient optimization techniques, such as DVFS and offers tools to monitor energy consumption in datacenter servers, switches, and other components. A three-tier tree datacenter topology consists of 1536 servers and arranged into 32 racks each holding 48 servers, served by 8 aggregation switches and 4 core (see Fig. 5.7), was used in all simulation experiments. We used 1 GE links for interconnecting servers in the inside racks while 10 GE links are used to form a fat-tree topology interconnecting access, core switches and aggregation. The propagation delay on all of the links was set to 10 ns. The workload generation events and user arrivals are exponentially distributed in time to mimic typical process. As soon as an arrival scheduling decision is taken for a newly sent the workload for execution to over the datacenter network and selected server. The size of the workload is equal to 15 KB. Being fragmented, it occupies 10 Ethernet
packets. While execution, to produce the workloads a constant bit rate stream of 1 Mb/s datacenter is directed. Such a stream is designed to mimic the behavior of the most common video sharing applications. To add uncertainties, during the execution, each workload communicates with another randomly chosen workload by sending 75 KB message internally. The size of the message is also sent out of the datacenter at the moment of task completion as an external communication. The schedulers server left by the idle powered down to reduce the power consumption while using dynamic power management technique. Another technique is used for unused network switches in access and aggregation network. The core network switches remain always operational at the full rate due to their crucial importance in communications. The evaluated datacenters are designed in such a way that the workload scheduler consolidates leaving the most (around 1016 on average) servers in idle mode. Those servers are powered down. Thus, performed server load is kept close to the maximum communication and congestion levels and no consideration of network. As a consequence, the designed scheduler produce a number of workloads that scheduled a combined load exceeding ToR switch forwarding capacity and causes network congestion. The round-robin scheduler follows a completely opposite policy. It distributes communicational and computing loads are equally among switches and servers; there by the network traffic is balanced and no server is overloaded. However, the drawback is that no server or network switch is left idle for powering down, making the round-robin scheduler as the least energy-efficient. The proposed technique achieves the workload consolidation for power efficiency while preventing computing servers and network switches from overloading. In fact, the average load of the rack switch uplink is around 0.95 and the average load of an operating server is around 0.9. Such load levels ensure that no additional delays in job communications are caused by network congestion. However, this advantage comes at a
price of a slight increase in the number of running servers. On average, the proposed scheduler left 956 servers as opposed to 1016 servers left idle by the Green scheduler. To explore the uplink load in detail, the traffic statistics at the most loaded ToR switches are measured. Under the proposed scheduler, the link is constantly overloaded and the queue remains almost constantly full, which causes multiple congestion losses. All queues were limited to 1000 Ethernet packets in our simulations. Under the proposed scheduler, the occupancy of buffer is mostly below the half of its size with an average of 213 packets. At certain instances of time the queue even remains empty having no packets to send. This fact results in a slightly reduced uplink utilization level of 0.95. The proposed system discussed is evaluated on java platform in 32 bit machine with windows OS of 1.84 GHz processor. The prototype design of DVFS protocol is layered into task-allocation, group creation, and processing elements. The virtual machines are developed for accepting number of task considering simulation time and processing speed of the system. The framework is developed to track down the communication in datacenter. The optimization is checked through JOULEX (See Appendix). The application server is taken for YOUTUBE as normally multimedia applications are heavy, frequently visited, and has sharing privileges for the users.
Figure 5.8. Joulex Interface Figure 5.9 Joulex interface with graphical representation of Power consumption The proposed system performs implementation of the study VFS scheme along with traditional DVS scheme, where consideration of both the schemes for energy optimization for the datacenters along with all its components is been considered. The considerations are as follows: No of server=1536 No. of Racks=32 No of application server by racks=48 Propagation delay on links=10 ns Experiment Task size=15 KB A communication link is created for connecting application server within the rack. The total set of work is exponentially dispersed to the executers
Figure 5.10. Server workload distribution performed by proposed system, previous DVFS implementation, and round-robin technique Figure 5.11 Joint uplink traffic loads Once the provisioning decision is undertaken for a recently designed task, it will be forwarded to the purpose of execution. At the processing mode, the task generates a uniform stream of bitrate of 1.6 Mb/s which is forwarded to the datacenter. To increase
the challenges at the simulation time, each of the tasks communicates with another task which is arbitrarily chosen by transmitting a dummy message of 125 KB implicitly. Also at the event of task accomplishment, the similar size packets will be forwarded to the datacenters. The mean task load in the datacenter is fixed to 45% which is dispersed among other application server. A round robin algorithm can be used for this purpose. Fig. 5.10 shows the result being accomplished from load distribution for all three of the evaluated schedulers e.g. previous DVFS system, proposed DVFS system, and conventional Round Robin. Fig. 5.11 shows a combined uplink load at the corresponding rack switches. The previous DVFS schema basically consolidates the workload leaving the most (around 1016 on average) servers idle in the evaluated datacenter. These storages areas are then subjected to power down. However, the load of the active servers is maintained very close to the highest and no consideration of network congestion levels and communication delays is performed at this moment. As a consequence, a quantity of workloads scheduled by the previous DVFS schema actually generates an integrated load exceeding the/ switch forwarding capacity and finally results in traffic congestion over cloud network. The proposed algorithm is experimented with the previously implemented DVFS techniques as well as conventional round robin algorithm. The comparative analysis is presented in fig 5.10 and fig 5.11. The Fig. 5.10 represents the load allocation to Server for all the analyzed provisionary and Fig 5.11 represents a Joint uplink traffic load. However, the round-robin scheduler follows a completely opposite policy for which it distributes computing and communicational loads equally among servers and switches; thereby the network traffic is balanced and no server is overloaded. However, the drawback is that no server or network switch is left idle for powering down, making the round-robin scheduler as the least energy-efficient in the proposed result analysis. The previous DVFS as well as proposed model is the most efficient schema for preserving power drainage in datacenters. The proposed schema is highly efficient to restore around 2/3 rd of the cloud servers along with network switches that potentially minimizes the score of unwanted power drainage. With the proposed system, the
datacenter energy consumption is reduced to half when compared to traditional and frequently used round-robin scheduler. The proposed technique when compared to the previous DVFS schema adds around: (a) 3.2 % to the total datacenter consumption, (b) 3% in servers energy consumption, and (c) 1.7 % in switches energy consumption. This minor increase in energy consumption is highly justified by the requirement of additional computing and communicational resources, detected by proposed technique, and required for keeping the quality of job execution at the desired level. In contrast to the previous DVFS schema, the proposed model basically deploys network awareness to detect congestion hotspots in the datacenter network and adjust its job consolidation methodology accordingly. It becomes particularly relevant for data intensive jobs which are constrained more by the availability of communication resources rather than by the available computing capacities. 5.11 Summary The proposed system identifies the function of transmission structure in the datacenter power utilization and highlights a unique process which amalgamates the power with network resources. Here, it also presents current challenges for achieving energy efficiency in local and regional datacenter systems. The power consumption of different layers of the datacenter is investigated and it is shown that there are great portions of power wasted both in non critical infrastructure and IT equipments. A review of the available metrics for energy efficiency was discussed and the effect of energy efficiency to carbon emissions and operational costs was computed. It is shown that there are great expectations for cost and carbon savings when the datacenter operates in a green manner. Strategies for developing and transforming a green datacenter were also presented based on operational and planning actions. It was shown that energy efficiency should target overall system optimization of both non critical and IT equipments with main focus placed on cooling, power delivery systems, virtualization and workload management. The proposed system maintains the equilibrium for the unique task
processing and accomplishment, power utilization of the datacenter, and optimal network requirements. The system optimizes the research gap for task grouping for reducing the quantity of the application server in the datacenter and dispersing of the network resources in order to mitigate the congestion which might possible occur in datacenter leading to unwanted power consumption.