1 HIGH RELIABILITY PRIVATE CLOUD STORAGE WITH MINIMAL OPERATION COST USING MULTI-OBJECTIVE OPTIMIZATION MUHD HANIS BIN HARUN A thesis submitted in fulfillment of the requirement for the award of the Master of Computer Science (Software Engineering) Faculty of Computer Science and Information Technology Universiti Tun Hussein Onn Malaysia APRIL 2014
2 ii ABSTRACT As far as internet connection brings people together to share idea and communicate through it, it continually invites service providers to share their resources with desired organizations in order to improve their business operations. Storage system used to store most of the organization s data needs to be accessed frequently. Thus, storage has become a vital part of an organization and needs to be available almost anytime. In order to maximize reliability and the availability of their data, many of organization converts their storage into a cloud storage platform. To set up a cloud server, a lot of consideration is needed. One of the main considerations is cost. Setting up a cloud storage without proper strategies will involve very high cost in future. In order to make sure the cloud storage server operates at maximum reliability in a certain amount of time, a high budget for servers is required. High server cost is required in order to obtain maximum reliability. Thus, setting up target duration will help organizations to control the cost without having to compromise the storage reliability. This research will focus on finding maximum server reliability based on target duration and cost constraint. In this thesis, maximum reliability is defined by using two main objectives which are budget and duration. The multi objective optimization technique is used in order to find the optimum point from the constraint. The optimal result will be the minimum number of servers required to achieve maximum reliability according to target duration.
3 iii ABSTRAK Saban hari talian internet membangun secara meluas membawa orang ramai untuk berkongsi idea dan berkomunikasi melaluinya, seterusnya menjemput pembekal perkhidmatan untuk berkongsi sumber mereka dengan organisasi ini untuk meningkatkan operasi perniagaan mereka. Sistem penyimpanan yang digunakan untuk menyimpan hampir keseluruhan maklumat organisasi dan perlu kepada akses yang lebih kerap. Oleh itu, pusat penyimpanan data telah menjadi bahagian yang penting dalam sesebuah organisasi dan perlu beroperasi hampir setiap masa. Bagi memaksimumkan kebolehpercayaan dan ketersediaan data mereka, kebanyakan organisasi menukar pusat penyimpanan data mereka ke platform Cloud. Dalam usaha untuk menubuhkan Cloud Server, banyak pertimbangan yang diperlukan. Salah satu pertimbangan utama adalah kos. Membina Cloud Server sendiri tanpa strategi yang betul akan melibatkan kos yang sangat tinggi pada masa hadapan. Bajet yang tinggi juga diperlukan bagi memastikan Cloud Server beroperasi pada kebolehpercayaan maksimum dalam jumlah masa tertentu. Kos server yang tinggi diperlukan untuk mendapatkan kebolehpercayaan maksimum. Oleh itu, menetapkan jangka masa sasaran akan membantu organisasi untuk mengawal kos tanpa perlu berkompromi dengan keupayaan Cloud Server. Dalam tesis ini, kebolehpercayaan maksimum ditentukan dengan menggunakan dua objektif utama iaitu bajet dan tempoh masa. Multi-objective optimization digunakan untuk mencari titik optimum dari kekangan yang ditetapkan. Hasil optimum akan bilangan minimum server yang diperlukan untuk mencapai kebolehpercayaan maksimum dalam jangka masa yang ditetapkan.
4 iv CONTENT ACKNOWLEDGEMENT ABSTRACT ABSTRAK CONTENT LIST OF TABLE LIST OF FIGURE i ii iii iv viii ix CHAPTER 1: INTRODUCTION Background Problem Statement Objective of Study Scope of Study Thesis Outline Summary 5
5 v CHAPTER 2: LITERATURE REVIEW Introduction Storage Reliability Previous Study Research Gap Failure Rate Failure Rate for Serial System Failure Rate for Parallel System Identical and Different Failure Rate Bathtub Curve Power Consumption Carbon Dioxide Emission Summary. 15 CHAPTER 3: METHODOLOGY Introduction Operational Framework Failure Rate Calculation Reliability Calculation Variable and Calculation Formula 20
6 vi 3.5 Multi Objective and Linear Prog Application Development Summary.. 23 CHAPTER 4: IMPLEMENTATION AND TESTING Introduction Program Implementation Flow Chart Flow Chart Explanation Data and Testing Element Expected Output Summary. 28 CHAPTER 5: RESULT AND DISCUSSION Introduction Optimization Result Testing Result Result Discussion Result for Target Duration Reliability after Target Duration. 36
7 vii 5.5 Summary. 37 CHAPTER 6: CONCLUSION AND RECOMMENDATION Introduction Contribution Future Work 39 REFERENCES 40
8 viii LIST OF TABLES 3.1 Variable for Reliability Calculation Input, Process, and Output for Application Expected Output Sample Data 29
9 ix LIST OF FIGURES 2.1 Serial Storage Connection Parallel Storage Connection Bathtub Curve Relation between Server Storage in Cloud Implementation Flow Chart Result Analysis Graph Comparison Analysis Optimization Result 34
10 CHAPTER 1 INTRODUCTION 1.1 Background Cloud Computing has turned to become one of the most popular and latest technology now days. The need of increasing processing power and efficient storage capacity influenced businesses to take part in cloud computing technology. As a general concept, we can imagine all business data are being stored in some place different from the organization. The data is replicated in a few forms as an assurance of 99.9% maximum uptime if anything happens to the server. This is in opposition to the distributed computing paradigm, where replication of data is the main issue to be solved. The term cloud computing is becoming quite synonyms with computing technology now days. The idea of cloud computing was coined by Leonard Kleinrock back then in 1969 where computer usage was likened to utilities. Even though, there are certain people working hard toward the idea, some computer users still do not have a clear view of what it is. What is cloud computing? Cloud computing refers to both the application delivered as a service over the internet and the hardware and systems software in the datacenters that
11 2 provide those services (Fox et al., 2009). By using the cloud, an organization does not have to deal with the technical issues on the infrastructure and does not have to invest the budget on buying and maintaining their vault of data. They will have to outsource their data to the cloud provider s server and need to pay for the size of data stored. They can also enlarge the storage size every time they need to. That is why the term pay-per-use was very synonymous with the cloud service. It is as if we have to pay for what we have used only. In general, there are three common service models offered in a cloud environment, namely: Application as a Service (AaaS), Platform as a Service (PaaS) and Infrastructure as a Service (IaaS) (Sourabh, 2012). As part of cloud computing, IaaS plays important features of its thriving compared to the other two services. This is because infrastructure serves as a third party provider for organization s hardware and operations. All of the company s confidential data will be outsourced to the provider and will reside outside their premises. The capability of a service provider to manage and handle huge amount of data is very crucial. At the same time, the providers have to manage their storage server and maintain the data to make sure it is available anytime when being requested by the organization. To maintain the storage reliability and the data availability, cloud offers the best framework with massive scalability, 99.99% reliability, high performance, and specific configuration by using multiple redundancy solution (Wu, 2010). 1.2 Problem Statement In terms of data availability, cloud offers a very good data reliability and performance as it offers 99.99% service up-time. In order to maintain the service up-time, it requires huge amount of storage capacity. Current cloud systems apply three replication strategies in order to achieve data availability. Such a 3-replica strategy means to store one gigabyte of data would need three gigabytes of data space and incur two times additional cost, which significantly affects the cost-effectiveness of the cloud system. In addition, such
12 3 huge resource allocation will indirectly increase the energy consumption to power up the storage server for 24/7 a week that will indirectly affect the operational cost. Organizations today tend to keep all of their business activity and data on an online server, which enables them and their customer to access the resource anytime, anywhere. Organization s size often affects the storage amount for their data. Most cloud storage providers nowadays provide various storage packages for consideration, but the main consideration is: in order for a certain amount of data to be stored for a particular period of time, how much storage would be required? Was the amount of storage able to maintain the reliability of storage in a certain amount of time? If it does, how long will it take before it changes to a failed state? Having additional servers to maintain storage reliability will increase the electricity consumption to power up and cooling down the server. According to Barett (2011) in his article on ehow.com, a server consumes electricity from 500 to 1200watt per hour. A server powered up by using natural gas, produced approximate 0.553kg carbon dioxide per hour (Carbon Fund, 2012). If an organization needs five servers to operate twentyfour hours a year to serve their customers for 10 years, the operational server itself will produce at least 242,214kg carbon dioxide. Even worst, twenty-four hours operated servers require a good cooling system that might double up the electricity usage. Therefore, reducing the number of servers without compromising the system reliability will save the service provider s operational cost and at the same time save the earth. 1.3 Objectives Of Study In order to achieve the research aim, few objectives has been set which are: I. To obtain highest reliability based on target and minimum storage based on the budget allocated. II. To solve the multi-objective problem using the linear programming method.
13 4 1.4 Scope Of Study Nowadays, large-scale cloud-based applications have put forward a higher demand for storage ability of data centers. Data in the cloud storage system needs to be stored with high efficiency and cost-effectiveness while meeting the requirement of reliability. In order to ensure data reliability in private cloud, cloud storage environment offers replication that sometimes is not needed by organizations. To achieve the costeffectiveness in storage infrastructure, strategies to minimize a storage system without compromising its reliability is the focus of this research. This will be the key factors of effective storage strategies in order to save high cost on large storage centers. The focus of this research is on the storage environment itself without considering the virtualization of the storage structure. 1.5 Thesis Outline Chapter 1 of this thesis explains on the research background, problem statements, the objectives and scope of research. Detail explanations on the problem statements will be discussed in Chapter 2 as well as the revision of related works, research gap, storage reliability, failure rate, power consumption and carbon dioxide. Research methodology in Chapter 3 will discuss in details on how to calculate system reliability by using failure rate and produce solutions for minimizing cost, but maximizing reliability, and apply effective solutions based on organization's constraint. Chapter 4 will discuss detail implementation of program, flow chart, codes as long as reliability calculation and expected output. Chapter 5 is discussing the details of testing results, and optimization in order to achieve minimal operation cost with maximum system reliability. Finally, Chapter 6 will conclude the entire thesis and provide recommendation for future work.
14 5 1.6 Summary Reliability is the main criteria for the cloud storage system. There are several ways to achieve system reliability, and some of them require high resource usage in term of operational cost and cost of server itself. Understanding system reliability is the main concern in order to be able to provide the solution to this problem. Next chapter will explain in detail about reliability and some of the previous research carried out by others.
15 CHAPTER 2 LITERATURE REVIEW 2.1 Introduction As more and more cloud-based applications tend to be data intensive, huge amount of data are stored in cloud infrastructure. The problem of managing the huge amount of data will require high amount of storage capacity and cost. Few researches are conducted to manage these issues. Most of them manage to reduce the redundancy by proposing a new method in term of backup and fault tolerance. Instead of providing fault tolerant backup, cloud storage could be optimized to reduce cost by managing data storage which requires huge storage size. This chapter will review some of techniques used by researchers to optimize reliability in cloud storage. At the end of this chapter, a strategy to manage storage reliability based on cost optimization will be proposed.
16 7 2.2 Storage Reliability A simple definition of storage is a place or locations where users save and manage data. In order to do such operations, users need to frequently access the storage device. In the internet era, the storage enables user to share its content through a network as a server. While the evolution of internet plays an important part in managing and improving businesses online, the server capability has become a great concern. As we know that the server stores vital organizational data which are being accessed frequently, hence the reliability of a server becomes a major concern. As we know that all electronic components in one way or another will exhaust its lifetime and therefore tend to crash by any factor. This keeps researchers working hard to find a solution to the problem. By implementing backups of current server with additional servers, they help organizations maintain the reliability of the cloud storage system. At the same time, the growths of data each day will also affect the cost of storage itself. Therefore, we found that, adding multiple backup servers without appropriate strategies is not a good solution to overcome storage reliability issue. In order to handle massive amounts of data with minimal cost, the appropriate storage reliability model needs to be implemented. Since the performance changes during the product's storage period, it has nothing to do with stress of storage servers. The main reasons responsible for the failure are the release of stress, which is created during the processing and assembling stages, the decay of materials, corrosion of electronic products, open and short circuit, and the material aging (Hong, 2012) 2.3 Previous Study Few frameworks and building blocks of cloud server architecture in the research paper by Bakshi (2011) serves as a guidance for enterprises and service providers before considering cloud data center. Major motivation of cloud storage implementation is to reduce operating cost and increase reliability of the system. In order to adopt cloud
17 8 architecture, the developer needs detailed explanations on how exactly cloud storage system is able to reduce the operational cost. The researcher groups the adoption steps into three phases which are: (1) Architecture development, (2) Migration and transition, and (3) Operation and optimization (Bakshi, 2011). The architecture development phase creates an accurate inventory of the application and data center related to the organization. The second phase applies to new data centers as well as redesign data centers and the third is focused on the operation and performance of the cloud system. The cloud s storage major consideration is about its reliability. Most of the cloud storage providers are likely to offer a 99.9 % uptime guarantee. This means, the system will be online and available for access twenty-four days a week without fail. In order to understand reliability in detail, few research papers were used as references. In a paper authored by Huang et al., (2011), redundancy scheme was classified into replication and erasure codes. Replication scheme is simple and intuitive which replicates the original data into the numbers n of that, then distribute them to the different network nodes. Erasure codes split the original data into k blocks and encode them into (n-k) blocks with n blocks in total, then distributed these n blocks into different nodes. The study on erasure codes implementation towards redundancy consumes smaller storage space with the same reliability as compared to replication technique (Huang et al., 2011). During the implementation process, this method takes several steps to complete. At the first stage, storage allocation will be built using generic function. Then it will be transformed into the problem to calculate the multiplication on several polynomials. After that an erasure codes, algorithm is applied to find the minimum redundancy but satisfying reliability. In order to evaluate the method, there are a few parameters which will be used. These are reliability for all nodes ( ṝ ), standard deviation of node s reliability ( ᵟ ), number of stored block s ( l ), number of blocks to reconstruct data ( k ), total number of blocks ( n ), power of the combinations ( p ), and cumulative data reliability for the power not more than p (CDR(p)). Based on the parameters, the impact of ṝ, ᵟ, and l toward data reliability and redundancy optimality is easily identified.
18 9 In the research proposed by Feng et al., (2012) he was more focused on finding a solution to make a trade-off between high reliability and low space overhead for cloud storage systems. High reliability and low space cost are always contradictory in existing storage systems. This research paper introduces Magicube, a storage architecture with high reliability and low space overhead for cloud computing. To reduce the space overhead of file storage, Magicube keeps only one copy of each file in HDFS (Feng et al., 2012), and to achieve high reliability, it uses a special encoding algorithm for fault-tolerance. Once the system parameters (n,k) have been set, the source file will be encoded and the chipper text will be split into n splits, then stores the corresponding file splits in the cluster (Feng et al., 2012). If the source file is lost, Magicube can repair the lost file from any k (k < n) of splits of that encoded file. It also allows users to adjust the system the system s faulttolerance level and space cost by changing the setting of (n,k). Based on the research paper, to achieve high reliability and low space coast, there are several methods that can be used, which are Triplication at the beginning, File splitting and distribution, Extra replication deletion and File repair. All experiments were carried out on a cluster of nine nodes, where a node served as a master and the others as client and slaves for MapReduces and HDFS (Feng et al., 2012). Each node has two quad-core 2.40GHz Xeon processors, 11GB of memory and two disks. The experiment result shows that the (n,k) coding introduced negligible influence on the performance of Magicube, and the file repairing performance was also acceptable. Meanwhile, LI et al., (2011) proposed a replication strategy that works especially for data that are used temporarily and/or have a relatively low reliability requirement. Based on the observation, the researcher found that the typical 3-replicas data replication strategy or any other fixed replica number replication strategy might not be the best solution for data with uncertain reliability duration. Most cloud systems contain a large number of tasks. During the execution, large volumes of intermediate data are generated and the amount could be much larger than the original data, but most of them are only aimed for temporary use. All the intermediate data are deleted after use or some of these will be stored for later use, but for an uncertain period of time (LI et al., 2011). The average failure rate of a single storage unit in most cloud infrastructure without
19 10 replication is 0.01%. Such reliability assurance and storage duration is sufficient to meet the requirement of most intermediate data in scientific applications without extra data replication. Thus, in order to reduce the storage cost and fully utilize the storage resource in current cloud systems, researchers propose a strategy named Cost-effective Incremental Replication (CIR). It is an incremental replication approach that works by calculating the replica creation time based on prediction, which indicates the storage duration to meet the reliability. To evaluate the performance of the proposed model, there are a few parameters which are being used, namely: number of data instance, storage unit number, reliability assurance, storage duration and cost. Based on the evaluation, the CIR strategy can reduce the data storage cost substantially, especially when the data is only stored for short duration or have a lower reliability requirement. 2.4 Research Gap The storage reliability of components is especially important to the electronic products which store for a long-term and are used at one time (Yang, 2011). Based on the previous studies, the main issues in the research are reliability and cost. Most of the researches focus on finding minimal redundancy using erasure code, increasing the storage reliability from time to time by creating an additional replica to current nodes. By doing these, redundancy issues are yet to be resolved because the existing storage size will keep on increasing over time. What will happen to the existing storage nodes if it failed to operate in a certain amount of time? Moving replicas from the existing server to another is a good solution as a backup, but for how long will the recovery process take to ensure the reliability of the storage. From the previous study, the reliability of the storage system is Rs = 1 qc, where q=1-p and p are the reliable server, c is the number of servers existing in the system. By using this formula, we can assume the reliability of a storage system is set to 1 qc, where q is the unreliability of the storage which is the system failure rate. In cloud storage system features, each of the data is stored using minimum 3-replication strategies, which mean
20 11 the reliability of the storage is equal to 1 q3 reliability. The basic storage rules provide good data access to the organization, but it will become insufficient if the data keep increasing from time to time. If a certain amount of data is to be stored in a certain amount of time, the exact number of storage system needs to minimize the server cost. Most applicable methods to improve storage reliability and availability is by using redundancy scheme. Redundancy scheme is divided two classes: Replication and Erasure codes. Replication works by replicating original data into n copies and distribute to different nodes in or outside the network while erasure codes split the original data into k blocks, and encode them into (n-k) blocks with n blocks in total, then distribute these n blocks into different network nodes. From this solution, reliability equation Rs=1-qc will be used to i. determine how many storage will be used, ii. iii. iv. predict how long it will be operated until it gets to the next failure state, how long it will take to recover and how many storage nodes are required to ensure reliability. 2.5 Failure Rate Failure rate (λ) is a number of failures per unit depending on operating hours. Each of storage devices consists of a failure rate during its operation, which is the number of units failing per time. Therefore, the failure rate is used to measure the reliability of a storage based on the probability of the storage itself. In the meantime, lower failure rates indicate higher reliability of storage. According to Rooney (1989), failure rates of storage are estimated at 2.64 and up to 5.90 failures per million hours. Failure rates determine the reliability of a storage framework to which it displays the frequency of failure a device undergoes within a specific period (Torell, 2004). In a server, system failure rate is being identified according to the position of the storage,
21 12 whether they are stacked in serial or parallel order. Another factor that directly affects the system reliability is the failure rate of each of the storage contained in the system. Each of the storage device contained in the system might have different failure rate or might not i.e. the system might contain identical or different storage failure rate Failure Rate for Serial System Serial systems work with two or more storage combination operating in series. The failure of either storage will result in failure of the combination. The failure rate of combined storage in the series is shown as follows: Rs = FR A * FR B (2.1) Storage A Figure 2.1: Serial Storage Connection Failure Rate for Parallel System Parallel systems work with two or more storage combination operating in side by side in parallel. The system remains operable unless all storages fail. The failure rate of parallel system shows as follows:
22 13 Rs = 1 (1 FR A )(1 FR B ) (2.2) Storage A Storage B Figure 2.2: Parallel Storage Connection Identical and Different Storage Failure Rate Cloud servers consist of multiple storage units, and in order to identify its reliability, the failure rate of the system is calculated by finding the failure rate of each storage unit. A set of storage units in a server that have the same failure rate is called identical failure rate, and a set of storage that have varying failure rates is called different failure rate. 2.6 BATHTUB CURVE Every single device comes with a lifespan. Failure rates and time are used to determine the reliability of a device during its life span. Normal life span for electronic device can be divided into three distinct periods (Klutke, 2003). Bathtub curve is used to show how the failure rates vs. time for each of an electronic device to help users predict the actual failure before it is happening. This helps a lot in managing user access and retrieval to the device.
23 14 Figure 2.3: Bathtub Curve (Klutke, 2003) The bathtub curve consists of three periodic stages which are infant mortality period (first stage) with a decreasing failure rate followed by normal operating stage period with a low, relatively constant failure rate and concluding with aging stage that exhibits an increasing failure rate. Bathtub curve does not depict the failure rate of a single device, but describes the relative failure rate of an entire population of devices over time. Some individual units will fail relatively early, the other will last until wearout and some will fail during normal operating state. 2.7 Power Consumption High-end servers use a lot of electricity to operate, and about the same is used to cool it down. Using electricity without proper control will affect global warming since the emission of carbon dioxide (CO2) from electric device is high especially a computer server with non-stop operation. Before developing a storage system, an organization needs to manage and study the system reliability prediction in order to minimize the power consumption.
24 Carbon Dioxide Emission Every electric device emits carbon dioxide (CO 2 ) for every kilowatts of usage. According to Barett (2011) in his article on ehow.com, a server system consumes electricity from 500 to 1200 watts per hour. According to Carbon Fund (2012), on average electricity usage emits 1.222lbs per kilowatts (kwh) or approximate kg per kwh. If on the average, a server used 850 watts per hour, and about 7,446 kilowatts per year, that means more than 4120kg carbon dioxide (CO 2 ) is released into the atmosphere. In order to reduce the carbon dioxide emissions, less storage system is required without compromising the system reliability Summary In next chapter, we will discuss about the strategies to minimize storage cost by reducing the number of server but maintaining the reliability of data. We will be focusing on the mean time between failure as well as mean time to repair and suggest a tradeoff between costs and performance based on the needs of the organization.
25 CHAPTER 3 METHODOLOGY 3.1 Introduction In order to determine storage reliability and minimal operation cost, several factors need to be considered. This research focuses on how organizations can obtain storage reliability using the allocated budget. Instead of having budget as the key point to the system reliability and optimization, few variables such as failure rate, number of server and duration of the storage system are used. The research methodology is used to identify the reliability formula based on previous study and analyzes the appropriateness to organizational requirements based on current cloud storage pricing structure. The research also focuses on identifying the minimum number of servers required based on the budget allocated.
26 Operational Framework The aim of this research is to get high reliability for cloud storage framework subject to minimal cost allocated. In order to achieve the objective, multi-objective optimization will be used to solve multi criteria decision making involving more than one objective function to be optimized simultaneously. At the end of this research, a table consisting of comparison from the selected variable will be published as analysis result and determine the best storage implementation depending on the number of servers required to maximize the system reliability. In order to maintain its reliability, cloud storage uses replication strategies to ensure that all of its servers are operating using the latest updated data whenever one or more of its server fails. To avoid wasting storage resource during replication strategies, an exact calculation on estimation of when and how many times the server will fail in certain period is needed. To achieve these objectives, storage failure rates are used. Figure 3.1shows the relation between servers in cloud storage: Storage C 1 q 1 Storage C 2 q 2 Storage C 3 q 3 Figure 3.1: Relation between Servers in Cloud Storage
27 18 Figure 3.1 indicates that in a cloud storage framework consist of three servers c. Each of the servers has a reliability of q. The overall storage framework reliability is represented as Rs = 1 q 3, that is achieved from the equation of R = 1-q 1 q 2 q 3 or R = 3 1 i=1 In order to determine the reliability of the storage, unreliability of each of the storage will have to be identified. The unreliability of each of the storage is calculated using failure rate (λ) or q. q i (3.1) 3.3 Failure Rate Calculation The failure rate of a storage device can be found from the hardware itself, which is provided by the device manufacturer. Failure rate enable user to identify how reliable the device is based on number of predicted fail per certain of times. For example, a device having 0.2%/1000 hours, which means the device will fail to operate two times within one million hours. This can be expressed on the following equation: q = * = (3.2) From the failure based on failure rate per 1000 hours, the probability of failures per year could be calculated as: q = * * 24*365 = Reliability Calculation Based on the failure rate, the reliability of a device with 0.2% of failure rate per 1000 hours can be express as:
28 19 Rs = 1 q Rs = Rs = (3.3) That means, if a device with such failure rate operates 24 hours a day for one year, the probability of a device will be reliable is 98%. In order to increase the reliability to 99%, one additional server with such failure rate is required. Rs = 1 q Rs = 1 ( ) Rs = The reliability will decrease by 1% for every server over a year and will require additional server to maintain the reliability. With minimum two servers, the storage reliability, maintained at 99% at least for five years. If an organization need to set up a storage system using three servers for the duration of three years, using the previous failure rate for each server, the following calculation applies. Year 1: Failure rate: q = * * 24 *365 = Reliability: Rs = 1 q Rs = 1 ( ) 3 3 Rs = Year 2: Failure rate: q = * *24*365*2 =
29 20 Reliability: Rs = 1 q Rs = 1 ( ) Rs = Year 3: Failure rate: q = * *24*365*3 = Reliability: Rs = 1 q Rs = 1 ( ) Rs = Variable and Calculation Formula In order to identify the reliabilities of a storage system for certain duration, there are few variable are used. The variable listed as follows: Table 3.1: Variable for Reliability Calculation Variable n h d y Description number of server constant number of hours server operate (24 hours) Constant number of the day server operates (365 days) total number of years to calculate the server reliability
30 21 mh million time is set to 10-6 nf number of failed per million hours b budget of the organization Sc cost per server Mc Minimum cost Tr Target reliability Rs System reliability Number of possible server (n) will be calculated from the organization budget (b) and cost per server (Sc) based on market price. (3.4) n = b/sc Reliability calculation formula Rs = 1-[(nF*mh)*(h*d*y)] n (3.5) Reliability of the system will be calculate by using the formula to identify what is the reliability of the storage based on variables that were listed previously. The result will be displayed is the system reliability (%) based on year and the number of storage used. 3.5 Multi Objective and Linear Programming In order to find system reliability, there are several objectives used as the constraint. Thus, the multi-objective optimization is used. Multi-objective optimization involves more than one objective function in order to optimize the result (Deb, 2014). Multi objective optimization is used to find minimum and maximum value from multi variable constraint. In this thesis, reliability is calculated using the constraint from cost and duration to find maximum reliability with lowest cost allocated.
31 22 The main objective of this research is to provide reliability solutions for organizations if they want to create cloud storage environment. For this purpose, several constraints are count for calculation. The focus is to find maximum server reliability based on target reliability and the minimum numbers of server involve in order of achieving maximum reliability in target duration. Number of server involve is relate to budget allocated. In order to achieve result base on several objective, optimization technique is required. In this problem, linear programming is use to find the maximum and minimum point in each calculation The following formula shows the relation of multi objective variable based on constraint below: Let, i is the number of server, q = 1 p is the probability of failure, The maximum reliability is given as, n Maximum Rs = 1 = q ) ( i 1 i (3.6) Subject to: n i=1 and i budget n i= 1 i duration
32 Application Development Based on reliability calculation formula, we were able to calculate the system reliability for a given number of servers and duration. In order to find the system reliability based on certain duration, an application is needed to automate the calculation. 3.7 Summary Research methodology is the most important part in which we will provide the guideline and procedure on how to get the data and verify what to do in the next chapters. With all variables used, input, expected result and appropriate steps were being explained within this chapter. This will make the implementation phase come in handy. Details of application to calculate the system reliability is discussed in the next chapter.
34 40 REFERENCES Bakshi, K. (2011, March). Considerations for cloud data centers: Framework, architecture and adoption. In Aerospace Conference, 2011 IEEE (pp. 1-7). IEEE. Borthakur, D. (2007). The hadoop distributed file system: Architecture and design. Carbon Fund (2012), How We Calculate, Retrieve October 13, 2013, from URL: Deb, K. (2014). Multi-objective optimization. In Search methodologies (pp ). Springer US. Feng, Q., Han, J., Gao, Y., & Meng, D. (2012, June). Magicube: High Reliability and Low Redundancy Storage Architecture for Cloud Computing. In Networking, Architecture and Storage (NAS), 2012 IEEE 7th International Conference on(pp ). IEEE. Fox, A., Griffith, R., Joseph, A., Katz, R., Konwinski, A., Lee, G.,... & Stoica, I. (2009). Above the clouds: A Berkeley view of cloud computing. Dept. Electrical Eng. and Comput. Sciences, University of California, Berkeley, Rep. UCB/EECS, 28. Hong, W., Wang, S., Wang, X., & Wang, W. (2012, July). Overview of storage reliability for high reliability products. In Industrial Informatics (INDIN), th IEEE International Conference on (pp ). IEEE.
35 41 Huang, Z., Yuan, Y., & Peng, Y. (2011, May). Storage allocation for redundancy scheme in reliability-aware cloud systems. In Communication Software and Networks (ICCSN), 2011 IEEE 3rd International Conference on (pp ). IEEE. J.T Barett (2011), How Much Electricity Does a Computer Use Per Hour?, Retrieve October 13, 2013, from URL: Klutke, G. A., Kiessler, P. C., & Wortman, M. A. (2003). A critical look at the bathtub curve. Reliability, IEEE Transactions on, 52(1), Li, W., Yang, Y., & Yuan, D. (2011, December). A novel cost-effective dynamic data replication strategy for reliability in cloud data centres. In Dependable, Autonomic and Secure Computing (DASC), 2011 IEEE Ninth International Conference on (pp ). IEEE. Rooney, J. P. (1989, January). Storage reliability. In Reliability and Maintainability Symposium, Proceedings., Annual (pp ). IEEE. Sourabh (2012), What is Cloud Computing, Retrieve March 10, 2013 from Torell, W., & Avelar, V. (2004). Mean time between failure: Explanation and standards. White Paper, 78. Wu, J., Ping, L., Ge, X., Wang, Y., & Fu, J. (2010, June). Cloud storage as the infrastructure of cloud computing. In Intelligent Computing and Cognitive Informatics (ICICCI), 2010 International Conference on (pp ). IEEE. Yang, Y., Fu, G., Wan, B., & Gu, H. (2011, May). Component storage reliability evaluation method based on multi-performance parameters degradation analysis. In Prognostics and System Health Management Conference (PHM-Shenzhen), 2011 (pp. 1-6). IEEE.
DEVELOP AND DESIGN SHEMATIC DIAGRAM AND MECHANISM ON ONE SEATER DRAG BUGGY MUHAMMAD IBRAHIM B MD NUJID A report in partial fulfillment of the requirements For award of the Diploma of Mechanical Engineering
HELP DESK SYSTEM IZZAT HAFIFI BIN AHMAD ARIZA A thesis submitted in fulfillment of the requirement for the awards of Bachelor of Computer Science (Computer Systems & Networking) with Honours Faculty of
DEVELOPING AN ISP FOR HOTEL INDUSTRY: A CASE STUDY ON PUTRA PALACE HOTEL A report submitted to the Graduate School in partial fulfillment of the requirement for the Degree Master of Science (Information
JOB AGENT MANAGEMENT SYSTEM LU CHUN LING A thesis submitted in partial fulfillment of the requirements for the award of the degree of Bachelor of Computer Sciences (Software Engineering) FACULTY OF COMPUTER
EXPERIMENTAL ANALYSIS OF PASSIVE BANDWIDTH ESTIMATION TOOL FOR MULTIPLE HOP WIRELESS NETWORKS NURUL AMIRAH BINTI ABDULLAH THESIS SUBMITTED IN FULFILLMENT OF THE DEGREE OF COMPUTER SCIENCE FACULTY OF COMPUTER
LIGHTNING AS A NEW RENEWABLE ENERGY SOURCE SARAVANA KUMAR A/L ARPUTHASAMY UNIVERSITI TEKNOLOGI MALAYSIA LIGHTNING AS A NEW RENEWABLE ENERGY SOURCE SARAVANA KUMAR A/L ARPUTHASAMY A project report submitted
2011 Ninth IEEE Ninth IEEE International Conference on Dependable, Autonomic and Secure Computing A Novel Cost-effective Dynamic Data Replication Strategy for Reliability in Cloud Data Centres Wenhao LI
ii CLIENT SERVER APPLICATION FOR SERVER FARM PERFORMANCE MONITORING ABDIRASHID HASSAN ABDI A project submitted in partial fulfillment of the requirements for the award of the degree of Master of Computer
SURVEY ON RAID Aishwarya Airen 1, Aarsh Pandit 2, Anshul Sogani 3 1,2,3 A.I.T.R, Indore. Abstract RAID stands for Redundant Array of Independent Disk that is a concept which provides an efficient way for
Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture He Huang, Shanshan Li, Xiaodong Yi, Feng Zhang, Xiangke Liao and Pan Dong School of Computer Science National
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT 1 SARIKA K B, 2 S SUBASREE 1 Department of Computer Science, Nehru College of Engineering and Research Centre, Thrissur, Kerala 2 Professor and Head,
Car Rental Management System (CRMS) Lee Chen Yong This report is submitted in partial fulfillment of the requirement for the Bachelor of Computer Science (Database Management) FACULTY OF INFORMATION AND
MASTER S PROJECT REPORT SUMMARY LEVEL OF SERVICE (LOS) FOR MULTILANE HIGHWAY AND ROAD ACCIDENT INFORMATION SYSTEM DEVELOPMENT OF BATU PAHAT AREA (LORIS) Prepared by: Mohd Ezree Bin Abdullah Master of Engineering
i INTEGRATING CONSUMER TRUST IN BUILDING AN E-COMMERCE WEBSITE NUR ZAILAN BIN OTHMAN A project report submitted in partial fulfillment of the requirements for the award of the degree of Master of Science
STRESS EFFECT STUDY ON 6 DIFFERENT PATTERN OF TYRES FOR SIZE 175/70 R13 SYAHRIL AZEEM ONG BIN HAJI MALIKI ONG A report submitted in partial fulfilment of the requirements for the award of the degree of
IMPROVING SERVICE REUSABILITY USING ENTERPRISE SERVICE BUS AND BUSINESS PROCESS EXECUTION LANGUAGE AKO ABUBAKR JAAFAR A project report submitted in partial fulfillment of the requirements for the award
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
PRODUCTIVITY IMPROVEMENT VIA SIMULATION METHOD (MANUFACTURING INDUSTRY) HASBULLAH BIN MAT ISA Thesis submitted in fulfillment of the requirements for the award of the degree of Bachelor of Mechanical Engineering
A Dynamic Resource Management with Energy Saving Mechanism for Supporting Cloud Computing Liang-Teh Lee, Kang-Yuan Liu, Hui-Yang Huang and Chia-Ying Tseng Department of Computer Science and Engineering,
SPAM FILTERING USING BAYESIAN TECHNIQUE BASED ON INDEPENDENT FEATURE SELECTION MASURAH BINTI MOHAMAD A project report submitted in partial fulfillment of the requirements for the award of the degree of
RESEARCH ARTICLE OPEN ACCESS Distributed Framework for Data Mining As a Service on Private Cloud Shraddha Masih *, Sanjay Tanwani** *Research Scholar & Associate Professor, School of Computer Science &
Introduction to Cloud Computing Cloud Computing I (intro) 15 319, spring 2010 2 nd Lecture, Jan 14 th Majd F. Sakr Lecture Motivation General overview on cloud computing What is cloud computing Services
Data Integrity Check using Hash Functions in Cloud environment Selman Haxhijaha 1, Gazmend Bajrami 1, Fisnik Prekazi 1 1 Faculty of Computer Science and Engineering, University for Business and Tecnology
Dynamic Load Balancing: Improve Efficiency in Cloud Computing Argha Roy * M.Tech CSE Netaji Subhash Engineering College West Bengal, India. Diptam Dutta M.Tech CSE Heritage Institute of Technology West
Load Balancing in Fault Tolerant Video Server # D. N. Sujatha*, Girish K*, Rashmi B*, Venugopal K. R*, L. M. Patnaik** *Department of Computer Science and Engineering University Visvesvaraya College of
Enhancing Dataset Processing in Hadoop YARN Performance for Big Data Applications Ahmed Abdulhakim Al-Absi, Dae-Ki Kang and Myong-Jong Kim Abstract In Hadoop MapReduce distributed file system, as the input
A Survey on Aware of Local-Global Cloud Backup Storage for Personal Purpose Abhirupa Chatterjee 1, Divya. R. Krishnan 2, P. Kalamani 3 1,2 UG Scholar, Sri Sairam College Of Engineering, Bangalore. India
ENERGY-EFFICIENT TASK SCHEDULING ALGORITHMS FOR CLOUD DATA CENTERS T. Jenifer Nirubah 1, Rose Rani John 2 1 Post-Graduate Student, Department of Computer Science and Engineering, Karunya University, Tamil
CHAPTER 1 INTRODUCTION 1.1 Background The command over cloud computing infrastructure is increasing with the growing demands of IT infrastructure during the changed business scenario of the 21 st Century.
Creating an IT Infrastructure that Adapts to Your Business PLAYBOOK F O R C H A N G E For decades, data centers have been over-provisioned two or even three times over in an attempt to plan for growth.
COST MINIMIZATION OF RUNNING MAPREDUCE ACROSS GEOGRAPHICALLY DISTRIBUTED DATA CENTERS Ms. T. Cowsalya PG Scholar, SVS College of Engineering, Coimbatore, Tamilnadu, India Dr. S. Senthamarai Kannan Assistant
Investigation of Cloud Computing: Applications and Challenges Amid Khatibi Bardsiri Anis Vosoogh Fatemeh Ahoojoosh Research Branch, Islamic Azad University, Sirjan, Iran Research Branch, Islamic Azad University,
Cloud Computing and Amazon Web Services Gary A. McGilvary edinburgh data.intensive research 1 OUTLINE 1. An Overview of Cloud Computing 2. Amazon Web Services 3. Amazon EC2 Tutorial 4. Conclusions 2 CLOUD
Big Data Storage Architecture Design in Cloud Computing Xuebin Chen 1, Shi Wang 1( ), Yanyan Dong 1, and Xu Wang 2 1 College of Science, North China University of Science and Technology, Tangshan, Hebei,
1 Apache Hadoop Alexandru Costan Big Data Landscape No one-size-fits-all solution: SQL, NoSQL, MapReduce, No standard, except Hadoop 2 Outline What is Hadoop? Who uses it? Architecture HDFS MapReduce Open
EFFICIENT AND SECURE DATA PRESERVING IN CLOUD USING ENHANCED SECURITY Siliveru Ashok kumar* S.G. Nawaz ## and M.Harathi # * Student of M.Tech, Sri Krishna Devaraya Engineering College, Gooty # Department
International Journal Of Scientific Research And Education Volume 2 Issue 10 Pages-2109-2115 October-2014 ISSN (e): 2321-7545 Website: http://ijsae.in ABSTRACT: Database Management System as a Cloud Computing
Journal homepage: http://www.journalijar.com INTERNATIONAL JOURNAL OF ADVANCED RESEARCH RESEARCH ARTICLE CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING R.Kohila
Non-Stop Hadoop Paul Scott-Murphy VP Field Techincal Service, APJ Cloudera World Japan November 2014 WANdisco Background WANdisco: Wide Area Network Distributed Computing Enterprise ready, high availability
High Availability of VistA EHR in Cloud ViSolve Inc. White Paper February 2015 1 Abstract Inspite of the accelerating migration to cloud computing in the Healthcare Industry, high availability and uptime
CSE-E5430 Scalable Cloud Computing P Lecture 5 Keijo Heljanko Department of Computer Science School of Science Aalto University firstname.lastname@example.org 12.10-2015 1/34 Fault Tolerance Strategies for Storage
A STUDY ON MOTIVATION TO START UP A BUSINESS AMONG CHINESE ENTREPRENEURS A master project submitted to the Graduate School in partial fulfillment of the requirements for the degree Master of Business Administration,
CLOUDDMSS: CLOUD-BASED DISTRIBUTED MULTIMEDIA STREAMING SERVICE SYSTEM FOR HETEROGENEOUS DEVICES 1 MYOUNGJIN KIM, 2 CUI YUN, 3 SEUNGHO HAN, 4 HANKU LEE 1,2,3,4 Department of Internet & Multimedia Engineering,
Coding Techniques for Efficient, Reliable Networked Distributed Storage in Data Centers Anwitaman Datta Joint work with Frédérique Oggier (SPMS) School of Computer Engineering Nanyang Technological University
Designing a Cloud Storage System End to End Cloud Storage When designing a cloud storage system, there is value in decoupling the system s archival capacity (its ability to persistently store large volumes
, pp.11-15 http://dx.doi.org/10.14257/astl.2014.53.03 UPS battery remote monitoring system in cloud computing Shiwei Li, Haiying Wang, Qi Fan School of Automation, Harbin University of Science and Technology
Dynamic resource management for energy saving in the cloud computing environment Liang-Teh Lee, Kang-Yuan Liu, and Hui-Yang Huang Department of Computer Science and Engineering, Tatung University, Taiwan
Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware Created by Doug Cutting and Mike Carafella in 2005. Cutting named the program after
Resource Scalability for Efficient Parallel Processing in Cloud ABSTRACT Govinda.K #1, Abirami.M #2, Divya Mercy Silva.J #3 #1 SCSE, VIT University #2 SITE, VIT University #3 SITE, VIT University In the
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
DESIGN OF A RECYCLE BIN TIN CAN CRUSHER MUHAMMAD HANIS BIN MUHAMMAD ZULKIFLI A report submitted in partial fulfillment of the requirements for the award of the degree of Diploma of Mechanical Engineering
Improving Apriori Algorithm to get better performance with Cloud Computing Zeba Qureshi 1 ; Sanjay Bansal 2 Affiliation: A.I.T.R, RGPV, India 1, A.I.T.R, RGPV, India 2 ABSTRACT Cloud computing has become
AN EFFICIENT LOAD BALANCING APPROACH IN CLOUD SERVER USING ANT COLONY OPTIMIZATION Shanmuga Priya.J 1, Sridevi.A 2 1 PG Scholar, Department of Information Technology, J.J College of Engineering and Technology
A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems Anton Beloglazov, Rajkumar Buyya, Young Choon Lee, and Albert Zomaya Present by Leping Wang 1/25/2012 Outline Background
International Journal of Electronics and Computer Science Engineering 2404 Available Online at www.ijecse.org ISSN- 2277-1956 Optimized and Secured Educlouds by Implementing Virtualization R. Rajesh 1,
JIOS, VOL. 35, NO. 1 (2011) SUBMITTED 02/11; ACCEPTED 06/11 UDC 004.75 Comparison of Cloud vs. Tape Backup Performance and Costs with Oracle Database University of Ljubljana Faculty of Computer and Information
ISSN:2320-0790 Dynamic Data Replication for HPC Analytics Applications in Hadoop Ragupathi T 1, Sujaudeen N 2 1 PG Scholar, Department of CSE, SSN College of Engineering, Chennai, India 2 Assistant Professor,
Introduction to Cloud : Cloud and Cloud Storage Lecture 2 Dr. Dalit Naor IBM Haifa Research Storage Systems 1 Advanced Topics in Storage Systems for Big Data - Spring 2014, Tel-Aviv University http://www.eng.tau.ac.il/semcom
Minimization of Energy Consumption Based on Various Techniques in Green Cloud Computing Jaswinder Kaur 1, Sahil Vashist 2, Rajwinder Singh 3, Gagandeep Singh 4 Student, Dept. of CSE, Chandigarh Engineering
Efficient Scheduling Of On-line Services in Cloud Computing Based on Task Migration 1 Harish H G, 2 Dr. R Girisha 1 PG Student, 2 Professor, Department of CSE, PESCE Mandya (An Autonomous Institution under
GraySort and at Yahoo on Hadoop.23 Thomas Graves Yahoo! May, 213 The Apache Hadoop software library is an open source framework that allows for the distributed processing of large data sets across clusters
i TRANSFORMATIONAL PROJECT MANAGER: AN ENABLER OF AN ENTERPRISE RESOURCE PLANNING (ERP) IMPLEMENTATION SUCCESS JOHN ONYEKACHI OKUGO A dissertation submitted in partial fulfillment of the requirements for
An Energy Efficient Server Load Balancing Algorithm Rima M. Shah 1, Dr. Priti Srinivas Sajja 2 1 Assistant Professor in Master of Computer Application,ITM Universe,Vadodara, India 2 Professor at Post Graduate
INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK IMPLEMENTATION OF AN APPROACH TO ENHANCE QOS AND QOE BY MIGRATING SERVICES IN CLOUD
Role of Cloud Computing in Big Data Analytics Using MapReduce Component of Hadoop Kanchan A. Khedikar Department of Computer Science & Engineering Walchand Institute of Technoloy, Solapur, Maharashtra,
Volume 5, Issue 7, July 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Configuration Tuning
Heterogeneous Workload Consolidation for Efficient Management of Data Centers in Cloud Computing Deep Mann ME (Software Engineering) Computer Science and Engineering Department Thapar University Patiala-147004
Chapter 7 Using Hadoop Cluster and MapReduce Modeling and Prototyping of RMS for QoS Oriented Grid Page 152 7. Using Hadoop Cluster and MapReduce for Big Data Problems The size of the databases used in
i WEB-BASED PROPERTY MANAGEMENT SYSTEM SAFURA ADEELA BINTI SUKIMAN A thesis submitted in fulfillment of the requirements for the award of the degree of Masters of Science (IT Management) Faculty of Computer
Multifunctional Barcode Inventory System for Retailing. Are You Ready for It? Ling Shi Cai, Leau Yu Beng, Charlie Albert Lasuin, Tan Soo Fun, Chin Pei Yee Abstract This paper explains the development of
Log Mining Based on Hadoop s Map and Reduce Technique ABSTRACT: Anuja Pandit Department of Computer Science, email@example.com Amruta Deshpande Department of Computer Science, firstname.lastname@example.org
Volume 5, Issue 1, January 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Survey on Load
Improving Job Scheduling in Hadoop MapReduce Himangi G. Patel, Richard Sonaliya Computer Engineering, Silver Oak College of Engineering and Technology, Ahmedabad, Gujarat, India. Abstract Hadoop is a framework
Introduction to Cloud Computing Srinath Beldona email@example.com Agenda Pre-requisites Course objectives What you will learn in this tutorial? Brief history Is cloud computing new? Why cloud computing?
Setting deadlines and priorities to the tasks to improve energy efficiency in cloud computing Problem description Cloud computing is a technology used more and more every day, requiring an important amount
Managing the Real Cost of On-Demand Enterprise Cloud Services with Chargeback Models A Guide to Cloud Computing Costs, Server Costs, Pricing Plans, and Chargeback Implementation and Systems Introduction
Assistance of Cloud Computing For Green Computing Rashmi R. Rathi Department of Masters in Computer Application Prof.Ram Meghe Institute Of Technology And Research, Badnera,Amravati firstname.lastname@example.org
Server Consolidation for SAP ERP on IBM ex5 enterprise systems with Intel Xeon Processors: Lowering Total Cost of Ownership An Alinean White Paper Published by: Alinean, Inc. 201 S. Orange Ave Suite 1210
Introducing the Singlechip Cloud Computer Exploring the Future of Many-core Processors White Paper Intel Labs Jim Held Intel Fellow, Intel Labs Director, Tera-scale Computing Research Sean Koehl Technology
WHITE PAPER Advanced Core Operating System (ACOS): Experience the Performance Table of Contents Trends Affecting Application Networking...3 The Era of Multicore...3 Multicore System Design Challenges...3
Oracle Database Backup Service Secure Backup in the Oracle Cloud Today s organizations are increasingly adopting cloud-based IT solutions and migrating on-premises workloads to public clouds. The motivation
Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com Image
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next