Distributed Storage Evaluation on a Three-Wide Inter-Data Center Deployment
|
|
|
- Shannon Kennedy
- 10 years ago
- Views:
Transcription
1 Distributed Storage Evaluation on a Three-Wide Inter-Data Center Deployment Yih-Farn Chen, Scott Daniels, Marios Hadjieleftheriou, Pingkai Liu, Chao Tian, Vinay Vaishampayan AT&T Labs-Research, Shannon Laboratory, 18 Park Ave., Florham Park, NJ {chen,daniels,marioh,pingkai,tian,vinay}@research.att.com Abstract The demand for cloud storage is exploding as an ever increasing number of enterprises and consumers are storing and processing their data in the cloud. Hence, distributed object storage solutions (e.g., Tahoe-LAFS, Riak, Swift, HDFS) are becoming very critical components of any cloud infrastructure. These systems are able to offer good reliability by distributing redundant information across a large number of commodity servers, easily achieving up to 1 nines of reliability. One drawback of these systems is that they are usually designed for deployment within a single data center, where node-to-node latencies are small. Geo-replication (i.e., distributing redundant information across data centers) for most open-source storage systems is, to the best of our knowledge, accomplished by asynchronously mirroring a given deployment. Given that georeplication is critical for ensuring very high degrees of reliability (e.g., for achieving 1 nines), in this work we evaluate how these storage systems perform when they are directly deployed in a WAN setting. To this end, three popular distributed object stores, namely Quantcast-QFS, Swift and Tahoe-LAFS, are considered and tested on a three-wide data center environment and our findings are reported. I. INTRODUCTION Modern distributed object storage solutions, like HDFS[1], Swift[2], Quantcast-QFS[3], Tahoe-LAFS[4], Riak[5], Azure[], Colossus[7], and Amazon Dynamo[8], offer very good availability and reliability at a low price point by distributing data across a very large set of inexpensive commodity servers that consist of unreliable components. Despite their success, most of these systems, to the best of our knowledge, have been designed to distribute data across large clusters of servers redundantly (either by using replication or erasure coding) within a single data center. Geo-replication, i.e., distribution of redundant information across data centers, is typically handled by asynchronously mirroring a given deployment. There are several limiting factors that affect the performance of distributed object store deployments in a WAN setting (in the presence of TCP slow start and shared bandwidth), as will become clear after an in-depth description of these systems in Section II. In this work, we evaluate three distributed object stores, namely Quantcast-QFS, Swift and Tahoe-LAFS on a three-wide data center environment, in order to assess the impact of WAN latencies on these systems. We focus on the read/write performance of these systems, and ignore other features such as repair, ease of use, maintainability, recovery performance, and compatibility to other system components. These considerations, though important, are more subjective and application dependent. Moreover, our eventual goal is to understand the weaknesses of these distributed storage systems Fig. 1. IL 32. MiB/s.2 MiB/s 25ms GA Multi-site Data Center Deployment with Network Bandwidths on a WAN setting, and thus we make the conscious choice on these restricted but most fundamental issues. II. QUANTCAST-QFS, TAHOE-LAFS AND SWIFT PDFill PDF Editor with Free Writer and Tools In this section, we provide some background information for the three open-source storage systems that we chose to evaluate, and briefly describe the characteristics that are relevant to us. A. Quantcast-QFS Quantcast-QFS is a high-performance, fault-tolerant, distributed file system developed by Quantcast corporation, implemented in C++, that underlies its MapReduce [9] infrastructure 1. File storage in QFS utilizes the (9, ) Reed-Solomon (RS) codes [1], or simple replication codes. The system consists of a large number of storage servers (or chunk servers in QFS terminology) and a single meta-data repository residing on a dedicated meta-server. Each chunk server is responsible for storing erasure-coded chunks of files. The meta-server is responsible for balancing the placement of chunks across the set of chunk servers, maintaining the chunk placement information, and issuing repair requests. When a client initializes a write request, it first retrieves relevant file chunk placement information from the meta server. Then, the file is divided into equal-sized stripes, and each stripe is erasure coded written out to the nine chunk servers returned by the meta-server, and the process repeats with the next set of nine chunks, and potentially a new set of nine chunk servers. 1 Note that Quantcast-QFS is not related to SAN-QFS developed by Sun corporation. In what follows we will refer to Quantcast-QFS as QFS. NJ
2 To read a file, QFS retrieves any six out of the nine chunks (for every chunk group associated with that file) and reconstructs all stripes therein. The meta-server in QFS is a single-point-of-failure (SPOF); however, there exists a checkpointing mechanism such that the whole system can recover fairly quickly, without losing files that had been successfully stored before the last checkpoint. In addition, to achieve high reliability, QFS allows for certain placement policies to be specified such that the file chunks are placed onto different failure domains (e.g., across nodes, racks, and zones). Finally, given that chunk servers issue frequent keep-alive heartbeats to the meta-server, the meta-server has a very accurate view of the status of the cluster, and is hence responsible for issuing repair requests for lost chunks. B. Tahoe-LAFS Tahoe-LAFS (Least-Authority File System) is an opensource, distributed file system, implemented in Python, that can tolerate multiple data server failures or attacks, while preserving the privacy and security of the data. The underlying idea is that users can store data on the Tahoe-LAFS cluster in an encrypted form, using standard cryptographic techniques. Clients are responsible for maintaining the necessary cryptographic keys (embedded in read/write/verify capability strings ) needed to access the data, and without those keys, no entity is able to learn any information about the data, including its placement across the cluster. In addition, data is erasure coded for increased reliability, and users can choose the erasure coding parameters on a per-file basis. A Tahoe-LAFS cluster consists of a set of storage peers and a single coordinator node, called the Introducer, whose primary purpose is to announce the addition of new storage peers to the existing pool of peers by following a publish/subscribe paradigm, and relay relevant node information to clients upon read/write requests. Tahoe-LAFS exposes a file system interface to clients, similar to QFS, but in Tahoe-LAFS the file system meta-data is erasure-coded and distributed within Tahoe-LAFS itself. The client holds a key for retrieving a client specific meta-data object that associates directories/files with distinct read/write keys. When a client creates a new file, a unique public/private key pair is generated for that file, and the meta-data object is retrieved, updated and replaced. Then, the file is encrypted, erasure coded, and distributed across storage peers. Alternatively, the system can also be used as a simple object storage service, where users are responsible for managing file keys directly. The placement of the erasure coded shares is decided by a server selection algorithm that hashes the private key of the file, in order to generate a distinct server permutation. Then, servers without enough storage space are removed from the permutation, and the rest of the servers are contacted in sequence and asked to hold one share of the file. When enough peers are found, the write proceeds. Reading a file involves asking all known servers in the cluster to list the number of shares of that file that they hold, and on a subsequent round, choosing which shares to request, based on various heuristics that take into account latency and peer load. The introducer in Tahoe-LAFS is a single-point-of-failure for new clients or new storage peers, since they need to rely on it to be able to join the storage network. However, it is not an SPOF in the traditional sense, because in Tahoe-LAFS the file placement information is decided by hashing the keys held by the clients, and the introducer does not have to maintain any file (or chunk) placement information, therefore, losing the introducer does not jeopardize the normal operation of existing clients and storage peers. The downside of this introducer-storage-nodes architecture is that due to the lack of a centralized meta-server, Tahoe- LAFS is not able to provide complete statistics and details for file storage and placement, though from a security point of view, this is a choice by design. Moreover only a lazy repair policy can be implemented; in other words, the clients are responsible for frequently iterating through their file space and verifying that all files are complete. However when a storage peer fails, the system is not able to initiate an efficient repair mechanism specifically to replace the given peer. C. Swift As part of the OpenStack architecture, Swift was originally designed to provide a reliable, distributed storage facility for maintaining virtual machine images. Despite its roots, it can be used as a generic object store, that uses replication for increased reliability. It is implemented in Python. A Swift cluster consists of a large number of storage servers, which are called object servers in Swift terminology. Each object server is only capable of storing, retrieving, and deleting objects. Objects are stored as raw binary files on the underlying file system, with meta data stored in the file s extended attributes. Swift also uses the concept of proxy servers. Proxy servers are responsible for accepting read/write/delete requests from clients, and coordinating object servers in order to accomplish those requests. The proxy servers also act as load-balancers, firewalls and a caching layer for the Swift cluster. Finally, the Proxy servers are responsible for maintaining user accounts, and a list of all containers and objects within those containers. This meta-data is stored in individual SQLite databases, that are themselves treated as objects, hence they are replicated and stored within the object servers. It should be noted that Swift does not maintain any object placement meta-data, unlike QFS, and is similar to Tahoe- LAFS. Instead, Swift uses the concept of a ring (or consistent hashing), borrowed from Amazon Dynamo [8]. A ring represents a mapping between the names of entities and their physical location on object servers. There are separate rings for accounts, containers, and objects. The ring conceptually divides the entity namespace into a pre-specified number of partitions, and each partition is assigned to multiple object servers (hence entities falling within a given partition are replicated multiple times; the replication factor is configurable). Each partition replica is guaranteed to reside in a different failure domain (or zones in Swift terminology), where the zones are statically defined during system configuration. Swift does not currently support erasure coding. In terms of repair, Swift takes a pro-active approach. Each object server periodically checks whether each partition the server is responsible for needs to be repaired. For each partition, the server polls all other object servers that should also be replicating this partition, and if any objects within the
3 partition are missing (or if an object server has failed), rsync is run to repair the partition (or replicate the partition to a temporary handoff server). 15 III. THE TEST ENVIRONMENT For our tests, two physical configurations are used. The main configuration consists of three geographically distant data center. The baseline configuration consists of a single data center. The three-site layout is depicted in Figure 1. The sites selected for the test were roughly arranged in a geographically equilateral triangle with several hundred miles separating each. Site 1 is in Lisle, IL, site 2 is in Secaucus, NJ, and site 3 is in Alpharetta, GA. The connectivity between sites varies significantly from site to site, and we observe that network characteristics are not symmetric. Maximum throughput between sites varied depending on direction. Furthermore, in one case throughput varied significantly from one measurement to the next, even when measurements were taken within very short time spans. While the network was far from ideal, and prevented us from determining the best performance that could be obtained from each storage system, direct comparison of the three systems is still possible and meaningful. For the three-site configuration, within each data center we used three hosts as storage servers and one host, referred to as meta host in what follows, for supporting tasks (i.e., the QFS meta-server, the Swift proxy, and the Tahoe-LAFS introducer). The meta host in GA was also used to drive the tests. Each host has an Intel Xeon E Core processor (2.9 GHz), 4GiB of main memory, and nine 72RPM SATA drives of 2TB each in a JBOD configuration running XFS. The operating system is Ubuntu 12.4 LTS with Linux kernel For the single site configuration we used a total of ten hosts, nine as storage servers and one meta host for everything else. All hosts within a data center were connected to the same switch using 1-gigabit Ethernet. All the tests were driven using Cosbench (Cloud Object Store benchmarking) [11], an open-source tool developed by Intel. Cosbench implements a driver that simulates user defined workload configurations. Users can choose the characteristics of the workload in terms of the size of files, the type and percentages of create/write/read/delete operations, and the number of concurrent threads of execution. We extended Cosbench by adding plug-ins to provide interfaces to QFS and Tahoe-LAFS; the interface to Swift is provided as part of the Cosbench package. The tests were organized such that data written to the object store was randomly generated, and data read from the object store was discarded; no disk I/O on the test host impacted the throughput measurements. Results presented in this paper with regard to performance are measured in (powers of 1) and are referred to by Cosbench as bandwidth. In Cosbench, bandwidth is not computed similarly to the traditional throughput measurement (i.e., total bytes over elapsed time), but it is a summation over the throughput of each individual thread of execution. This method of calculation can yield larger values than would be observed by the traditional computation, especially for tests involving files of variable sizes, because it does not capture the idle time for threads in-between job scheduling. Clearly, Fig. 2. Fig QFS/Swift/Tahoe-LAFS Multi-Site Read Performance QFS/Swift/Tahoe-LAFS Multi-Site Write Performance bandwidth is a good measure in practice, because it does not reflect any design choices related to the Cosbench job scheduler itself. IV. RESULTS Measurements were collected for each system either reading or writing fixed-sized objects. Tests with concurrent reading and writing were not conducted. In our tests we used workloads consisting of 1MB and 1GB objects. For brevity, we present averages over all workloads. The tests were executed with a varying number of concurrent threads. In most cases, the addition of threads, up to a point, increased overall bandwidth, but also resulted in the degradation of the throughput of each individual thread, as expected, due to contention on resources. A. Multi-site Test Results The results are shown in Figures 2 and 3. There are a few observations that should be noted. Writing of larger objects (1 GB vs 1 MB; not shown in the figures), was observed to be slightly faster for all three storage systems for the same total volume of data written. This is likely due to the reduced number of interactions per MB with the meta/proxy server. The difference, though, was not significant. Read performance from multiple sites was slightly better than write performance for all three storage systems. This is expected in erasure coded systems, since reads in fact transfer less data than writes. Performance increased as more threads, to a point, were allowed to access the objects.
4 Overall, we can see that QFS exhibits the best performance and scalability with respect to concurrent requests. Note that the y-axis is in log-scale. Also notice that as more threads are used, the read/write performance flattens to a point where the available bandwidth dictates what is the maximum throughput that can be obtained, as expected. B. Single-Site Test Results In order to understand the impact that the network imposes on a multi-site environment we established a single site with nine storage hosts and one meta host. Two sets of tests were conducted: one is when the reads/writes are initiated from a node located in the same site, which is referred to as local read/write; the other is when they are initiated from a node located in a different geo-location, which is referred to as remote read/write. For the former, QFS, Tahoe-LAFS and Swift are all tested, while for the latter we focus on QFS. Figures 4 and 5 show the single-site test results, while Figures and 7 show the remote read/write performance for QFS. A few observations are noted below: In terms of local read/write performance, QFS is the clear winner among the three systems. Notice the logarithmic scale on the y-axis, once again. Accessing the single site object stores from a local host results in nearly one order of magnitude increase in performance, as shown in Figures and 7. Even taking into account the bandwidth limitation between data centers, this significant difference is remarkable. We suspect that this is due to the system optimization done in QFS for single-site deployment (further discussion is included in Section V). When accessing the single site object store from a remote location, the performance drops slightly below the performance observed with the object store distributed across multiple sites. This small difference might be caused by the fact that when testing on the multi-site configuration, one set of storage hosts is local to the host running the Cosbench driver and, thus, one third of the data transfers are local. Notice that in any application scenario where we expect the majority of client requests to originate from remote locations, this implies that in fact deploying these systems in a multi-site configuration is preferable than the single-site configuration (both in terms of reliability and performance). Of course, this is not the case, for example, for MapReduce deployments. V. DISCUSSION There are certain limiting factors in terms of performance when trying to deploy distributed storage systems in a WAN setting, mainly due to the latency introduced by the physical connection, TCP slow start, and of course due to shared bandwidth. Systems that rely on a meta-server, like QFS (and, for example, HDFS), introduce large latencies when reading/writing objects across the WAN, because every read/write request for each chunk of the file incurs one round-trip delay to the metaserver for the client and each storage server involved (although Fig. 4. Fig QFS/Swift/Tahoe-LAFS Single-Site Read Performance QFS/Swift/Tahoe-LAFS Single-Site Write Performance storage servers will typically aggregate multiple chunk replies into a single message digest for the meta-server). If the metaserver is located in a remote data center from the client, we expect this architecture to add a significant overhead for read/write requests. In addition, given that QFS splits files into 4MiB chunks, and chunks are uniformly distributed across all storage servers, we expect QFS to suffer significantly from TCP slow start. Systems that use consistent hashing to determine object placement (e.g., Riak and Amazon Dynamo), need to use a distributed consensus protocol (e.g., Paxos [12]) in order to keep the state of the cluster up-to-date, which requires at least one round-trip delay per server for each cluster update. But read/write requests happen independently of the cluster management protocol. Hence read/write requests go directly to the relevant storage servers, without any additional round-trip delays. From that respect, even though Swift uses consistent hashing in order to determine object placement, the ring is statically allocated during system configuration (and can change only manually). Hence, Swift does not have the overhead of running a distributed consensus protocol to maintain the consistent hashing ring up-to-date. On the other hand, Swift does have to maintain meta-data information (accounts and containers) as native objects within the object store itself, hence, incurring at least one round-trip delay for every update of each replica of the meta-data object. Although, in Swift this latency can be hidden completely due to the caching layer at the Proxy servers. We also observe that under-provisioning the proxy setup in Swift can have a detrimental effect in terms of scalability. Swift does not scale well as the number of concurrent threads increases, resulting in a large number of dropped operations (not shown in our figures since we plot averages across all workloads for successful operations only). This is because all data transfers have to go through
5 Fig.. Fig Multi-site Single-loc Single-rmt QFS Single-Site Read Performance QFS Single-Site Write Performance Multi-site Single-loc Single-rmt the proxy server, which eventually becomes oversubscribed and starts to shed requests. Clearly this is an indication that they proxy server is not designed to scale gracefully as the number of clients increases, given that, ostensibly, for all storage systems eventually all data has to go through the sole Cosbench driver running on the meta host, in which case we would expect Cosbench to become oversubscribed and we should be observing the same behavior for QFS and Tahoe- LAFS, which is not the case. Nevertheless, a more robust Swift configuration would include several proxy servers that would load-balance requests, but this is something that we did not test in our configuration, and we plan to do as future work, since it would require a multi-driver configuration of Cosbench. Tahoe-LAFS is similar to Swift, in that the meta-data objects are stored within Tahoe-LAFS itself, necessitating at least one round-trip delay for each erasure coded share of the meta-data object, for every write/delete request, and an additional round-trip to all relevant servers to execute the request. On the other hand, reads are accomplished by submitting requests to all known storage peers (given by the introducer) simultaneously, hence the relevant peers are found with one round-trip to every server. In Tahoe-LAFS currently, a second round-trip is incurred, after choosing the peers to read the file from (the intention here is to be able to select which peers to read from, after the initial negotiation phase, based on various heuristics). But the poor performance of Tahoe-LAFS, on the multi-site and single-site environment, can be attributed to several factors, already pointed out by the developers themselves. First, the default stripe size is optimized for reading/writing small objects (the stripe size determines the granularity at which data is being encrypted and erasure coded). Second, due to several implementation issues such as inefficient message passing, and the expensive and frequent hash read/write seeks that are needed in order to reconstruct shares on the storage peers. Third, Tahoe-LAFS has to deal with the overhead of reading/writing the file system meta-data objects (i.e., the mutable directory objects), every time an object is accessed. Fourth, when creating new objects, Tahoe-LAFS has to generate a new public/private key, which is an expensive operation. Surprisingly, reads exhibit the same performance as writes, even though reads ideally have to transfer less data than writes. This is probably because both reads and writes of shares happen simultaneously to all relevant storage peers, hence the extra data transfers are hidden by parallelism. Moreover, this is an indication that pinging all available peers and requesting shares within two round-trips, as well as the fact that every read request has to read the mutable directory object, dominate the overall cost. VI. CONCLUSION We conducted extensive experiments with QFS, Swift and Tahoe-LAFS, which are three very popular distributed storage systems. Our focus was to deploy these systems in a multi-site environment and measure the impact of WAN characteristics on these systems. In addition, as a baseline, we also measured performance on a single-site configuration. Overall, we observe that WAN characteristics have an even larger than expected impact on the performance of these systems, mainly due to several design choices of these systems. Ideally, across the WAN, we would like to reduce the amount of round-trips to a minimum, which is something not particularly important on a LAN. In addition, we notice that good system design and extensive optimizations can have a significant effect on performance, as is seen by the relative difference between QFS, Swift and Tahoe-LAFS. It is important to point out here that QFS is implemented in C++, Swift and Tahoe-LAFS in Python. In addition, QFS is heavily optimized for MapReduce style processing. For future work we are planning to also test Riak and HDFS, as well as our own proprietary solution that is designed, from the ground up, for WAN deployment. REFERENCES [1] D. Borthakur, The hadoop distributed file system: Architecture and design, design.pdf, 27. [2] Swift, [3] QFS, [4] Tahoe-LAFS, [5] Riak, [] B. C. J. Wang, A. Ogus, N. Nilakantan, A. Skjolsvold, S. McKelvie, Y. Xu, S. Srivastav, J. Wu, H. Simitci et al., Windows azure storage: a highly available cloud storage service with strong consistency, in Proceedings of the Twenty-Third ACM Symposium on Operating Systems Principles. ACM, 211, pp [7] A. Fikes, Storage architecture and challenges, Faculty Summit, 29. [8] G. DeCandia, D. Hastorun, M. Jampani, G. Kakulapati, A. Lakshman, A. Pilchin, S. Sivasubramanian, P. Vosshall, and W. Vogels, Dynamo: Amazon s highly available key-value store, in SOSP, vol. 7, 27, pp [9] J. Dean and S. Ghemawat, MapReduce: simplified data processing on large clusters, Communications of the ACM, vol. 51, no. 1, pp , 28. [1] S. B. Wicker and V. K. Bhargava, Reed-Solomon codes and their applications. Wiley.com, [11] J. Duan, COSBench: A benchmark tool for cloud object storage services, OpenStack Summit Fall 212, 212. [12] L. Lamport, The part-time parliament, ACM Transactions on Computer Systems (TOCS), vol. 1, no. 2, pp , 1998.
Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms
Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes
Lecture 5: GFS & HDFS! Claudia Hauff (Web Information Systems)! [email protected]
Big Data Processing, 2014/15 Lecture 5: GFS & HDFS!! Claudia Hauff (Web Information Systems)! [email protected] 1 Course content Introduction Data streams 1 & 2 The MapReduce paradigm Looking behind
Distributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
GraySort and MinuteSort at Yahoo on Hadoop 0.23
GraySort and at Yahoo on Hadoop.23 Thomas Graves Yahoo! May, 213 The Apache Hadoop[1] software library is an open source framework that allows for the distributed processing of large data sets across clusters
CSE-E5430 Scalable Cloud Computing Lecture 2
CSE-E5430 Scalable Cloud Computing Lecture 2 Keijo Heljanko Department of Computer Science School of Science Aalto University [email protected] 14.9-2015 1/36 Google MapReduce A scalable batch processing
Key Components of WAN Optimization Controller Functionality
Key Components of WAN Optimization Controller Functionality Introduction and Goals One of the key challenges facing IT organizations relative to application and service delivery is ensuring that the applications
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack
Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack May 2015 Copyright 2015 SwiftStack, Inc. swiftstack.com Page 1 of 19 Table of Contents INTRODUCTION... 3 OpenStack
Distributed File Systems
Distributed File Systems Mauro Fruet University of Trento - Italy 2011/12/19 Mauro Fruet (UniTN) Distributed File Systems 2011/12/19 1 / 39 Outline 1 Distributed File Systems 2 The Google File System (GFS)
Evaluating HDFS I/O Performance on Virtualized Systems
Evaluating HDFS I/O Performance on Virtualized Systems Xin Tang [email protected] University of Wisconsin-Madison Department of Computer Sciences Abstract Hadoop as a Service (HaaS) has received increasing
Scalable Multiple NameNodes Hadoop Cloud Storage System
Vol.8, No.1 (2015), pp.105-110 http://dx.doi.org/10.14257/ijdta.2015.8.1.12 Scalable Multiple NameNodes Hadoop Cloud Storage System Kun Bi 1 and Dezhi Han 1,2 1 College of Information Engineering, Shanghai
CSE-E5430 Scalable Cloud Computing P Lecture 5
CSE-E5430 Scalable Cloud Computing P Lecture 5 Keijo Heljanko Department of Computer Science School of Science Aalto University [email protected] 12.10-2015 1/34 Fault Tolerance Strategies for Storage
How To Make A Backup System More Efficient
Identifying the Hidden Risk of Data De-duplication: How the HYDRAstor Solution Proactively Solves the Problem October, 2006 Introduction Data de-duplication has recently gained significant industry attention,
Berkeley Ninja Architecture
Berkeley Ninja Architecture ACID vs BASE 1.Strong Consistency 2. Availability not considered 3. Conservative 1. Weak consistency 2. Availability is a primary design element 3. Aggressive --> Traditional
Benchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014
VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup
International Journal of Advance Research in Computer Science and Management Studies
Volume 2, Issue 8, August 2014 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online
Storage Systems Autumn 2009. Chapter 6: Distributed Hash Tables and their Applications André Brinkmann
Storage Systems Autumn 2009 Chapter 6: Distributed Hash Tables and their Applications André Brinkmann Scaling RAID architectures Using traditional RAID architecture does not scale Adding news disk implies
Hadoop Architecture. Part 1
Hadoop Architecture Part 1 Node, Rack and Cluster: A node is simply a computer, typically non-enterprise, commodity hardware for nodes that contain data. Consider we have Node 1.Then we can add more nodes,
Comparative analysis of mapreduce job by keeping data constant and varying cluster size technique
Comparative analysis of mapreduce job by keeping data constant and varying cluster size technique Mahesh Maurya a, Sunita Mahajan b * a Research Scholar, JJT University, MPSTME, Mumbai, India,[email protected]
Comparative analysis of Google File System and Hadoop Distributed File System
Comparative analysis of Google File System and Hadoop Distributed File System R.Vijayakumari, R.Kirankumar, K.Gangadhara Rao Dept. of Computer Science, Krishna University, Machilipatnam, India, [email protected]
Service Description Cloud Storage Openstack Swift
Service Description Cloud Storage Openstack Swift Table of Contents Overview iomart Cloud Storage... 3 iomart Cloud Storage Features... 3 Technical Features... 3 Proxy... 3 Storage Servers... 4 Consistency
1. Comments on reviews a. Need to avoid just summarizing web page asks you for:
1. Comments on reviews a. Need to avoid just summarizing web page asks you for: i. A one or two sentence summary of the paper ii. A description of the problem they were trying to solve iii. A summary of
Aspera Direct-to-Cloud Storage WHITE PAPER
Transport Direct-to-Cloud Storage and Support for Third Party April 2014 WHITE PAPER TABLE OF CONTENTS OVERVIEW 3 1 - THE PROBLEM 3 2 - A FUNDAMENTAL SOLUTION - ASPERA DIRECT-TO-CLOUD TRANSPORT 5 3 - VALIDATION
Introduction to Gluster. Versions 3.0.x
Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster
PARALLELS CLOUD STORAGE
PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...
Online Remote Data Backup for iscsi-based Storage Systems
Online Remote Data Backup for iscsi-based Storage Systems Dan Zhou, Li Ou, Xubin (Ben) He Department of Electrical and Computer Engineering Tennessee Technological University Cookeville, TN 38505, USA
MANAGEMENT OF DATA REPLICATION FOR PC CLUSTER BASED CLOUD STORAGE SYSTEM
MANAGEMENT OF DATA REPLICATION FOR PC CLUSTER BASED CLOUD STORAGE SYSTEM Julia Myint 1 and Thinn Thu Naing 2 1 University of Computer Studies, Yangon, Myanmar [email protected] 2 University of Computer
Accelerating and Simplifying Apache
Accelerating and Simplifying Apache Hadoop with Panasas ActiveStor White paper NOvember 2012 1.888.PANASAS www.panasas.com Executive Overview The technology requirements for big data vary significantly
MinCopysets: Derandomizing Replication In Cloud Storage
MinCopysets: Derandomizing Replication In Cloud Storage Asaf Cidon, Ryan Stutsman, Stephen Rumble, Sachin Katti, John Ousterhout and Mendel Rosenblum Stanford University [email protected], {stutsman,rumble,skatti,ouster,mendel}@cs.stanford.edu
POSIX and Object Distributed Storage Systems
1 POSIX and Object Distributed Storage Systems Performance Comparison Studies With Real-Life Scenarios in an Experimental Data Taking Context Leveraging OpenStack Swift & Ceph by Michael Poat, Dr. Jerome
The OpenStack TM Object Storage system
The OpenStack TM Object Storage system Deploying and managing a scalable, open- source cloud storage system with the SwiftStack Platform By SwiftStack, Inc. [email protected] Contents Introduction...
Dynamo: Amazon s Highly Available Key-value Store
Dynamo: Amazon s Highly Available Key-value Store Giuseppe DeCandia, Deniz Hastorun, Madan Jampani, Gunavardhan Kakulapati, Avinash Lakshman, Alex Pilchin, Swaminathan Sivasubramanian, Peter Vosshall and
GeoGrid Project and Experiences with Hadoop
GeoGrid Project and Experiences with Hadoop Gong Zhang and Ling Liu Distributed Data Intensive Systems Lab (DiSL) Center for Experimental Computer Systems Research (CERCS) Georgia Institute of Technology
How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda
How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda 1 Outline Build a cost-efficient Swift cluster with expected performance Background & Problem Solution Experiments
Journal of science STUDY ON REPLICA MANAGEMENT AND HIGH AVAILABILITY IN HADOOP DISTRIBUTED FILE SYSTEM (HDFS)
Journal of science e ISSN 2277-3290 Print ISSN 2277-3282 Information Technology www.journalofscience.net STUDY ON REPLICA MANAGEMENT AND HIGH AVAILABILITY IN HADOOP DISTRIBUTED FILE SYSTEM (HDFS) S. Chandra
Parallel Processing of cluster by Map Reduce
Parallel Processing of cluster by Map Reduce Abstract Madhavi Vaidya, Department of Computer Science Vivekanand College, Chembur, Mumbai [email protected] MapReduce is a parallel programming model
HDFS Architecture Guide
by Dhruba Borthakur Table of contents 1 Introduction... 3 2 Assumptions and Goals... 3 2.1 Hardware Failure... 3 2.2 Streaming Data Access...3 2.3 Large Data Sets... 3 2.4 Simple Coherency Model...3 2.5
TECHNICAL WHITE PAPER: ELASTIC CLOUD STORAGE SOFTWARE ARCHITECTURE
TECHNICAL WHITE PAPER: ELASTIC CLOUD STORAGE SOFTWARE ARCHITECTURE Deploy a modern hyperscale storage platform on commodity infrastructure ABSTRACT This document provides a detailed overview of the EMC
Jeffrey D. Ullman slides. MapReduce for data intensive computing
Jeffrey D. Ullman slides MapReduce for data intensive computing Single-node architecture CPU Machine Learning, Statistics Memory Classical Data Mining Disk Commodity Clusters Web data sets can be very
Performance Analysis of Mixed Distributed Filesystem Workloads
Performance Analysis of Mixed Distributed Filesystem Workloads Esteban Molina-Estolano, Maya Gokhale, Carlos Maltzahn, John May, John Bent, Scott Brandt Motivation Hadoop-tailored filesystems (e.g. CloudStore)
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at distributing load b. QUESTION: What is the context? i. How
Apache Hadoop. Alexandru Costan
1 Apache Hadoop Alexandru Costan Big Data Landscape No one-size-fits-all solution: SQL, NoSQL, MapReduce, No standard, except Hadoop 2 Outline What is Hadoop? Who uses it? Architecture HDFS MapReduce Open
Fault Tolerance in Hadoop for Work Migration
1 Fault Tolerance in Hadoop for Work Migration Shivaraman Janakiraman Indiana University Bloomington ABSTRACT Hadoop is a framework that runs applications on large clusters which are built on numerous
ENABLING GLOBAL HADOOP WITH EMC ELASTIC CLOUD STORAGE
ENABLING GLOBAL HADOOP WITH EMC ELASTIC CLOUD STORAGE Hadoop Storage-as-a-Service ABSTRACT This White Paper illustrates how EMC Elastic Cloud Storage (ECS ) can be used to streamline the Hadoop data analytics
Introduction to Cloud : Cloud and Cloud Storage. Lecture 2. Dr. Dalit Naor IBM Haifa Research Storage Systems. Dalit Naor, IBM Haifa Research
Introduction to Cloud : Cloud and Cloud Storage Lecture 2 Dr. Dalit Naor IBM Haifa Research Storage Systems 1 Advanced Topics in Storage Systems for Big Data - Spring 2014, Tel-Aviv University http://www.eng.tau.ac.il/semcom
How to Choose Between Hadoop, NoSQL and RDBMS
How to Choose Between Hadoop, NoSQL and RDBMS Keywords: Jean-Pierre Dijcks Oracle Redwood City, CA, USA Big Data, Hadoop, NoSQL Database, Relational Database, SQL, Security, Performance Introduction A
Hadoop: Embracing future hardware
Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop
Quantcast Petabyte Storage at Half Price with QFS!
9-131 Quantcast Petabyte Storage at Half Price with QFS Presented by Silvius Rus, Director, Big Data Platforms September 2013 Quantcast File System (QFS) A high performance alternative to the Hadoop Distributed
Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem
Identifying the Hidden Risk of Data Deduplication: How the HYDRAstor TM Solution Proactively Solves the Problem Advanced Storage Products Group Table of Contents 1 - Introduction 2 Data Deduplication 3
Index Terms : Load rebalance, distributed file systems, clouds, movement cost, load imbalance, chunk.
Load Rebalancing for Distributed File Systems in Clouds. Smita Salunkhe, S. S. Sannakki Department of Computer Science and Engineering KLS Gogte Institute of Technology, Belgaum, Karnataka, India Affiliated
A Study on Workload Imbalance Issues in Data Intensive Distributed Computing
A Study on Workload Imbalance Issues in Data Intensive Distributed Computing Sven Groot 1, Kazuo Goda 1, and Masaru Kitsuregawa 1 University of Tokyo, 4-6-1 Komaba, Meguro-ku, Tokyo 153-8505, Japan Abstract.
Parallels Cloud Storage
Parallels Cloud Storage White Paper Best Practices for Configuring a Parallels Cloud Storage Cluster www.parallels.com Table of Contents Introduction... 3 How Parallels Cloud Storage Works... 3 Deploying
Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1
Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System
A Brief Analysis on Architecture and Reliability of Cloud Based Data Storage
Volume 2, No.4, July August 2013 International Journal of Information Systems and Computer Sciences ISSN 2319 7595 Tejaswini S L Jayanthy et al., Available International Online Journal at http://warse.org/pdfs/ijiscs03242013.pdf
A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems*
A Content-Based Load Balancing Algorithm for Metadata Servers in Cluster File Systems* Junho Jang, Saeyoung Han, Sungyong Park, and Jihoon Yang Department of Computer Science and Interdisciplinary Program
On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform
On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform Page 1 of 16 Table of Contents Table of Contents... 2 Introduction... 3 NoSQL Databases... 3 CumuLogic NoSQL Database Service...
PERFORMANCE ANALYSIS OF KERNEL-BASED VIRTUAL MACHINE
PERFORMANCE ANALYSIS OF KERNEL-BASED VIRTUAL MACHINE Sudha M 1, Harish G M 2, Nandan A 3, Usha J 4 1 Department of MCA, R V College of Engineering, Bangalore : 560059, India [email protected] 2 Department
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
Cloud Computing at Google. Architecture
Cloud Computing at Google Google File System Web Systems and Algorithms Google Chris Brooks Department of Computer Science University of San Francisco Google has developed a layered system to handle webscale
Can High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
Sector vs. Hadoop. A Brief Comparison Between the Two Systems
Sector vs. Hadoop A Brief Comparison Between the Two Systems Background Sector is a relatively new system that is broadly comparable to Hadoop, and people want to know what are the differences. Is Sector
Design and Implementation of a Storage Repository Using Commonality Factoring. IEEE/NASA MSST2003 April 7-10, 2003 Eric W. Olsen
Design and Implementation of a Storage Repository Using Commonality Factoring IEEE/NASA MSST2003 April 7-10, 2003 Eric W. Olsen Axion Overview Potentially infinite historic versioning for rollback and
Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF
Panasas at the RCF HEPiX at SLAC Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory Centralized File Service Single, facility-wide namespace for files. Uniform, facility-wide
ZooKeeper. Table of contents
by Table of contents 1 ZooKeeper: A Distributed Coordination Service for Distributed Applications... 2 1.1 Design Goals...2 1.2 Data model and the hierarchical namespace...3 1.3 Nodes and ephemeral nodes...
THE HADOOP DISTRIBUTED FILE SYSTEM
THE HADOOP DISTRIBUTED FILE SYSTEM Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Presented by Alexander Pokluda October 7, 2013 Outline Motivation and Overview of Hadoop Architecture,
Hadoop and Map-Reduce. Swati Gore
Hadoop and Map-Reduce Swati Gore Contents Why Hadoop? Hadoop Overview Hadoop Architecture Working Description Fault Tolerance Limitations Why Map-Reduce not MPI Distributed sort Why Hadoop? Existing Data
GraySort on Apache Spark by Databricks
GraySort on Apache Spark by Databricks Reynold Xin, Parviz Deyhim, Ali Ghodsi, Xiangrui Meng, Matei Zaharia Databricks Inc. Apache Spark Sorting in Spark Overview Sorting Within a Partition Range Partitioner
CS2510 Computer Operating Systems
CS2510 Computer Operating Systems HADOOP Distributed File System Dr. Taieb Znati Computer Science Department University of Pittsburgh Outline HDF Design Issues HDFS Application Profile Block Abstraction
CS2510 Computer Operating Systems
CS2510 Computer Operating Systems HADOOP Distributed File System Dr. Taieb Znati Computer Science Department University of Pittsburgh Outline HDF Design Issues HDFS Application Profile Block Abstraction
SOLVING LOAD REBALANCING FOR DISTRIBUTED FILE SYSTEM IN CLOUD
International Journal of Advances in Applied Science and Engineering (IJAEAS) ISSN (P): 2348-1811; ISSN (E): 2348-182X Vol-1, Iss.-3, JUNE 2014, 54-58 IIST SOLVING LOAD REBALANCING FOR DISTRIBUTED FILE
Diagram 1: Islands of storage across a digital broadcast workflow
XOR MEDIA CLOUD AQUA Big Data and Traditional Storage The era of big data imposes new challenges on the storage technology industry. As companies accumulate massive amounts of data from video, sound, database,
www.basho.com Technical Overview Simple, Scalable, Object Storage Software
www.basho.com Technical Overview Simple, Scalable, Object Storage Software Table of Contents Table of Contents... 1 Introduction & Overview... 1 Architecture... 2 How it Works... 2 APIs and Interfaces...
Processing of Hadoop using Highly Available NameNode
Processing of Hadoop using Highly Available NameNode 1 Akash Deshpande, 2 Shrikant Badwaik, 3 Sailee Nalawade, 4 Anjali Bote, 5 Prof. S. P. Kosbatwar Department of computer Engineering Smt. Kashibai Navale
SiteCelerate white paper
SiteCelerate white paper Arahe Solutions SITECELERATE OVERVIEW As enterprises increases their investment in Web applications, Portal and websites and as usage of these applications increase, performance
Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage
Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage Technical white paper Table of contents Executive summary... 2 Introduction... 2 Test methodology... 3
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
Introduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7
Introduction 1 Performance on Hosted Server 1 Figure 1: Real World Performance 1 Benchmarks 2 System configuration used for benchmarks 2 Figure 2a: New tickets per minute on E5440 processors 3 Figure 2b:
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011 Executive Summary Large enterprise Hyper-V deployments with a large number
GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
Distributed Data Stores
Distributed Data Stores 1 Distributed Persistent State MapReduce addresses distributed processing of aggregation-based queries Persistent state across a large number of machines? Distributed DBMS High
Scala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
marlabs driving digital agility WHITEPAPER Big Data and Hadoop
marlabs driving digital agility WHITEPAPER Big Data and Hadoop Abstract This paper explains the significance of Hadoop, an emerging yet rapidly growing technology. The prime goal of this paper is to unveil
How To Virtualize A Storage Area Network (San) With Virtualization
A New Method of SAN Storage Virtualization Table of Contents 1 - ABSTRACT 2 - THE NEED FOR STORAGE VIRTUALIZATION 3 - EXISTING STORAGE VIRTUALIZATION METHODS 4 - A NEW METHOD OF VIRTUALIZATION: Storage
BookKeeper. Flavio Junqueira Yahoo! Research, Barcelona. Hadoop in China 2011
BookKeeper Flavio Junqueira Yahoo! Research, Barcelona Hadoop in China 2011 What s BookKeeper? Shared storage for writing fast sequences of byte arrays Data is replicated Writes are striped Many processes
MAGENTO HOSTING Progressive Server Performance Improvements
MAGENTO HOSTING Progressive Server Performance Improvements Simple Helix, LLC 4092 Memorial Parkway Ste 202 Huntsville, AL 35802 [email protected] 1.866.963.0424 www.simplehelix.com 2 Table of Contents
Understanding Microsoft Storage Spaces
S T O R A G E Understanding Microsoft Storage Spaces A critical look at its key features and value proposition for storage administrators A Microsoft s Storage Spaces solution offers storage administrators
Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000
Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000 Alexandra Carpen-Amarie Diana Moise Bogdan Nicolae KerData Team, INRIA Outline
Data-Intensive Computing with Map-Reduce and Hadoop
Data-Intensive Computing with Map-Reduce and Hadoop Shamil Humbetov Department of Computer Engineering Qafqaz University Baku, Azerbaijan [email protected] Abstract Every day, we create 2.5 quintillion
Purpose... 3. Computer Hardware Configurations... 6 Single Computer Configuration... 6 Multiple Server Configurations... 7. Data Encryption...
Contents Purpose... 3 Background on Keyscan Software... 3 Client... 4 Communication Service... 4 SQL Server 2012 Express... 4 Aurora Optional Software Modules... 5 Computer Hardware Configurations... 6
AppSense Environment Manager. Enterprise Design Guide
Enterprise Design Guide Contents Introduction... 3 Document Purpose... 3 Basic Architecture... 3 Common Components and Terminology... 4 Best Practices... 5 Scalability Designs... 6 Management Server Scalability...
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi
EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Applied Technology Abstract Microsoft SQL Server includes a powerful capability to protect active databases by using either
Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh
1 Hadoop: A Framework for Data- Intensive Distributed Computing CS561-Spring 2012 WPI, Mohamed Y. Eltabakh 2 What is Hadoop? Hadoop is a software framework for distributed processing of large datasets
