Survey of Technologies for Wide Area Distributed Storage
|
|
|
- Candice Jackson
- 10 years ago
- Views:
Transcription
1 Survey of Technologies for Wide Area Distributed Storage Project : GigaPort3 Project Year : 2010 Project Manager : Rogier Spoor Author(s) Completion Date : Version : 1.0 : Arjan Peddemors, Christiaan Kuun, Rogier Spoor, Paul Dekkers Christiaan den Besten Summary This report gives on overview of existing technologies that may be used to offer distributed storage services to SURFnet and its connected institutes. It provides an analysis of the requirements that are relevant for such services, which is used to compare the different products. Furthermore, this report gives a description of the environment that will be used to evaluate candidate products. This publication is licensed under Creative Commons Attribution 3.0 Unported. More information on this license can be found at
2 Colophon Programme line Part Activity Deliverable Access rights External party : Enabling Dynamic Services : Task 3 - Storage Clouds : Technology Scouting Storage Clouds : EDS-3R Report on scouting results on Storage Clouds Technology : Public : Novay, CSIR, Prolocation This project was made possible by the Economic Structure Enhancing Fund (FES) of the Dutch Government.
3 Contents 1 Introduction Use cases Virtual machine distributed storage Database replication Scientific data storage Requirements Overview of existing products Candidate products Lustre GlusterFS GPFS Ceph Candidate product comparison Non-candidate products XtreemFS MogileFS NFS version 4.1 (pnfs) ZFS VERITAS File System ParaScale CAStor Tahoe-LAFS DRBD Future work Conclusion References... 22
4 1 Introduction Cheap storage technology and fast Internet access have brought online storage services within range of a large group of users and organizations. Where traditional storage capacity was closely tied to the computing facilities in use, it is now possible to store data in the cloud at reasonable prices. Current public cloud storage services, however, provide moderate levels of support for high performance and high availability applications running at the edges of the Internet. These applications come from various domains, ranging from those deployed in corporate datacenters to applications in research. They typically operate in an environment with local highend storage services in the form of storage area network (SAN) services or network attached storage (NAS), which are not easily replaced by public cloud alternatives. A different kind of online storage technology, a distributed file system with parallel access and fault tolerant features, may be better suited to support these applications. With such a system, local storage resources as well as resources at remote sites are joined to support concurrent access by multiple clients on copies of the same data. Additionally, it continues to serve data in case of failure of a component at one of the sites. Ideally, such a system regulates the placement of the data in such a way that access to that data by applications is optimal, i.e., by keeping the data close to where it is used on storage resources that match the application needs (in terms of speed, latency, etc.). We expect that a distributed storage facility can be successfully implemented in the SURFnet context, because of the high-quality and high-speed network infrastructure offered to SURFnet participants. This report provides an overview of existing products and building blocks that deliver this distributed file system functionality. We will use this survey to select a few products as candidate products, which will be further investigated and deployed in a distributed test environment (as part of the GigaPort3 project). The outline of this document is as follows. The remainder of this section presents a number of use cases in which selected products are to be used, and gives a listing of relevant requirements. In section 2, we provide an overview of products we considered, and place these products in two groups (depending on how well they match with the requirements): candidate products that will be evaluated further in the future, and non-candidate products that will not be further considered (but are included to put candidates into context). Section 3 describes the environment in which candidate products will be further tested and evaluated. Section 4 provides the survey conclusions. 4
5 1.1 Use cases To illustrate in which situations the distributed storage products will be used, we now describe a number of use cases. The main users of the system are SURFnet participants, although the use cases are also relevant for others. All use cases assume the availability of storage capacity at different locations, which are linked to each other through a fast, wide-area IP-based optical lightpath SURFnet lightpaths [19] Virtual machine distributed storage In recent years, hardware virtualization has become a mainstream technology used in many server environments to consolidate resources and to allow for flexible server configurations. Virtual machines (VMs) can be moved from one physical server machine to another with minimal interruption, which is important for those applications that have high-availability requirements. In case of hardware failure, VMs can also be powered up on different machines while using centralized storage. In configurations where hardware resources are dispersed over a large geographical area, virtual machines may be moved to a new physical machine at a long distance from the initial physical machine. A common setup for virtual machines is to access storage through a storage area network (SAN). When moving over a long distance, access to such storage may suffer from high latency. A solution to this problem is to replicate the storage data accessed by a virtual machine to those locations where it may run in the future [23][25]. When a VM is moved to a new location, it will start using the replicated storage that is nearby (instead of the storage that was close to the initial location). In this use case, selected distributed storage solution is applied to keep data stored at multiple locations in sync, to support local access to storage for virtual machines that migrate over a long distance. This mechanism allows for easier disaster recovery between geographically dispersed locations Database replication Databases play an important role in the IT infrastructure of many organizations. They offer core functionality to a wide range of software services that are crucial for day-to-day operation; they may store, for instance, messages, transaction results, customer information, document management information, etc. Usually, great care is taken to keep the information stored in databases safe, i.e., to make sure that databases are stored in a durable fashion. Additionally, databases that are necessary for the core operation of an organization are often configured in such a way (e.g. through replication [24]) that high availability is guaranteed, so that, even in case of hardware failure, the database stays online. 5
6 In this use case, a database is replicated at different nodes in a distributed storage environment, at locations that are far apart. Contrary to situations where data is replicated at nearby locations (in the order of kilometers), the geographic distribution of data over a wide area makes sure that the data is highly durable. In case of calamities such as large-scale industrial accidents or natural disasters, a short distance between replicas may not be enough to prevent data loss. In case hardware is malfunctioning at one location and the network remains operational, access to the database may be relayed to a remote node, so that the availability of the database remains intact. Database access may be such that it requires fast local as well as distributed storage Scientific data storage Scientific experiments generate data in the form of measurements. The instruments used in various scientific domains, such as DNA sequencers in biology, satellite remote sensors in geology, and particle colliders in experimental physics, are currently capable of generating vast amounts of data within a short timeframe [18]. Experiments using these instruments require substantial capacity to store the raw measurement data, but also to store data as output of analysis and processing steps executed on this raw data after the experiment. Typically, once the experimental data is captured, storage performance and high-availability is less important. Durability of the data, low storage costs, and the possibility to share experimental results between groups of scientists spread all over the world are important. In this use case, we assume a setup where a wide range of different types of experimental data is stored online in a distributed manner. The raw measurement data is entered into the system at a single location, while it is read and expanded (i.e., processed to generate derived data) at various locations, by a group of scientists working together over a substantial distance. It must be easy for scientists and system administrators to start using the storage facility and to maintain and expand it over time. Additionally, it must be possible to indicate the required level of durability offered by the storage system. 1.2 Requirements An e-science infrastructure demands a high performance, high volume and scalable data storage architecture. Many storage solutions that are currently used in the e- Science infrastructure have a limited, non-scalable approach like Storage Area Networks (SAN) or they use specially developed storage solutions that are only suitable in a specific research area like nuclear physics or astronomy. Scalable, high-capacity commercial solutions could be applied in these environments, but these are (too) expensive. Additionally, existing technologies both commercial and non-commercial often do not utilize to the fullest the unique network infrastructure available in the SURFnet setting: a very high-speed, state-of-the-art network offered at many locations in the Netherlands, with very good connections to the rest of the Internet. This network supports configurations where data storage is realized and used in a distributed fashion, over a wider area than previously feasible. 6
7 Cloud storage services, such as those offered by Amazon and Google, are examples of a fully distributed storage facility. These services have a number of attractive features, such as pay per usage and (some) guarantees that data is safely stored (through redundancy). They do not, however, provide high speed access and support for a wide range of different applications. This is also true for storage solutions applied to grid computing; in a grid infrastructure, storage facilities are focused on supporting computation at grid nodes, not on providing high-speed data access to applications at the edges of the network. There is, however, some overlap in functionality, and products used in grid environments may also be suitable for our needs. When data is stored in a distributed system, it is beneficial to place data where it is most often used, and migrate less important or less frequently accessed data to places with cheap capacity. Such a system is likely to have different storage level (or tiers) that form a storage hierarchy. The aim of this project is to determine the feasibility of building a high-capacity, fully distributed, hierarchical storage solution, exploiting the SURFnet infrastructure. Given the use cases and the aspects described above, it is clear that the system must be general purpose. We identify the following high level requirements as relevant for the distributed storage system. They are mostly qualitative requirements, as these are sufficient to survey existing products and sufficient to make an initial product selection (as is the purpose of this report). We are aware that a (wide-area) distributed storage system may not be able to offer features in any combination, i.e., that the CAP theorem applies which prescribes that only two out of the three properties data consistency, system availability, and partition tolerance can be supported by a distributed system [6]. Ideally, products must be able to explicitly balance these requirements, such that in a future system a property trade-off is easy to configure. - Scalable The system must be scalable in terms of capacity, performance and in terms of concurrent access. For instance, it must be easy to expand the total amount of storage without degrading performance and concurrency. Additionally, it must be easy, when need arises, to configure the system such that a large number of users concurrently may have access also concurrently to individual storage objects without degrading performance too much. - High Availability The system must have high-availability functionality that keeps data available to applications and clients, even in the event of malfunctioning software or hardware. This implies that the system must be capable of replicating data at multiple locations. Also, this means that it must be possible to maintain and reconfigure the system on-the-fly, i.e., while the system as a whole keeps running. In case of component failure, the system must support bringing back online the component after repair, which is likely to include synchronization actions and data consistency checks. It implies that capacity and storage locations may be added or removed while the system is operating. - Durability The system must support the storage of data in a durable manner, i.e., when a single software or hardware component fails, no data is lost. Durability functionality that must be supported is replication of data to disks at 7
8 other (remote) locations, which includes maintenance to make sure that a minimum number of replicas are available. Additionally, the system may support backup (parts) of the data on backup media such as tape. - Performance at Traditional SAN/NAS Level To support existing applications running in server environment at SURFnet institutes, the system must have to be able to support a level of performance (bandwidth and latency) comparable to that found in a traditional (non-distributed) SAN/NAS environment. Being a general purpose system, it is clear that the supported performance for specific applications cannot match the performance of dedicated storage solutions (tailored for that application). However, by offering different kinds of storage tiers within the distributed system, i.e., SSDs, RAID nodes, and SAS/SATA disks, it is possible to offer different levels of performance. Additionally, by extending existing systems based on slower hardware (e.g., SATA disks) with fast components (e.g., SSDs), system performance may be improved. The system must be able to incorporate different kinds of storage technologies and must be able to match these technologies given application requirements. It is clear that data read and write performance is closely linked with aspects such as replica management and the available WAN bandwidth required to distribute replicas. We assume that by balancing parameters such as object placement strategy, level of consistency, and level of durability, it is possible to reach a high performance level. - Dynamic Operation The level of availability, durability and performance must be configurable per application. This prevents the system to always run at the highest supported level of functionality, which reduces costs. It also allows user, application developers and system administrators to balance cost versus features. Preferably, the system must be self-configurable and self-tunable, in the sense that it changes parameters to optimize its own operation. The system must support to move data between different kinds of storage technologies offering in this way tiered functionality so that data objects that are accessed frequently are stored on the disks with highest performance and those that are infrequent accessed are stored on slower disks. It must be possible to manually override this behavior (for instance, to force a database to be available on high performance storage, even when not accessed very frequently). - Cost Effective It must be possible to build, configure, run, and maintain the system in a cost-effective manner. The system must work with commodity hardware, which means that individual hardware components may not be as reliable as when high-end hardware is used: due to the requirements of scalability, availability, durability and performance, however, the system already must be able to cope with failures. The configuration of the system, as well as the maintenance (including the resolution of failure situations) must be easy and straightforward. Preferably, the operation of the system is energy efficient. License fees for software, when applicable, must be limited. - Generic Interfaces The system must offer generic interfaces to applications and clients. In particular, it preferably supports the POSIX file system interface [10] as closely as possible; in that way, a wide range of different applications can be supported. It is understood that the POSIX standard defines local file system operations and semantics and that these may be selectively extended and adapted to better support performance in a distributed setting (e.g., 8
9 consistency semantics may be relaxed to obtain better write performance). An example of such an extension is the POSIX I/O extension for high performance computing (HPC) [16]. Alternatively, the system may support a block device interface, because such an interface can be used to implement arbitrary file systems. Note that it might be easier to apply smart placement policies with objects (files) than with raw blocks. - Protocols Based on Open Standards The system must be built using protocols based on open standards as much as possible. This will reduce the chance on vendor lock-in and improves extensibility. It is likely that a storage service, based on open standards, in the long run will be more economic to maintain then a proprietary based service. - Multi-Party Access The system must support access by multiple, geographically dispersed parties at the same time. This enables collaboration between these parties over long distances on the same data. 9
10 2 Overview of existing products Different kinds of distributed storage systems have evolved over the years. Early products with initial implementations appearing in the mid-80 s were based on a client-server paradigm, with a server at a single location storing all the data and multiple clients accessing this data. These network file systems are now very widely used in local area networks, in particular systems based on the Common Internet File System (CIFS) and the Network File System (NFS). Other systems with more explicit focus on multiple servers that store data are the Andrew File System (AFS) and the DCE Distributed File System (DFS). These early systems have in common that they do not operate well in a wide area network environment and that they are not fault tolerant. More recently, cloud storage technologies have emerged that operate well in the Internet at large, and also, in many cases, provide facilities for durability in the form of data replication. The best known cloud storage services, such as Amazon S3 and Google Storage, are not available as standalone software products and therefore cannot be implemented on own hardware. The Hadoop Distributed File System [8] and CloudStore [3] are open source products that can be used, but by default these provide interfaces at the application level 1, and work with a single metadata server (which is a single point of failure). Therefore we do not consider these products here. Like traditional distributed file systems and cloud storage systems, a whole range of other products have interesting features and do at least meet some of the requirements we described in the previous section. Many of them, however, focus on dedicated environments or are designed for specific applications. An example is OCFS2 [14], which provides a shared file system dedicated for operation in SANs. Another example is dcache [4], which has been developed to store the large amounts of data generated by experiments with the Large Hadron Collider at CERN. As an initial filter for product selection, we picked those that claim to be distributed, fault-tolerant and having supporting parallel access. We observed that a broad classification of products can be made by looking at the used data primitive and by the handling of metadata. The data primitive is the form in which data is handled within the system at the most basic level, which can be either at the block level, or at the object or file level. For the handling of metadata, there is a distinction between products that treat metadata separate from regular data by storing it on dedicated nodes, and products that treat metadata in the same way as regular data (i.e., storing metadata and regular data on the same nodes). In this section we provide an overview of existing products and technologies that may be used to implement a distributed storage system that fulfills the requirements completely or to a large extent. The products are subdivided in those 1 Both HDFS and CloudStore have facilities to mount the file system under Linux using FUSE. 10
11 that are most promising (the candidate products; section 2.1) and those that meet many requirements but are not further considered for evaluation (the non-candidate products; section 2.2). Furthermore, a number of building blocks are described (in section 2.3) that may be used to extend or improve a selected product. 2.1 Candidate products The candidate products described here have a good match with the requirements identified in the previous section. The list of candidates will be used to make a final selection of products that will be tested and evaluated in an wide-area distributed environment with different types of hardware and interconnects between the storage nodes. The order in which the candidates are described does not indicate the order of preference Lustre Lustre [12] is a massively parallel distributed file system running on Linux and used at many high-performance computing (HPC) centres worldwide. Originally developed at CMU in 1999, it is now owned by Oracle and the software is available as open source (under a GNU GPL license). The Lustre architecture defines different kinds of roles and nodes, following an object storage paradigm where metadata is separated from the file data. A typical Lustre cluster can have tens of thousands of clients, thousands of object storage servers (OSDs) and a failover pair of metadata servers (currently still a work in progress, metadata servers will in the future be able to form a cluster comprising of dozens of nodes). Lustre assumes that OSDs are reliable, i.e., use such techniques as RAID to prevent data loss. Servers can currently be added dynamically and the file system is POSIX compliant. Other features that Lustre provides are ADIO interfaces, it can disable locking, and perform direct I/O that is usable for databases. Lustre also has other tuneable settings. It can currently be installed on Linux where it can interoperate amongst all supported processor architectures. It is still being developed for other operating systems. The main advantage of Lustre is that it has very high parallel performance, it also has good file I/O and can handle requests for thousands of files. Lustre does not seem to be deployed frequently in clusters that stretch over large distances 2, however, which raises questions about the performance and other characteristics when nodes are interconnected by a wide-area network. Additionally, Lustre appears to have little support for tiered operation, and setting policies for placement of files at particular tiers. 2 Lustre operation in a wide-area network setting with sites at various locations in the US is currently explored at Teragrid / PSC (focusing on high-performance computing applications) 11
12 2.1.2 GlusterFS GlusterFS [7] is a parallel network file system developed by Gluster Inc., which is used primarily for clustered storage consisting of a (potentially large) number of simple storage servers (also referred to as storage bricks ). GlusterFS is an open source product, available under the GNU GPL license. GlusterFS stores data at the file level (not at the block level) and does not use, contrary to many other similar products, separate metadata nodes. Instead, the location of files is found through a hash algorithm, which maps file names to storage servers. This algorithm takes into account that storage nodes may join and leave the system dynamically and, according to Gluster, makes the system scale in a linear fashion. All storage aggregates into a single global namespace. GlusterFS has a POSIX file system interface for general purpose access and a dedicated client library for direct access to storage nodes. Most functionality of the system resides at the client with the server nodes being relatively simple. The client software consists of modules with dedicated responsibilities such as volume management and file replication. All access to the storage nodes is done by clients, i.e., server nodes do not exchange data. The typical GlusterFS environment seems to be a high-performance, high-capacity datacenter with a fast, low-latency local network (such as Infiniband RDMA) between clients and servers. Any commodity hardware can be used to implement a storage brick and scale the system to several peta-bytes. GlusterFS can handle thousands of clients. It also includes a configuration utility. Like Lustre, it is unclear how well GlusterFS will operate in a wide area network. For instance, to which nodes will be written by a client when files must be redundantly stored? For the sake of durability, these files should be far apart, but for the sake of write performance, these files should be written to storage nodes close to the client GPFS The General Parallel File System (GPFS) [9] is a commercial shared-disk clustered file system from IBM. GPFS is being used by many large supercomputer centers because of its distributed file system capabilities and high speed parallel file access. Other areas in which GPFS is being used are streaming digital media, grid computing and scalable file storage. The GPFS file system works like a traditional UNIX file system. From a user/application level perspective this is very convenient. GPFS works internally with data blocks instead of objects. In order to grant concurrent access from multiple applications or nodes to the same file and to keep the data consistent, GPFS has developed a token management system on the block level. The size of these blocks is configurable and these data blocks can be striped across multiple nodes and disk. This is beneficial for the throughput of the file system. To improve the performance further it is possible to connect GPFS to a SAN/NAS via fiber channel instead of using local disks. Other performance improvements are cache mechanism, read ahead and write behind a file. 12
13 A very interesting feature of GPFS is its capability to define storage tiers. Different tiers can be created based on the location or performance of the hardware. Data can be dynamically moved between tiers based on the policies that have been created (a placement policy allows to determine the best tier based on simple file characteristics). This gives the opportunity to dynamically shift data, depending on the actual usage, between a faster or slower tier. When designing a multi tier GPFS file system the complexity of the policies and management of the storage will increase while adding additional tiers. Another interesting feature is the concept of failure groups, which allows GPFS to replicate data between hardware components that do not share a single point of failure, or for instance between geographically dispersed locations. GPFS provides a scalable metadata management instead of a centralize metadata server that is common in other file systems. This makes all GPFS nodes of the cluster involved with metadata management operations. Practically when a file is accessed on a certain node, that particular node is responsible for metadata management of that file. In case of parallel access a dynamically selected node will be made authoritative. The features mentioned above have provided GPFS a unique position in the widearea file system research environment. An example of such a usage is TeraGrid [22] in which the nodes are geographically spread in the USA. Such a large cluster setup can be build on top of commodity hardware, although commercial support is only available for specific hardware brands. Access to the data stored in a GPFS environment is granted via these interfaces: fiber channel, CIFS, NFS, HTTP, SCP and FTP. Unfortunately a common protocol as iscsi is not officially supported for system-internal data storage Ceph Ceph [2] is a distributed file system originally designed by the Storage Systems Research Center at the University of California, Santa Cruz. It is developed as an open source project, with code under the GNU LGPL license, and, since May 2010, has its client integrated in the Linux kernel. The objectives of Ceph are to provide a fully distributed file system without a single point of failure, with a POSIX-style interface. It claims high I/O performance and a high level of scalability. Ceph is based on an object storage paradigm, where file data is stored by object storage devices (OSDs) and metadata is stored by metadata servers (MDSs). Contrary to some distributed file systems relying on dumb OSDs, the Ceph OSDs have responsibilities for data migration, replication and failure handling and communicate between each other. Metadata management is completely distributed, using a cluster of MDSs to handle metadata request from clients. The operation is adapted dynamically based on the workload generated by the clients (e.g., moving and replicating metadata depending on how often a file is accessed). An MDS does not keep track of which OSDs store the data for a particular file. Instead, Ceph uses a special function called CRUSH to determine the location of 13
14 objects on storage nodes: it first maps an object to a placement group, and then calculates which OSDs belong to that placement group (and provides an ordering in the list of OSDs for a placement group). While doing so, it takes care of the replication of file data on different OSDs. CRUSH automatically takes into account that the set of storage nodes is dynamic over time. To clients, the data within a Ceph configuration (potentially consisting of thousands of OSDs) is presented as a single logical object store called RADOS. Replication of data is organized by writing to the first OSD in the placement group, after which this OSD replicates the data to others. The client receives an ack when all data has reached the buffer caches on all OSDs, and receives a commit when the data has been safely stored on all involved OSDs. RADOS has mechanisms for failure detection and automatic re-replication. Furthermore, Ceph implements a mechanism for recovery in case of system outages or large configuration changes. The features of Ceph as described above provide a reasonably good match with the requirements. However, a number of issues remain unclear. A major drawback is the immaturity of Ceph as a platform: as far as we know, it has not been widely used in a production environment, and the Ceph documentation also explicitly warns for the beta state of the code. Another issue is the uncertainty about the operation of Ceph in a WAN environment (as is the case for other products). The placement of the file data at OSDs does not take into account that links between storage nodes have variable quality (bandwidth, latency). Additionally, the operation of the mechanisms for automatic adaptation (adjustment of placement group to OSD mapping, and failure detection) may be non-optimal or worse for a WAN Ceph configuration Candidate product comparison Lustre GlusterFS GPFS Ceph Owner Oracle Gluster IBM Newdream (?) License GNU GPL GNU GPL commercial GNU LGPL Data object (file) object (file) block object (file) primitive Data placement strategy Metadata handling Storage tiers based on round robin and free space heuristics max. of 2 metadata servers (> 2 in beta version) pools of object storage targets different strategies through plugin modules stored with file data on storage servers policy based metadata distributed over storage servers placement groups, pseudorandom mapping multiple metadata servers unknown policy defined through CRUSH rules 14
15 Failure handling Replication WAN example deployment client interfacing Node types Lustre GlusterFS GPFS Ceph assuming assuming assuming assuming reliable nodes unreliable reliable nodes unreliable nodes nodes server side client side server side server side (failover pairs) TeraGrid [22] (scientific data) native client file system, FUSE, clients may export NFS, CIFS client, metadata, object no known deployment native library, FUSE TeraGrid [22] (scientific data) native client file system, clients may export NFS, no known deployment native client file system, FUSE CIFS,etc. client, data client, data client, metadata, object 2.2 Non-candidate products XtreemFS XtreemFS [26] is a globally distributed and replicated file system that has been developed in order to make grid data available in a distributed environment. It is developed as part of the XtreemOS EU project [27] which aims at creating an open source grid operating system. XtreemFS is a multiplatform file system, so client and server components can be installed on most common platforms (Linux, OS X, Windows). XtreemOS is an object-based file system, with metadata and regular data stored on different types of nodes. XtreemFS is POSIX compatible, failure tolerant and can be installed on commodity hardware. One big disadvantage of XtreemFS is the lack of support for a tiered storage approach: it will therefore not have a performance at the traditional SAN/NAS level MogileFS MogileFS [13] is an open source distributed file system with specific focus on data archiving and deployment on commodity hardware. It is fault tolerant (no single point of failure) by spreading data and metadata over different server nodes, where the replication level depends on the type of the file. MogileFS defines three different kinds of nodes (trackers, storage, and database) of which multiple instances may exist in a given configuration. A tracker is responsible for handling client sessions and requests, a database node stores file system metadata, and storage nodes store the actual data. Although MogileFS can be accessed through a variety of APIs and libraries, it does not provide a POSIX or block device interface to clients. It therefore is not suitable for our purposes. 15
16 2.2.3 NFS version 4.1 (pnfs) NFS [17] is a network file system that has been in common use for many years. It allows users to access files over a network similar to the manner they would access local storage. The protocol is an open protocol with many implementations from different vendors (also open source). NFS version 4 minor version 1 (NFSv4.1) has been approved by the IESG and received an RFC number in January Apart from bug fixes, the NFSv4.1 specification aims to provide further protocol support that will enable users to take advantage of secure, clustered server deployments. The ability to provide parallel scalable access to files when they are distributed among multiple servers is also supported. Various pnfs implementations are currently available (Linux, Solaris, ). Our impression is that these implementations are not stable enough to use in our environment, unfortunately. Aside from this, pnfs looks promising and may be reconsidered to become a candidate product ZFS ZFS [28] has been designed by Sun Microsystems (now Oracle) and is both a file system and a logical volume manager. ZFS provides support for (amongst others) the following features: very high storage capacities, snapshots, copy-on-write clones, integrity checking, automatic repair, RAID-Z (RAID-5 and RAID-6 via RAID- Z2) and native NFSv4 ACLs. It has been implemented as an open source product and can be freely downloaded. ZFS has many features used for error correction, hardware failure etc. Another useful feature is that block devices can be grouped according to the physical implementation (i.e. chassis), which allows the file system to continue in the case of a failure of an entire chassis. The main disadvantage is that ZFS is a local file system, not designed to run in a clustered, widely distributed network environment. This makes ZFS unsuitable for our purposes VERITAS File System The VERITAS File system [20] is a file system that was developed by VERITAS Software (now owned by Symantec), and capable to run as a cluster storage system. It runs on a variety of operating systems. The file-system can perform online defragmentation and resizing, and when running in clustered mode supports fail over between storage nodes. To ensure data consistency within the cluster, VERITAS uses a strict consistency model between storage nodes. A maximum of 32 storage nodes within the cluster is supported. VERITAS is focused on clustered storage within a local network. Due to its strict coherency model and its limit in number of nodes within a single cluster, it is not suitable to run as a wide area distributed system ParaScale ParaScale [15] is a private cloud solution. The basic belief is that a private cloud should be easy to manage and to scale. Scaling is easy because commodity 16
17 hardware can be added or extracted from the cloud when needed without much hassle. Files can be transferred into Parascale via protocols like nfs, ftp, http or webdav. The advantage of these protocols is that most applications do support them ParaScale can provide massive write bandwidth in parallel across multiple storage nodes. This enables it to provide an ideal solution for archiving near-line storage and provision of disk backup. It can cluster tens to hundreds of servers together; this can then be used to provide large file repositories with good parallel throughput. Parascale is a typical cloud storage solution and thus only using a single-tier architecture. In this way this product is not capable of handling a SAN/NAS-like performance CAStor CAStor [1] is an object-based storage software platform that can run on commodity hardware. It provides high performance and good scalability and is quite costeffective. CAStor virtualizes storage capacity creating a single pool of storage. It has the ability to scale easily and is thus able to meet dynamic capacity demands. It has a number of internal algorithms that perform self-management and self-healing. Additionally, CAStor supports the operation in a WAN environment. CAStor has a flat, single-tier architecture, which is its main disadvantage considering our requirements. By not dealing with multiple tiers (each providing different levels of performance), it is unlikely that CAStor will be able to support the traditional SAN/NAS performance requirement Tahoe-LAFS Tahoe [21] is an open source distributed file system that supports storing files into a network of peers. It is the enabling technology behind the Allmydata cloud storage company, where it is used to organize the back-end servers in a P2P manner. All data in a Tahoe configuration is written in an encrypted manner to the storage nodes. It uses erasure coding to spread the pieces of a file over a number of nodes in a redundant way and to improve data durability. It runs on various popular platforms such as Windows, OS X, Linux, etc. and is now part of the Ubuntu Linux distribution. Tahoe provides very little control over on which nodes data is stored, which makes it not suitable for tiered functionality. Furthermore, Tahoe assumes a flat, local network environment and therefore is not suitable to run in a WAN. These aspects make Tahoe unsuitable for our purposes DRBD DRBD (Distributed Replicated Block Device) [5] is a software-based storage replication program that allows a system administrator to mirror the content of block devices across different servers. It is released under the GNU GPL license. DRBD can do this replication in real time, transparently and synchronously or asynchronously. DRBD is implemented by way of a core Linux kernel module. It 17
18 exports a block device on top of which a file system can be configured, which makes DRBD quite flexible and versatile. Although DRBD does not offer the features of a complete distributed file system, and therefore is not suitable to build a system matching our requirement, it may be used as a building block that complements other product. 18
19 3 Future work To further explore the characteristics of candidate products, we will install and configure one or multiple products in a distributed test environment. This allows us to obtain hands-on experience with the chosen system(s) and provides us with markers to better assess the suitability for deployment on a larger scale. In this section we give short description of the environment we consider suitable for a firststep evaluation. Figure 1: example evaluation environment consisting of two sites connected by a high-speed, low latency WAN. The nodes at each site have a heterogeneous configuration and are interconnected through different LAN types. The objective is to obtain a number of basic performance indicators and evaluate a few straightforward use cases in a wide area setup with a limited number of nodes. We realize that a limited test environment as proposed here is not capable of a 19
20 broad system test investigating aspects such as scalability, large system performance and fault tolerance; such a test would require a very large setup and considerable resources to execute. In terms of distribution, the evaluation environment must consist of at least two sites, between which there is a reasonable distance, i.e., sites that are not part of the same metropolitan area. A distance of at least 100 km will introduce a packet delay that cannot be neglected when data is accessed and replicated between sites. The WAN connection between the sites must be realized using the SURFnet infrastructure and must have a considerable capacity. Each site must consist of multiple nodes that can take on different kinds of roles in the larger storage system (object storage node, metadata node, client nodes, ). As is the case in data centers, the nodes within a site are connected through a high speed LAN. To explore system characteristics in different circumstance, multiple local interconnecting technologies may be used (e.g. Gigabit Ethernet, Infiniband, ). Nodes consist of different kinds of hardware, to reflect that the storage system runs in a heterogeneous setting: within the nodes, different kinds of storage technologies such as SSD, SAS disks and RAID controllers are available for different kinds of nodes. At each site, a sufficient amount of nodes can act as client nodes, to generate test load that can stretch the system to its limits. 20
21 4 Conclusion This report provides an oversight of technologies for wide area distributed storage. It is a first step of a larger project that aims at designing a distributed storage facility that combines features of traditional SAN/NAS storage and cloud storage for applications in research and corporate datacenters. This report concludes that four products Lustre, GlusterFS, GPFS and Ceph are promising to investigate further. Surprisingly only one of these products, namely GPFS, has a purely commercial license; the others are provided with an open-source license. GPFS and Lustre are the most mature products and have a known track record as a file system in high performance computing environments. GlusterFS is also reasonably wide deployed, but Ceph is still under development. Unfortunately all promising product have been developed with a different mindset than what we are aiming for in this project. So it might be a challenge to tune and tweak these products so they will match our needs. So far only literature has been consulted to decide which products are promising. A next step in this project will be a more hands-on investigation of these products to obtain more in-depth knowledge. This investigation will be conducted in a distributed test environment within the SURFnet network. The outcome of this investigation, based on the requirements that have been set, will be a shortlist of two products that are most viable for an implementation phase. Finally a proof of concept architecture will be build in cooperation with a pilot partner and after a pilot phase the final product will be selected. A report on the experiences of this product and a proposal for a distributed storage design will also be delivered. 21
22 5 References [1] CAstor storage software, [2] Ceph open source distributed storage, [3] CloudStore home page, [4] dcache home page, [5] DRBD Home Page, [6] S. Gilbert and N. Lynch, Brewer s Conjecture and the Feasibility of Consistent, Available, Partition-Tolerant Web Services, ACM SIGACT News, Vol. 33, Issue 2, 2002 [7] Gluster Community, [8] Hadoop Distributed File System, [9] IBM General Parallel File System, [10] Institute of Electrical and Electronics Engineers (IEEE) Standard / POSIX:2008, Base Specifications, Issue 7, December 2008 [11] Isilon systems, [12] Lustre home page, [13] MogileFS home page, [14] OCFS2 project home, [15] ParaScale cloud storage software, [16] Posix Extensions for High-Performance Computing, [17] S. Shepler, M. Eisler, and D. Noveck, Network File System (NFS) Version 4 Minor Version 1 Protocol, RFC 5661, Internet Engineering Task Force, January 2010 [18] A. Shoshani, and D. Rotem (Ed.), Scientific Data Management: Challenges, Technology and Deployment, Chapman & Hall/CRC, 2010 [19] SURFnet lightpaths, px [20] Symantec storage solutions, [21] Tahoe The Least-Authority Filesystem (LAFS), [22] TeraGrid, [23] F. Travostino, P. Dasit, L. Gommans, C. Jog, C. de Laat, J. Mambretti, I. Monga, B. van Oudenaarde, S. Raghunath, and P. Wang, Seamless Live Migration of Virtual Machines over the MAN/WAN, Future Generation Computer Systems, Vol. 22, Issue 8, 2006 [24] M. Wiesmann, F. Pedone, A. Schiper, B. Kemme, and G. Alonso, Understanding Replication in Databases and Distributed Systems, In Proceedings of the International Conference on Distributed Computing Systems (ICDCS 00), April 2000 [25] T. Wood, K. Ramakrishnan, J. van der Merwe, and P. Shenoy. CloudNet: A Platform for Optimized WAN Migration of Virtual Machines, University of Massachusetts Technical Report TR , January 2010 [26] XtreemFS a cloud file system, 22
23 [27] XtreemOS: a Linux-based Operating System to Support Virtual Organizations for Next Generation Grids, [28] ZFS home page, 23
Introduction to Gluster. Versions 3.0.x
Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster
Using DRBD over Wide Area Networks
Using DRBD over Wide Area Networks Project : GigaPort3 Project Year : 2010 Project Manager : Rogier Spoor Author(s) : Arjan Peddemors, Rogier Spoor, Paul Dekkers, Christiaan den Besten Completion Date
POWER ALL GLOBAL FILE SYSTEM (PGFS)
POWER ALL GLOBAL FILE SYSTEM (PGFS) Defining next generation of global storage grid Power All Networks Ltd. Technical Whitepaper April 2008, version 1.01 Table of Content 1. Introduction.. 3 2. Paradigm
UNINETT Sigma2 AS: architecture and functionality of the future national data infrastructure
UNINETT Sigma2 AS: architecture and functionality of the future national data infrastructure Authors: A O Jaunsen, G S Dahiya, H A Eide, E Midttun Date: Dec 15, 2015 Summary Uninett Sigma2 provides High
Scala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
Big data management with IBM General Parallel File System
Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers
Data Storage in Clouds
Data Storage in Clouds Jan Stender Zuse Institute Berlin contrail is co-funded by the EC 7th Framework Programme 1 Overview Introduction Motivation Challenges Requirements Cloud Storage Systems XtreemFS
Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000
Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000 Clear the way for new business opportunities. Unlock the power of data. Overcoming storage limitations Unpredictable data growth
StorPool Distributed Storage Software Technical Overview
StorPool Distributed Storage Software Technical Overview StorPool 2015 Page 1 of 8 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard
Big Data Storage Options for Hadoop Sam Fineberg, HP Storage
Sam Fineberg, HP Storage SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies and individual members may use this material in presentations
Testing of several distributed file-system (HadoopFS, CEPH and GlusterFS) for supporting the HEP experiments analisys. Giacinto DONVITO INFN-Bari
Testing of several distributed file-system (HadoopFS, CEPH and GlusterFS) for supporting the HEP experiments analisys. Giacinto DONVITO INFN-Bari 1 Agenda Introduction on the objective of the test activities
New Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
Designing a Cloud Storage System
Designing a Cloud Storage System End to End Cloud Storage When designing a cloud storage system, there is value in decoupling the system s archival capacity (its ability to persistently store large volumes
How to Choose your Red Hat Enterprise Linux Filesystem
How to Choose your Red Hat Enterprise Linux Filesystem EXECUTIVE SUMMARY Choosing the Red Hat Enterprise Linux filesystem that is appropriate for your application is often a non-trivial decision due to
Product Spotlight. A Look at the Future of Storage. Featuring SUSE Enterprise Storage. Where IT perceptions are reality
Where IT perceptions are reality Product Spotlight A Look at the Future of Storage Featuring SUSE Enterprise Storage Document # SPOTLIGHT2013001 v5, January 2015 Copyright 2015 IT Brand Pulse. All rights
Storage Architectures for Big Data in the Cloud
Storage Architectures for Big Data in the Cloud Sam Fineberg HP Storage CT Office/ May 2013 Overview Introduction What is big data? Big Data I/O Hadoop/HDFS SAN Distributed FS Cloud Summary Research Areas
(Scale Out NAS System)
For Unlimited Capacity & Performance Clustered NAS System (Scale Out NAS System) Copyright 2010 by Netclips, Ltd. All rights reserved -0- 1 2 3 4 5 NAS Storage Trend Scale-Out NAS Solution Scaleway Advantages
XtreemFS Extreme cloud file system?! Udo Seidel
XtreemFS Extreme cloud file system?! Udo Seidel Agenda Background/motivation High level overview High Availability Security Summary Distributed file systems Part of shared file systems family Around for
IBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand.
IBM Global Technology Services September 2007 NAS systems scale out to meet Page 2 Contents 2 Introduction 2 Understanding the traditional NAS role 3 Gaining NAS benefits 4 NAS shortcomings in enterprise
CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT
SS Data & Storage CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT HEPiX Fall 2012 Workshop October 15-19, 2012 Institute of High Energy Physics, Beijing, China SS Outline
Block based, file-based, combination. Component based, solution based
The Wide Spread Role of 10-Gigabit Ethernet in Storage This paper provides an overview of SAN and NAS storage solutions, highlights the ubiquitous role of 10 Gigabit Ethernet in these solutions, and illustrates
Sep 23, 2014. OSBCONF 2014 Cloud backup with Bareos
Sep 23, 2014 OSBCONF 2014 Cloud backup with Bareos OSBCONF 23/09/2014 Content: Who am I Quick overview of Cloud solutions Bareos and Backup/Restore using Cloud Storage Bareos and Backup/Restore of Cloud
Research Data Storage Infrastructure (RDSI) Project. DaSh Straw-Man
Research Data Storage Infrastructure (RDSI) Project DaSh Straw-Man Recap from the Node Workshop (Cherry-picked) *Higher Tiered DCs cost roughly twice the cost of Lower Tiered DCs. * However can provide
IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE
White Paper IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE Abstract This white paper focuses on recovery of an IBM Tivoli Storage Manager (TSM) server and explores
WHITE PAPER. Software Defined Storage Hydrates the Cloud
WHITE PAPER Software Defined Storage Hydrates the Cloud Table of Contents Overview... 2 NexentaStor (Block & File Storage)... 4 Software Defined Data Centers (SDDC)... 5 OpenStack... 5 CloudStack... 6
BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything
BlueArc unified network storage systems 7th TF-Storage Meeting Scale Bigger, Store Smarter, Accelerate Everything BlueArc s Heritage Private Company, founded in 1998 Headquarters in San Jose, CA Highest
Building Storage as a Service with OpenStack. Greg Elkinbard Senior Technical Director
Building Storage as a Service with OpenStack Greg Elkinbard Senior Technical Director MIRANTIS 2012 PAGE 1 About the Presenter Greg Elkinbard Senior Technical Director at Mirantis Builds on demand IaaS
How To Build A Clustered Storage Area Network (Csan) From Power All Networks
Power-All Networks Clustered Storage Area Network: A scalable, fault-tolerant, high-performance storage system. Power-All Networks Ltd Abstract: Today's network-oriented computing environments require
ovirt and Gluster Hyperconvergence
ovirt and Gluster Hyperconvergence January 2015 Federico Simoncelli Principal Software Engineer Red Hat ovirt and GlusterFS Hyperconvergence, Jan 2015 1 Agenda ovirt Architecture and Software-defined Data
The Design and Implementation of the Zetta Storage Service. October 27, 2009
The Design and Implementation of the Zetta Storage Service October 27, 2009 Zetta s Mission Simplify Enterprise Storage Zetta delivers enterprise-grade storage as a service for IT professionals needing
June 2009. Blade.org 2009 ALL RIGHTS RESERVED
Contributions for this vendor neutral technology paper have been provided by Blade.org members including NetApp, BLADE Network Technologies, and Double-Take Software. June 2009 Blade.org 2009 ALL RIGHTS
Accelerating and Simplifying Apache
Accelerating and Simplifying Apache Hadoop with Panasas ActiveStor White paper NOvember 2012 1.888.PANASAS www.panasas.com Executive Overview The technology requirements for big data vary significantly
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
Red Hat Storage Server
Red Hat Storage Server Marcel Hergaarden Solution Architect, Red Hat [email protected] May 23, 2013 Unstoppable, OpenSource Software-based Storage Solution The Foundation for the Modern Hybrid
EMC VPLEX FAMILY. Continuous Availability and data Mobility Within and Across Data Centers
EMC VPLEX FAMILY Continuous Availability and data Mobility Within and Across Data Centers DELIVERING CONTINUOUS AVAILABILITY AND DATA MOBILITY FOR MISSION CRITICAL APPLICATIONS Storage infrastructure is
Storage Virtualization in Cloud
Storage Virtualization in Cloud Cloud Strategy Partners, LLC Sponsored by: IEEE Educational Activities and IEEE Cloud Computing Course Presenter s Biography This IEEE Cloud Computing tutorial has been
The Panasas Parallel Storage Cluster. Acknowledgement: Some of the material presented is under copyright by Panasas Inc.
The Panasas Parallel Storage Cluster What Is It? What Is The Panasas ActiveScale Storage Cluster A complete hardware and software storage solution Implements An Asynchronous, Parallel, Object-based, POSIX
High Availability with Windows Server 2012 Release Candidate
High Availability with Windows Server 2012 Release Candidate Windows Server 2012 Release Candidate (RC) delivers innovative new capabilities that enable you to build dynamic storage and availability solutions
Netapp @ 10th TF-Storage Meeting
Netapp @ 10th TF-Storage Meeting Wojciech Janusz, Netapp Poland Bogusz Błaszkiewicz, Netapp Poland Ljubljana, 2012.02.20 Agenda Data Ontap Cluster-Mode pnfs E-Series NetApp Confidential - Internal Use
Ultimate Guide to Oracle Storage
Ultimate Guide to Oracle Storage Presented by George Trujillo [email protected] George Trujillo Twenty two years IT experience with 19 years Oracle experience. Advanced database solutions such
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
XtreemFS a Distributed File System for Grids and Clouds Mikael Högqvist, Björn Kolbeck Zuse Institute Berlin XtreemFS Mikael Högqvist/Björn Kolbeck 1
XtreemFS a Distributed File System for Grids and Clouds Mikael Högqvist, Björn Kolbeck Zuse Institute Berlin XtreemFS Mikael Högqvist/Björn Kolbeck 1 The XtreemOS Project Research project funded by the
Diagram 1: Islands of storage across a digital broadcast workflow
XOR MEDIA CLOUD AQUA Big Data and Traditional Storage The era of big data imposes new challenges on the storage technology industry. As companies accumulate massive amounts of data from video, sound, database,
Getting performance & scalability on standard platforms, the Object vs Block storage debate. Copyright 2013 MPSTOR LTD. All rights reserved.
Getting performance & scalability on standard platforms, the Object vs Block storage debate 1 December Webinar Session Getting performance & scalability on standard platforms, the Object vs Block storage
Quantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
High Availability Databases based on Oracle 10g RAC on Linux
High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN, June 2006 Luca Canali, CERN IT Outline Goals Architecture of an HA DB Service Deployment at the CERN Physics Database
Ceph. A file system a little bit different. Udo Seidel
Ceph A file system a little bit different Udo Seidel Ceph what? So-called parallel distributed cluster file system Started as part of PhD studies at UCSC Public announcement in 2006 at 7 th OSDI File system
Technology Insight Series
Evaluating Storage Technologies for Virtual Server Environments Russ Fellows June, 2010 Technology Insight Series Evaluator Group Copyright 2010 Evaluator Group, Inc. All rights reserved Executive Summary
Private Cloud Storage for Media Applications. Bang Chang Vice President, Broadcast Servers and Storage [email protected]
Private Cloud Storage for Media Bang Chang Vice President, Broadcast Servers and Storage [email protected] Table of Contents Introduction Cloud Storage Requirements Application transparency Universal
GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
Building Storage Service in a Private Cloud
Building Storage Service in a Private Cloud Sateesh Potturu & Deepak Vasudevan Wipro Technologies Abstract Storage in a private cloud is the storage that sits within a particular enterprise security domain
Distributed File System Choices: Red Hat Storage, GFS2 & pnfs
Distributed File System Choices: Red Hat Storage, GFS2 & pnfs Ric Wheeler Architect & Senior Manager, Red Hat June 27, 2012 Overview Distributed file system basics Red Hat distributed file systems Performance
Analisi di un servizio SRM: StoRM
27 November 2007 General Parallel File System (GPFS) The StoRM service Deployment configuration Authorization and ACLs Conclusions. Definition of terms Definition of terms 1/2 Distributed File System The
OPTIMIZING PRIMARY STORAGE WHITE PAPER FILE ARCHIVING SOLUTIONS FROM QSTAR AND CLOUDIAN
OPTIMIZING PRIMARY STORAGE WHITE PAPER FILE ARCHIVING SOLUTIONS FROM QSTAR AND CLOUDIAN CONTENTS EXECUTIVE SUMMARY The Challenges of Data Growth SOLUTION OVERVIEW 3 SOLUTION COMPONENTS 4 Cloudian HyperStore
Virtual SAN Design and Deployment Guide
Virtual SAN Design and Deployment Guide TECHNICAL MARKETING DOCUMENTATION VERSION 1.3 - November 2014 Copyright 2014 DataCore Software All Rights Reserved Table of Contents INTRODUCTION... 3 1.1 DataCore
Introduction to NetApp Infinite Volume
Technical Report Introduction to NetApp Infinite Volume Sandra Moulton, Reena Gupta, NetApp April 2013 TR-4037 Summary This document provides an overview of NetApp Infinite Volume, a new innovation in
A Virtual Filer for VMware s Virtual SAN A Maginatics and VMware Joint Partner Brief
A Virtual Filer for VMware s Virtual SAN A Maginatics and VMware Joint Partner Brief With the massive growth of unstructured data in today s enterprise environments, storage IT administrators are constantly
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011 Executive Summary Large enterprise Hyper-V deployments with a large number
GlusterFS Distributed Replicated Parallel File System
GlusterFS Distributed Replicated Parallel File System SLAC 2011 Martin Alfke Agenda General Information on GlusterFS Architecture Overview GlusterFS Translators GlusterFS
RED HAT STORAGE SERVER TECHNICAL OVERVIEW
RED HAT STORAGE SERVER TECHNICAL OVERVIEW Ingo Börnig Solution Architect, Red Hat 24.10.2013 NEW STORAGE REQUIREMENTS FOR THE MODERN HYBRID DATACENTER DESIGNED FOR THE NEW DATA LANDSCAPE PETABYTE SCALE
Cloud storage reloaded:
Cloud storage reloaded: Some aspects on operating distributed Linux file systems like Ceph and GlusterFS Udo Seidel Agenda Introduction/motivation Distributed storage Ceph and GlusterFS Operational considerations
Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.
Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance
EMC XTREMIO EXECUTIVE OVERVIEW
EMC XTREMIO EXECUTIVE OVERVIEW COMPANY BACKGROUND XtremIO develops enterprise data storage systems based completely on random access media such as flash solid-state drives (SSDs). By leveraging the underlying
Next Generation Tier 1 Storage
Next Generation Tier 1 Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies, Matthew Viljoen HEPiX Beijing 16th October 2012 Why are we doing this?
With Red Hat Enterprise Virtualization, you can: Take advantage of existing people skills and investments
RED HAT ENTERPRISE VIRTUALIZATION DATASHEET RED HAT ENTERPRISE VIRTUALIZATION AT A GLANCE Provides a complete end-toend enterprise virtualization solution for servers and desktop Provides an on-ramp to
Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
Microsoft Private Cloud Fast Track
Microsoft Private Cloud Fast Track Microsoft Private Cloud Fast Track is a reference architecture designed to help build private clouds by combining Microsoft software with Nutanix technology to decrease
Building Reliable, Scalable AR System Solutions. High-Availability. White Paper
Building Reliable, Scalable Solutions High-Availability White Paper Introduction This paper will discuss the products, tools and strategies available for building reliable and scalable Action Request System
Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage
Moving Virtual Storage to the Cloud Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage Table of Contents Overview... 1 Understanding the Storage Problem... 1 What Makes
High Availability with Postgres Plus Advanced Server. An EnterpriseDB White Paper
High Availability with Postgres Plus Advanced Server An EnterpriseDB White Paper For DBAs, Database Architects & IT Directors December 2013 Table of Contents Introduction 3 Active/Passive Clustering 4
Hitachi NAS Platform and Hitachi Content Platform with ESRI Image
W H I T E P A P E R Hitachi NAS Platform and Hitachi Content Platform with ESRI Image Aciduisismodo Extension to ArcGIS Dolore Server Eolore for Dionseq Geographic Uatummy Information Odolorem Systems
High Availability Storage
High Availability Storage High Availability Extensions Goldwyn Rodrigues High Availability Storage Engineer SUSE High Availability Extensions Highly available services for mission critical systems Integrated
Mirror File System for Cloud Computing
Mirror File System for Cloud Computing Twin Peaks Software Abstract The idea of the Mirror File System (MFS) is simple. When a user creates or updates a file, MFS creates or updates it in real time on
Amazon Cloud Storage Options
Amazon Cloud Storage Options Table of Contents 1. Overview of AWS Storage Options 02 2. Why you should use the AWS Storage 02 3. How to get Data into the AWS.03 4. Types of AWS Storage Options.03 5. Object
High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software
High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software White Paper Overview The Micron M500DC SSD was designed after months of close work with major data center service providers and
T a c k l i ng Big Data w i th High-Performance
Worldwide Headquarters: 211 North Union Street, Suite 105, Alexandria, VA 22314, USA P.571.296.8060 F.508.988.7881 www.idc-gi.com T a c k l i ng Big Data w i th High-Performance Computing W H I T E P A
Solaris For The Modern Data Center. Taking Advantage of Solaris 11 Features
Solaris For The Modern Data Center Taking Advantage of Solaris 11 Features JANUARY 2013 Contents Introduction... 2 Patching and Maintenance... 2 IPS Packages... 2 Boot Environments... 2 Fast Reboot...
EMC ISILON OneFS OPERATING SYSTEM Powering scale-out storage for the new world of Big Data in the enterprise
EMC ISILON OneFS OPERATING SYSTEM Powering scale-out storage for the new world of Big Data in the enterprise ESSENTIALS Easy-to-use, single volume, single file system architecture Highly scalable with
Zadara Storage Cloud A whitepaper. @ZadaraStorage
Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers
Understanding Microsoft Storage Spaces
S T O R A G E Understanding Microsoft Storage Spaces A critical look at its key features and value proposition for storage administrators A Microsoft s Storage Spaces solution offers storage administrators
Client/Server and Distributed Computing
Adapted from:operating Systems: Internals and Design Principles, 6/E William Stallings CS571 Fall 2010 Client/Server and Distributed Computing Dave Bremer Otago Polytechnic, N.Z. 2008, Prentice Hall Traditional
Clustering Windows File Servers for Enterprise Scale and High Availability
Enabling the Always-On Enterprise Clustering Windows File Servers for Enterprise Scale and High Availability By Andrew Melmed Director of Enterprise Solutions, Sanbolic, Inc. April 2012 Introduction Microsoft
Caringo Swarm 7: beyond the limits of traditional storage. A new private cloud foundation for storage needs at scale
Caringo Swarm 7: beyond the limits of traditional storage. A new private cloud foundation for storage needs at scale Prepared for: Caringo May 2014 TABLE OF CONTENTS TABLE OF CONTENTS 1 EXECUTIVE SUMMARY
THE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved.
THE EMC ISILON STORY Big Data In The Enterprise 2012 1 Big Data In The Enterprise Isilon Overview Isilon Technology Summary 2 What is Big Data? 3 The Big Data Challenge File Shares 90 and Archives 80 Bioinformatics
E4 UNIFIED STORAGE powered by Syneto
E4 UNIFIED STORAGE powered by Syneto THE E4 UNIFIED STORAGE (US) SERIES POWERED BY SYNETO From working in the heart of IT environment and with our major customers coming from Research, Education and PA,
Cisco Active Network Abstraction Gateway High Availability Solution
. Cisco Active Network Abstraction Gateway High Availability Solution White Paper This white paper describes the Cisco Active Network Abstraction (ANA) Gateway High Availability solution developed and
Object Storage: A Growing Opportunity for Service Providers. White Paper. Prepared for: 2012 Neovise, LLC. All Rights Reserved.
Object Storage: A Growing Opportunity for Service Providers Prepared for: White Paper 2012 Neovise, LLC. All Rights Reserved. Introduction For service providers, the rise of cloud computing is both a threat
BlobSeer: Towards efficient data storage management on large-scale, distributed systems
: Towards efficient data storage management on large-scale, distributed systems Bogdan Nicolae University of Rennes 1, France KerData Team, INRIA Rennes Bretagne-Atlantique PhD Advisors: Gabriel Antoniu
Optimizing Storage for Better TCO in Oracle Environments. Part 1: Management INFOSTOR. Executive Brief
Optimizing Storage for Better TCO in Oracle Environments INFOSTOR Executive Brief a QuinStreet Excutive Brief. 2012 To the casual observer, and even to business decision makers who don t work in information
Informix Dynamic Server May 2007. Availability Solutions with Informix Dynamic Server 11
Informix Dynamic Server May 2007 Availability Solutions with Informix Dynamic Server 11 1 Availability Solutions with IBM Informix Dynamic Server 11.10 Madison Pruet Ajay Gupta The addition of Multi-node
Scalable filesystems boosting Linux storage solutions
Scalable filesystems boosting Linux storage solutions Daniel Kobras science + computing ag IT-Dienstleistungen und Software für anspruchsvolle Rechnernetze Tübingen München Berlin Düsseldorf Motivation
Distributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
Cloud Optimize Your IT
Cloud Optimize Your IT Windows Server 2012 The information contained in this presentation relates to a pre-release product which may be substantially modified before it is commercially released. This pre-release
