Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers

Size: px
Start display at page:

Download "Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers"

Transcription

1 DDN Whitepaper Improving Time to Results for Seismic Processing with Paradigm and DDN James Coomer and Laurent Thiers 2014 DataDirect Networks. All Rights Reserved.

2 Executive Summary Companies in the oil and gas industry continue to push the limits of technology to gain a competitive advantage in their quest for the richest and most profitable reservoirs. For fastest time to results, Paradigm Echos and DataDirect Networks provide cost-effective processing and storage solutions that deliver the highest speed, precision, capacity, and scalability. This whitepaper summarizes Paradigm Echos workload testing performed on DDN storage, taking advantage of SSDs and high performance networks to improve the time to value for a seismic processing problem at lower cost. Targeted workload testing enables analysis of the bottlenecks in the interactions between the software, the network fabric and the storage systems; and removing these bottlenecks leads to improved efficiency. For example, if the number of jobs per compute node can be increased, we can either improve overall performance for a given number of systems or produce the same overall performance with fewer systems. The Need for I/O Improvements in Seismic Processing Environments Companies and partners engaged in oil and gas exploration and production recognize the value of speed and precision, investing tens of billions of dollars in new projects. With so much money at stake, there's little room for error when collecting the most precise and accurate information on where to drill, how much to bid on a site, and how to maximize reservoir performance. The third critical piece of the process requires companies to apply the most sophisticated modeling and simulation technology for the transformation of field data into valuable geophysics. While technology advances within the industry are adding much improved precision, they are also significantly increasing data and processing volumes. Wide/Multi/Rich-azimuth methods are using multi-sensor arrays and sophisticated methods to produce higher-fidelity images. The deployment of these richer formats results in a vast increase for the amount of data the company must process and manage. Furthermore, new analytics techniques allow for continued advancement in the interpretation of seismic data for new data coming in and for historical oil field data as well. This means more data needs to be analyzed at greater speeds. In the field of seismic data acquisition and processing, the conversion of raw field recordings to geological images is a highly intensive process in both data and compute. The requirement for processing more data with less resources (power and capital cost) tends to push the compute threads ever more densely onto the available compute clients, and this in-turn impacts on the I/O subsystems both network and storage. As core counts and memory bandwidths steadily rise, the challenge of shifting data to feed the compute nodes becomes increasingly important. Paradigm Echos: The Standard for Seismic Processing and Data Analysis For over 25 years, Paradigm Echos has been the oil and gas industry's leading system for seismic processing. Echos is a software suite that links seismic data and processing parameters and sequences, with both interactive and 2014 DataDirect Networks. All Rights Reserved. 2

3 batch components. The suite of seismic processing modules are provided to transform converted wave recordings to interpretable images of the subsurface, which are generally run in an embarrassingly parallel-like fashion with many simultaneous instances running on a large seismic cluster. Widely used, companies prefer Echos because of its breadth of geophysical applications, unique combination of production and interactive seismic data processing, maturity and stability, and versatile programming environment for client-guided customization. DDN: Driving Big Data Storage Solutions for Seismic Processing High-performance storage solutions from DDN meet the most pressing storage challenges of the seismic processing industry, helping oil and gas companies achieve both their scientific goals and their long-term business goals. To support the most demanding environments, DDN solutions deliver game-changing performance, capacity and scalability for ingest, processing and visualization. For example, DDN's Storage Fusion Architecture (SFA) combines SATA, SAS and solid-state disks into a simply managed, massively scalable, multi-petabyte platform that can be tailored to a balance of throughput and capacity. The combination of modularity and performance feature allows users in seismic processing environments to configure systems that scale up and/or scale out in independent dimensions. This is why over half the top producers worldwide are leveraging DDN storage solutions in their big data oil and gas workflows. PARADIGM and DDN Seismic data sets can be very large, with the raw recording data ranging into the 10s or 100s of Terabytes. The ability to efficiently view a particular seismic data set in different orders or domains, without creating multiple copies of that data set is a key design feature of the seismic file handling capabilities of the Echos system. Initial loading of raw seismic recording data is usually done in the order of which the data is recorded, based on sequential shot numbering and a secondary order associated with the receivers associated with each shot. However, this primary and secondary key ordering, shot-receiver ordered data may not be the most effective order for certain subsurface analysis techniques. The Common Depth Point (CDP) Method is a particular seismic data acquisition and processing technique that transforms field recordings from seismic surveys into pseudo-cross-sectional images of the earth's geologic layering beneath the surface locations. Such images are key to geophysical analysis used to pinpoint likely drilling locations for oil and gas reservoirs. Three micro-workloads are used in this study, which are components of the Paradigm Echos. All three involve heavy I/O, but in differing ways. Reading and writing a shot-receiver ordered data set in sequential order (SEQIN) efficiently is a critical I/O workload for conditioning of seismic data for further analysis. Another important I/O workload is reading a very large shotreceiver ordered data set in CDP order (STRIDE) which can result in file system access pattern that are challenging for 2014 DataDirect Networks. All Rights Reserved. 3

4 general-purpose storage systems. An un-tuned system can limit the read performance to just a few MB/sec. The third micro-workload (SEQOUT) performs sequential writes rather than reads. The seismic file handling codes underlying the Echos software are multi threaded, with different tasks for reading, processing, timing etc. In the Echos processing system, a logical seismic file consists of multiple actual files on the file systems that enable specialized handling of different logical elements that make up the seismic file. The SFA Advantage Currently supporting 70% of the Top10 supercomputers in the world, DDN s SFA technology combines 15 years of hardware optimization in High Performance Computing with sophisticated I/O software (SFAOS). The result is a highly efficient system that delivers the highest throughput, density and IOPs of any system on the market. The SFA product line is divided into the high-end SFA12KX and the mid-range SFA7700. Both run SFAOS, which leverages Storage Fusion Xcelerator (SFX) for hybrid deployments of traditional SAS drives with SSDs. SFX integrates application-centric intelligence with flash media to accelerate awkward read intensive workloads. The GS12K Appliance is the highest performing file system appliance in the world, with a single appliance delivering over 20GB/s to the clients. The system uses the SFAOS embedded technology to run complete parallel filesystem services within the controller. This approach removes the need for external file servers typically required for parallel file systems, allowing for far better single client performance than can be reached with scale-out NAS offerings. GS12K supports very large file systems with extreme density. A single appliance can start with just 60 drives, but be scaled up to 10PB in just two racks. If a larger file system is required, further appliances can be added into the single namespace. The GS12K also supports advanced Enterprise features such as windows support, snapshots, replication, HSM, and NFS/CIFS exports DataDirect Networks. All Rights Reserved. 4

5 GS12K Features Client Access and connectivity Native Linux and Windows Clients Highly Available NFS v3 Native Linux and Windows Clients Client Connectivity 16 x 10/40GbE Host Port or 16 x 56Gb In-finiBand Host Port Support for InfiniBand RDMA I/O Client to Gateway Load Balancing Many to One, and One to Many Client to Gateway Access Volume Scalability Up to 3.45PBs of usable capacity (840 drives) Optional Metadata accelerator drives Up to 3.45PBs of usable capacity (840 drives) Over 10PB+ of Aggregated Capacity Performance 20GB/s per GS12K System (read & write) Aggregated Performance up to 200GB/sec 100,000s of File Operations/sec Data Protection Data Protection Data: RAID 6 (8+2) Metadata: RAID Snapshots per file system DirectProtect Data Integrity Synchronous File System Replication Optimization of SFA with Paradigm DDN s history in deploying systems for seismic processing has included significant benchmarking and tuning with the largest oil and gas companies in the world. This report discusses the results of some particular major work with one such company where a large GS12K environment is deployed with 40Gigabit Ethernet. We present three sets of results, each of which are of significance to O&G seismic processing systems: 1. GS12KE baseline. Maximum data throughput of the GS12K in 40GE environment as measured by synthetic benchmarks 2. The potential for 40GE vs 10GE in heavy I/O seismic environments 3. Efficiency of Paradigm Echos I/O in large scale systems Data and Systems Centralized con figuration and Management The benchmark system comprised a single monitoring solution Online File System Growth GS12K system connected via 8 40GE ports to a Object-Based File Locking Monitoring and Event Notification System 40GE backbone. Each GS12K uses 5 SS8460 Disk Enclosures with between 300 and 400 NL-SAS drives for data. Data protection was provided via RAID 6. GRIDScaler metadata was held on a small dedicated set of drives separate from the data drives. The clients used varied in the test, but all were a mid-range clock speed Intel system connected, with either 10Gigabit Ethernet or 40Gigabit Ethernet. SAS 6Gb x DataDirect Networks. All Rights Reserved. 5

6 GS12KE Baseline The IOR benchmark ( is a popular parallel I/O benchmark used to characterise storage systems and file systems. IOR allows the benchmark engineer to run multiple threads, reading or writing from many clients simultaneously. A number of parameters are tuneable to avoid caching, alter the size of the data transfers issued from the IOR tasks and change the nature of the I/O topology. In the following case, we present results from executing IOR across 16 clients simultaneously. Each client is connected to the network with 40GE and runs between 8 and 32 threads per client, each thread read or writing with varying transfer sizes to a single shared file DataDirect Networks. All Rights Reserved. 6

7 Paradigm Echos Microbenchmarks The three I/O intensive benchmarks were compared, over time to assess the optimal value configuration. In particular, a single-socket system was compared with a dual-socket Intel system. Secondly, a system configured with dual-port 10GE was compared with a system fitted with 40Gigabit IO. The relative merits of the configurations in terms of I/O performance could then be balanced against other factors of CPU/memory bandwidth bottlenecks to arrive at the ideal system. For all benchmarks, we also show a dotted line indicating the throughput achieved by a single-threaded IOZONE sequential I/O benchmark on the systems. Note that with GRIDScaler, the single thread results are actually quite representative of maximum throughput as GRIDScaler delivers extremely strong single threaded I/O particularly for writes. In the results below for the write-benchmark, it is clear that the 10Gigabit node is achieving a good proportion of available bandwidth, with job counts above eight per node hitting the I/O ceiling. For 40Gigabit Ethernet the efficiency is still very strong, but of-course allows higher scaling. Only with 2 socket nodes, does the I/O performance really merit the additional cost of the 40G interface, where throughputs above 3GB/s are see to the application. Figure 1: ECHOs Write Benchmark showing the measured write performance when loading a single client with multiple copies of the I/O benchmark. The sequential read benchmark shows a similar overall pattern with the 2-socket node, clearly delivering benefits over a single socket node. The maximum read performance is achieved with just eight concurrent jobs per node and reaches very close to 4GB/s within 80% of the IOZONE measurements DataDirect Networks. All Rights Reserved. 7

8 Finally the Strided Read benchmarks were tested. In this case, the added dimension of a more onerous I/O task for the file system and storage subsystem has an impact on the efficiency of the application I/O for both 40G and 10G networked systems. In this case, the application benchmark still achieves over 2.5GB/s for the 40G connected client. The above throughput tests demonstrate strong scaling to around eight concurrent jobs before the I/O and network bottleneck significantly constrains the workload. In fact, for other reasons including CPU/memory bandwidth considerations, the chosen optimal value was three. Thus with three concurrent jobs per node we further investigated the number of nodes supported efficiently via a single GS12K appliance. It was found that a strong compromise involved 15 nodes each running three jobs per node. In this case, the measured throughput on the storage system indicated that the I/O benchmarks attained near the maximum throughput of ~20GB/s for Sequential Reads DataDirect Networks. All Rights Reserved. 8

9 Sequential Write Sequential Read Strided Read Measured throughput for 45 concurrent jobs across 15 nodes 13,155 MB/s 18,153 MB/s 6,756 MB/s One more preliminary test was performed to establish the impact of SSDs particularly on the Strided Read benchmark, where the I/O pattern is evidently more onerous on the storage system. To this end, we took advantage of the SFX read cache facility within the GS12K. A smaller test system was used and compared the performance of 100 spindles alone, with 100 spindles accelerated with 12 SSDs. Strided Reads No SFX 1639MB/s SFX Accelerated 5110 MB/s The smaller test system used 4 clients, with each running the 3 instances of the Strided Read benchmark. It is clear that the use of SFX read caching on the SSDs eliminates the disk contention that was the major bottleneck for the benchmark on spinning disk. This result is important as SFX allows a relatively small expenditure on a handful of SSDs to have a large impact on runtimes. Summary The requirement for increasing levels of I/O performance in large seismic processing environments is clear. The GS12K delivers very strong, scalable performance with a small footprint and enterprise features. Each GS12K appliance delivers 20GB/s to the clients and each single client can achieve over 4 GB/s. A good portion of the theoretical bandwidth is attainable by Paradigm Echos I/O intensive regions particularly alongside a strong 40GE interconnect. We show that synthetic benchmarks are a good indicator of Echos performance and can be used to assess the suitability of a good I/O subsystem in a seismic environment. In addition to providing a system that helps minimize the hardware environment, it was also of benefit to reduce the number of file systems. In such a large-scale production environment, there is an always a balance between the manageability of an extremely large file system, and the usability of multiple file systems. Today we provide file systems comprising of four GS12K appliances for production systems, each housing 400 drives. The economics at the procurement time drives the choice of disk capacity, but currently with 4TB drives, this would provide single file systems of over 5PBs. The key benefit of a true parallel file system over scale-out NAS is that very high single client performance can be delivered and sustained when many hundreds of clients are working concurrently with intensive seismic I/O. The single client I/O performance is particularly important because with increasing core counts per system, this I/O 2014 DataDirect Networks. All Rights Reserved. 9

10 quickly becomes the major bottleneck. If a client is limited to 10Gigabit Ethernet or the files system does not achieve near-wire speed access, the number of concurrent jobs that can be efficiently run per node decreases, hugely impacting the total cost of ownership of the system and increasing the time to value. This report concludes that the GS12K provides a highly efficient platform for the I/O-heavy aspects of Paradigm Echos, demonstrating an ability to scale to large file systems over 5PB with only four GS12K appliances. This can support real-world production workloads involving dense computation with high bandwidth networks. We also introduce SFX as a cost-effective way to gain benefits to application throughput without major investment on a large SSD estate. As evidenced with the testing and results presented in this paper, oil and gas companies can be confident in leveraging the Paradigm Echos and DDN solution to accelerate discovery and time to results for seismic processing in their own implementations. DDN About Us DataDirect Networks (DDN) is the world leader in massively scalable storage. Our data storage and processing solutions and professional services enable content-rich and high growth IT environments to achieve the highest levels of systems scalability, efficiency and simplicity. DDN enables enterprises to extract value and deliver business results from their information. Our customers include the world s leading online content and social networking providers, high performance cloud and grid computing, life sciences, media production, and security and intelligence organizations. Deployed in thousands of mission critical environments worldwide, DDN s solutions have been designed, engineered and proven in the world s most scalable data centers to ensure competitive business advantage for today s information powered enterprise. For more information, go to www. or call DataDirect Networks, Inc. All Rights Reserved. DDN, GS12K, SFA7700, SFA, SFX, Storage Fusion Architecture, Storage Fusion Xcelerator, is a trademark of DataDirect Networks. Other Names and Brands May Be Claimed as the Property of Others. Version-1 10/ DataDirect Networks. All Rights Reserved. 10

HadoopTM Analytics DDN

HadoopTM Analytics DDN DDN Solution Brief Accelerate> HadoopTM Analytics with the SFA Big Data Platform Organizations that need to extract value from all data can leverage the award winning SFA platform to really accelerate

More information

With DDN Big Data Storage

With DDN Big Data Storage DDN Solution Brief Accelerate > ISR With DDN Big Data Storage The Way to Capture and Analyze the Growing Amount of Data Created by New Technologies 2012 DataDirect Networks. All Rights Reserved. The Big

More information

ANY SURVEILLANCE, ANYWHERE, ANYTIME

ANY SURVEILLANCE, ANYWHERE, ANYTIME ANY SURVEILLANCE, ANYWHERE, ANYTIME WHITEPAPER DDN Storage Powers Next Generation Video Surveillance Infrastructure INTRODUCTION Over the past decade, the world has seen tremendous growth in the use of

More information

Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com

Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com DDN Technical Brief Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. A Fundamentally Different Approach To Enterprise Analytics Architecture: A Scalable Unit

More information

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution Arista 10 Gigabit Ethernet Switch Lab-Tested with Panasas ActiveStor Parallel Storage System Delivers Best Results for High-Performance and Low Latency for Scale-Out Cloud Storage Applications Introduction

More information

Integrated Grid Solutions. and Greenplum

Integrated Grid Solutions. and Greenplum EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving

More information

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage

More information

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5

More information

MAKING THE BUSINESS CASE

MAKING THE BUSINESS CASE MAKING THE BUSINESS CASE LUSTRE FILE SYSTEMS ARE POISED TO PENETRATE COMMERCIAL MARKETS table of contents + Considerations in Building the.... 1... 3.... 4 A TechTarget White Paper by Long the de facto

More information

WOS Cloud. ddn.com. Personal Storage for the Enterprise. DDN Solution Brief

WOS Cloud. ddn.com. Personal Storage for the Enterprise. DDN Solution Brief DDN Solution Brief Personal Storage for the Enterprise WOS Cloud Secure, Shared Drop-in File Access for Enterprise Users, Anytime and Anywhere 2011 DataDirect Networks. All Rights Reserved DDN WOS Cloud

More information

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010 Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,

More information

ECMWF HPC Workshop: Accelerating Data Management

ECMWF HPC Workshop: Accelerating Data Management October 2012 ECMWF HPC Workshop: Accelerating Data Management Massively-Scalable Platforms and Solutions Engineered for the Big Data and Cloud Era Glenn Wright Systems Architect, DDN Data-Driven Paradigm

More information

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products MaxDeploy Ready Hyper- Converged Virtualization Solution With SanDisk Fusion iomemory products MaxDeploy Ready products are configured and tested for support with Maxta software- defined storage and with

More information

Scala Storage Scale-Out Clustered Storage White Paper

Scala Storage Scale-Out Clustered Storage White Paper White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current

More information

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a

More information

Accelerate > Converged Storage Infrastructure. DDN Case Study. ddn.com. 2013 DataDirect Networks. All Rights Reserved

Accelerate > Converged Storage Infrastructure. DDN Case Study. ddn.com. 2013 DataDirect Networks. All Rights Reserved DDN Case Study Accelerate > Converged Storage Infrastructure 2013 DataDirect Networks. All Rights Reserved The University of Florida s (ICBR) offers access to cutting-edge technologies designed to enable

More information

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory Customer Success Story Los Alamos National Laboratory Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory June 2010 Highlights First Petaflop Supercomputer

More information

Object storage in Cloud Computing and Embedded Processing

Object storage in Cloud Computing and Embedded Processing Object storage in Cloud Computing and Embedded Processing Jan Jitze Krol Systems Engineer DDN We Accelerate Information Insight DDN is a Leader in Massively Scalable Platforms and Solutions for Big Data

More information

Performance in a Gluster System. Versions 3.1.x

Performance in a Gluster System. Versions 3.1.x Performance in a Gluster System Versions 3.1.x TABLE OF CONTENTS Table of Contents... 2 List of Figures... 3 1.0 Introduction to Gluster... 4 2.0 Gluster view of Performance... 5 2.1 Good performance across

More information

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer Stan Posey, MSc and Bill Loewe, PhD Panasas Inc., Fremont, CA, USA Paul Calleja, PhD University of Cambridge,

More information

Benchmarking Cassandra on Violin

Benchmarking Cassandra on Violin Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract

More information

ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter

ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter WHITEPAPER ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter THE BIG DATA CHALLENGE AND OPPORTUNITY The.proliferation,.management.and.analysis.of.intelligence.data.is.a.fast.growing.concern.

More information

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications

More information

Isilon IQ Scale-out NAS for High-Performance Applications

Isilon IQ Scale-out NAS for High-Performance Applications Isilon IQ Scale-out NAS for High-Performance Applications Optimizing Performance with Isilon IQ Storage By Shai Harmelin, Sr. Solutions Architect An Isilon Systems Technical Whitepaper July 2009 ISILON

More information

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything BlueArc unified network storage systems 7th TF-Storage Meeting Scale Bigger, Store Smarter, Accelerate Everything BlueArc s Heritage Private Company, founded in 1998 Headquarters in San Jose, CA Highest

More information

EMC XTREMIO EXECUTIVE OVERVIEW

EMC XTREMIO EXECUTIVE OVERVIEW EMC XTREMIO EXECUTIVE OVERVIEW COMPANY BACKGROUND XtremIO develops enterprise data storage systems based completely on random access media such as flash solid-state drives (SSDs). By leveraging the underlying

More information

Netapp @ 10th TF-Storage Meeting

Netapp @ 10th TF-Storage Meeting Netapp @ 10th TF-Storage Meeting Wojciech Janusz, Netapp Poland Bogusz Błaszkiewicz, Netapp Poland Ljubljana, 2012.02.20 Agenda Data Ontap Cluster-Mode pnfs E-Series NetApp Confidential - Internal Use

More information

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director brett.weninger@adurant.com Dave Smelker, Managing Principal dave.smelker@adurant.com

More information

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one

More information

EXPLORATION TECHNOLOGY REQUIRES A RADICAL CHANGE IN DATA ANALYSIS

EXPLORATION TECHNOLOGY REQUIRES A RADICAL CHANGE IN DATA ANALYSIS EXPLORATION TECHNOLOGY REQUIRES A RADICAL CHANGE IN DATA ANALYSIS EMC Isilon solutions for oil and gas EMC PERSPECTIVE TABLE OF CONTENTS INTRODUCTION: THE HUNT FOR MORE RESOURCES... 3 KEEPING PACE WITH

More information

www.xenon.com.au STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

www.xenon.com.au STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING www.xenon.com.au STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING GPU COMPUTING VISUALISATION XENON Accelerating Exploration Mineral, oil and gas exploration is an expensive and challenging

More information

www.thinkparq.com www.beegfs.com

www.thinkparq.com www.beegfs.com www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a

More information

Zadara Storage Cloud A whitepaper. @ZadaraStorage

Zadara Storage Cloud A whitepaper. @ZadaraStorage Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers

More information

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator

More information

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance

More information

Software-defined Storage Architecture for Analytics Computing

Software-defined Storage Architecture for Analytics Computing Software-defined Storage Architecture for Analytics Computing Arati Joshi Performance Engineering Colin Eldridge File System Engineering Carlos Carrero Product Management June 2015 Reference Architecture

More information

How To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory)

How To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory) WHITE PAPER Oracle NoSQL Database and SanDisk Offer Cost-Effective Extreme Performance for Big Data 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Abstract... 3 What Is Big Data?...

More information

Application Performance for High Performance Computing Environments

Application Performance for High Performance Computing Environments Application Performance for High Performance Computing Environments Leveraging the strengths of Computationally intensive applications With high performance scale out file serving In data storage modules

More information

Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era

Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era Enterprise Strategy Group Getting to the bigger truth. White Paper Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era Apeiron introduces NVMe-based storage innovation designed

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

Hadoop on the Gordon Data Intensive Cluster

Hadoop on the Gordon Data Intensive Cluster Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,

More information

Software-defined Storage at the Speed of Flash

Software-defined Storage at the Speed of Flash TECHNICAL BRIEF: SOFTWARE-DEFINED STORAGE AT THE SPEED OF... FLASH..................................... Intel SSD Data Center P3700 Series and Symantec Storage Foundation with Flexible Storage Sharing

More information

Microsoft SQL Server 2014 Fast Track

Microsoft SQL Server 2014 Fast Track Microsoft SQL Server 2014 Fast Track 34-TB Certified Data Warehouse 103-TB Maximum User Data Tegile Systems Solution Review 2U Design: Featuring Tegile T3800 All-Flash Storage Array http:// www.tegile.com/solutiuons/sql

More information

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor

More information

Geospatial Imaging Cloud Storage Capturing the World at Scale with WOS TM. ddn.com. DDN Whitepaper. 2011 DataDirect Networks. All Rights Reserved.

Geospatial Imaging Cloud Storage Capturing the World at Scale with WOS TM. ddn.com. DDN Whitepaper. 2011 DataDirect Networks. All Rights Reserved. DDN Whitepaper Geospatial Imaging Cloud Storage Capturing the World at Scale with WOS TM Table of Contents Growth and Complexity Challenges for Geospatial Imaging 3 New Solutions to Drive Insight, Simplicity

More information

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services ALPS Supercomputing System A Scalable Supercomputer with Flexible Services 1 Abstract Supercomputing is moving from the realm of abstract to mainstream with more and more applications and research being

More information

Maginatics Cloud Storage Platform for Elastic NAS Workloads

Maginatics Cloud Storage Platform for Elastic NAS Workloads Maginatics Cloud Storage Platform for Elastic NAS Workloads Optimized for Cloud Maginatics Cloud Storage Platform () is the first solution optimized for the cloud. It provides lower cost, easier administration,

More information

EMC SOLUTION FOR SPLUNK

EMC SOLUTION FOR SPLUNK EMC SOLUTION FOR SPLUNK Splunk validation using all-flash EMC XtremIO and EMC Isilon scale-out NAS ABSTRACT This white paper provides details on the validation of functionality and performance of Splunk

More information

Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS

Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS M ost storage vendors now offer all-flash storage arrays, and many modern organizations recognize the need for these highperformance

More information

Milestone Solution Partner IT Infrastructure Components Certification Summary

Milestone Solution Partner IT Infrastructure Components Certification Summary Milestone Solution Partner IT Infrastructure Components Certification Summary Dell FS8600 NAS Storage 12-1-2014 Table of Contents Introduction:... 2 Dell Storage Architecture:... 3 Certified Products:...

More information

IEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV

IEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV IEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV 11 Manager May 04, 2010 Joe Rotiroti Client Systems IBM, Federal 484 433 9756 cell 845 491 5227 fax rotiroti@us.ibm.com How is GPFS different?

More information

Understanding Enterprise NAS

Understanding Enterprise NAS Anjan Dave, Principal Storage Engineer LSI Corporation Author: Anjan Dave, Principal Storage Engineer, LSI Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA

More information

HPC Advisory Council

HPC Advisory Council HPC Advisory Council September 2012, Malaga CHRIS WEEDEN SYSTEMS ENGINEER WHO IS PANASAS? Panasas is a high performance storage vendor founded by Dr Garth Gibson Panasas delivers a fully supported, turnkey,

More information

Deploying Affordable, High Performance Hybrid Flash Storage for Clustered SQL Server

Deploying Affordable, High Performance Hybrid Flash Storage for Clustered SQL Server Deploying Affordable, High Performance Hybrid Flash Storage for Clustered SQL Server Flash storage adoption has increased in recent years, as organizations have deployed it to support business applications.

More information

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads 89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com @EdisonGroupInc 212.367.7400 IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads A Competitive Test and Evaluation Report

More information

Panasas: High Performance Storage for the Engineering Workflow

Panasas: High Performance Storage for the Engineering Workflow 9. LS-DYNA Forum, Bamberg 2010 IT / Performance Panasas: High Performance Storage for the Engineering Workflow E. Jassaud, W. Szoecs Panasas / transtec AG 2010 Copyright by DYNAmore GmbH N - I - 9 High-Performance

More information

Platfora Big Data Analytics

Platfora Big Data Analytics Platfora Big Data Analytics ISV Partner Solution Case Study and Cisco Unified Computing System Platfora, the leading enterprise big data analytics platform built natively on Hadoop and Spark, delivers

More information

Optimizing Large Arrays with StoneFly Storage Concentrators

Optimizing Large Arrays with StoneFly Storage Concentrators Optimizing Large Arrays with StoneFly Storage Concentrators All trademark names are the property of their respective companies. This publication contains opinions of which are subject to change from time

More information

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview Pivot3 Desktop Virtualization Appliances vstac VDI Technology Overview February 2012 Pivot3 Desktop Virtualization Technology Overview Table of Contents Executive Summary... 3 The Pivot3 VDI Appliance...

More information

Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk

Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk WHITE PAPER Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk 951 SanDisk Drive, Milpitas, CA 95035 2015 SanDisk Corporation. All rights reserved. www.sandisk.com Table of Contents Introduction

More information

ioscale: The Holy Grail for Hyperscale

ioscale: The Holy Grail for Hyperscale ioscale: The Holy Grail for Hyperscale The New World of Hyperscale Hyperscale describes new cloud computing deployments where hundreds or thousands of distributed servers support millions of remote, often

More information

POWER ALL GLOBAL FILE SYSTEM (PGFS)

POWER ALL GLOBAL FILE SYSTEM (PGFS) POWER ALL GLOBAL FILE SYSTEM (PGFS) Defining next generation of global storage grid Power All Networks Ltd. Technical Whitepaper April 2008, version 1.01 Table of Content 1. Introduction.. 3 2. Paradigm

More information

How To Get The Most Out Of A Large Data Set

How To Get The Most Out Of A Large Data Set DDN Solution Brief Overcoming > The Big Data Technology Hurdle Turning Data into Answers with DDN & Vertica 20 Networks. All Rights Reserved. Executive Summary Networks and Vertica have collaborated to

More information

Collaborative Research Infrastructure Deployments. ddn.com. Accelerate > DDN Case Study

Collaborative Research Infrastructure Deployments. ddn.com. Accelerate > DDN Case Study DDN Case Study Accelerate > Collaborative Research Infrastructure Deployments University College London Transforms Research Collaboration and Data Preservation with Scalable Cloud Object Storage Appliance

More information

(Scale Out NAS System)

(Scale Out NAS System) For Unlimited Capacity & Performance Clustered NAS System (Scale Out NAS System) Copyright 2010 by Netclips, Ltd. All rights reserved -0- 1 2 3 4 5 NAS Storage Trend Scale-Out NAS Solution Scaleway Advantages

More information

Netapp HPC Solution for Lustre. Rich Fenton (fenton@netapp.com) UK Solutions Architect

Netapp HPC Solution for Lustre. Rich Fenton (fenton@netapp.com) UK Solutions Architect Netapp HPC Solution for Lustre Rich Fenton (fenton@netapp.com) UK Solutions Architect Agenda NetApp Introduction Introducing the E-Series Platform Why E-Series for Lustre? Modular Scale-out Capacity Density

More information

NetApp High-Performance Computing Solution for Lustre: Solution Guide

NetApp High-Performance Computing Solution for Lustre: Solution Guide Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5

More information

High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology

High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology Evaluation report prepared under contract with Brocade Executive Summary As CIOs

More information

LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment

LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment Evaluation report prepared under contract with LSI Corporation Introduction Interest in solid-state storage (SSS) is high, and

More information

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server

More information

High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software

High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software White Paper Overview The Micron M500DC SSD was designed after months of close work with major data center service providers and

More information

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router HyperQ Hybrid Flash Storage Made Easy White Paper Parsec Labs, LLC. 7101 Northland Circle North, Suite 105 Brooklyn Park, MN 55428 USA 1-763-219-8811 www.parseclabs.com info@parseclabs.com sales@parseclabs.com

More information

The Ultimate in Scale-Out Storage for HPC and Big Data

The Ultimate in Scale-Out Storage for HPC and Big Data Node Inventory Health and Active Filesystem Throughput Monitoring Asset Utilization and Capacity Statistics Manager brings to life powerful, intuitive, context-aware real-time monitoring and proactive

More information

Everything you need to know about flash storage performance

Everything you need to know about flash storage performance Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices

More information

THE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved.

THE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved. THE EMC ISILON STORY Big Data In The Enterprise 2012 1 Big Data In The Enterprise Isilon Overview Isilon Technology Summary 2 What is Big Data? 3 The Big Data Challenge File Shares 90 and Archives 80 Bioinformatics

More information

WOS. High Performance Object Storage

WOS. High Performance Object Storage Datasheet WOS High Performance Object Storage The Big Data explosion brings both challenges and opportunities to businesses across all industry verticals. Providers of online services are building infrastructures

More information

Flash Memory Technology in Enterprise Storage

Flash Memory Technology in Enterprise Storage NETAPP WHITE PAPER Flash Memory Technology in Enterprise Storage Flexible Choices to Optimize Performance Mark Woods and Amit Shah, NetApp November 2008 WP-7061-1008 EXECUTIVE SUMMARY Solid state drives

More information

SMB Direct for SQL Server and Private Cloud

SMB Direct for SQL Server and Private Cloud SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server

More information

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine An Oracle White Paper May 2011 Exadata Smart Flash Cache and the Oracle Exadata Database Machine Exadata Smart Flash Cache... 2 Oracle Database 11g: The First Flash Optimized Database... 2 Exadata Smart

More information

Data management challenges in todays Healthcare and Life Sciences ecosystems

Data management challenges in todays Healthcare and Life Sciences ecosystems Data management challenges in todays Healthcare and Life Sciences ecosystems Jose L. Alvarez Principal Engineer, WW Director Life Sciences jose.alvarez@seagate.com Evolution of Data Sets in Healthcare

More information

A Survey of Shared File Systems

A Survey of Shared File Systems Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...

More information

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers Masood Ahmed EMEA Infrastructure Solutions Oracle/SAP Relationship Overview First SAP R/3 release

More information

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007

More information

Solving Agencies Big Data Challenges: PED for On-the-Fly Decisions

Solving Agencies Big Data Challenges: PED for On-the-Fly Decisions White Paper Solving Agencies Big Data Challenges: PED for On-the-Fly Decisions Carina Veksler, NetApp March 2012 WP-7158 ABSTRACT With the growing volumes of rich sensor data and imagery used today to

More information

Data Center Solutions

Data Center Solutions Data Center Solutions Systems, software and hardware solutions you can trust With over 25 years of storage innovation, SanDisk is a global flash technology leader. At SanDisk, we re expanding the possibilities

More information

Data Center Storage Solutions

Data Center Storage Solutions Data Center Storage Solutions Enterprise software, appliance and hardware solutions you can trust When it comes to storage, most enterprises seek the same things: predictable performance, trusted reliability

More information

Easier - Faster - Better

Easier - Faster - Better Highest reliability, availability and serviceability ClusterStor gets you productive fast with robust professional service offerings available as part of solution delivery, including quality controlled

More information

Mellanox Accelerated Storage Solutions

Mellanox Accelerated Storage Solutions Mellanox Accelerated Storage Solutions Moving Data Efficiently In an era of exponential data growth, storage infrastructures are being pushed to the limits of their capacity and data delivery capabilities.

More information

NEXSAN NST STORAGE FOR THE VIRTUAL DESKTOP

NEXSAN NST STORAGE FOR THE VIRTUAL DESKTOP NST STORAGE FOR THE VIRTUAL DESKTOP Nexsan s innovative product, the NST5000, is a hybrid storage system with unified protocols and highly dense storage for a combination of great performance, low cost,

More information

Block based, file-based, combination. Component based, solution based

Block based, file-based, combination. Component based, solution based The Wide Spread Role of 10-Gigabit Ethernet in Storage This paper provides an overview of SAN and NAS storage solutions, highlights the ubiquitous role of 10 Gigabit Ethernet in these solutions, and illustrates

More information

Data Center Performance Insurance

Data Center Performance Insurance Data Center Performance Insurance How NFS Caching Guarantees Rapid Response Times During Peak Workloads November 2010 2 Saving Millions By Making It Easier And Faster Every year slow data centers and application

More information

How to Choose your Red Hat Enterprise Linux Filesystem

How to Choose your Red Hat Enterprise Linux Filesystem How to Choose your Red Hat Enterprise Linux Filesystem EXECUTIVE SUMMARY Choosing the Red Hat Enterprise Linux filesystem that is appropriate for your application is often a non-trivial decision due to

More information

Introduction to Gluster. Versions 3.0.x

Introduction to Gluster. Versions 3.0.x Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster

More information

Big data management with IBM General Parallel File System

Big data management with IBM General Parallel File System Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers

More information

Colgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP

Colgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP selects SAP HANA to improve the speed of business analytics with IBM and SAP Founded in 1806, is a global consumer products company which sells nearly $17 billion annually in personal care, home care,

More information

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup

More information

Maximum performance, minimal risk for data warehousing

Maximum performance, minimal risk for data warehousing SYSTEM X SERVERS SOLUTION BRIEF Maximum performance, minimal risk for data warehousing Microsoft Data Warehouse Fast Track for SQL Server 2014 on System x3850 X6 (95TB) The rapid growth of technology has

More information

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system Christian Clémençon (EPFL-DIT)  4 April 2013 GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID

More information