Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers
|
|
- Vernon Ford
- 8 years ago
- Views:
Transcription
1 DDN Whitepaper Improving Time to Results for Seismic Processing with Paradigm and DDN James Coomer and Laurent Thiers 2014 DataDirect Networks. All Rights Reserved.
2 Executive Summary Companies in the oil and gas industry continue to push the limits of technology to gain a competitive advantage in their quest for the richest and most profitable reservoirs. For fastest time to results, Paradigm Echos and DataDirect Networks provide cost-effective processing and storage solutions that deliver the highest speed, precision, capacity, and scalability. This whitepaper summarizes Paradigm Echos workload testing performed on DDN storage, taking advantage of SSDs and high performance networks to improve the time to value for a seismic processing problem at lower cost. Targeted workload testing enables analysis of the bottlenecks in the interactions between the software, the network fabric and the storage systems; and removing these bottlenecks leads to improved efficiency. For example, if the number of jobs per compute node can be increased, we can either improve overall performance for a given number of systems or produce the same overall performance with fewer systems. The Need for I/O Improvements in Seismic Processing Environments Companies and partners engaged in oil and gas exploration and production recognize the value of speed and precision, investing tens of billions of dollars in new projects. With so much money at stake, there's little room for error when collecting the most precise and accurate information on where to drill, how much to bid on a site, and how to maximize reservoir performance. The third critical piece of the process requires companies to apply the most sophisticated modeling and simulation technology for the transformation of field data into valuable geophysics. While technology advances within the industry are adding much improved precision, they are also significantly increasing data and processing volumes. Wide/Multi/Rich-azimuth methods are using multi-sensor arrays and sophisticated methods to produce higher-fidelity images. The deployment of these richer formats results in a vast increase for the amount of data the company must process and manage. Furthermore, new analytics techniques allow for continued advancement in the interpretation of seismic data for new data coming in and for historical oil field data as well. This means more data needs to be analyzed at greater speeds. In the field of seismic data acquisition and processing, the conversion of raw field recordings to geological images is a highly intensive process in both data and compute. The requirement for processing more data with less resources (power and capital cost) tends to push the compute threads ever more densely onto the available compute clients, and this in-turn impacts on the I/O subsystems both network and storage. As core counts and memory bandwidths steadily rise, the challenge of shifting data to feed the compute nodes becomes increasingly important. Paradigm Echos: The Standard for Seismic Processing and Data Analysis For over 25 years, Paradigm Echos has been the oil and gas industry's leading system for seismic processing. Echos is a software suite that links seismic data and processing parameters and sequences, with both interactive and 2014 DataDirect Networks. All Rights Reserved. 2
3 batch components. The suite of seismic processing modules are provided to transform converted wave recordings to interpretable images of the subsurface, which are generally run in an embarrassingly parallel-like fashion with many simultaneous instances running on a large seismic cluster. Widely used, companies prefer Echos because of its breadth of geophysical applications, unique combination of production and interactive seismic data processing, maturity and stability, and versatile programming environment for client-guided customization. DDN: Driving Big Data Storage Solutions for Seismic Processing High-performance storage solutions from DDN meet the most pressing storage challenges of the seismic processing industry, helping oil and gas companies achieve both their scientific goals and their long-term business goals. To support the most demanding environments, DDN solutions deliver game-changing performance, capacity and scalability for ingest, processing and visualization. For example, DDN's Storage Fusion Architecture (SFA) combines SATA, SAS and solid-state disks into a simply managed, massively scalable, multi-petabyte platform that can be tailored to a balance of throughput and capacity. The combination of modularity and performance feature allows users in seismic processing environments to configure systems that scale up and/or scale out in independent dimensions. This is why over half the top producers worldwide are leveraging DDN storage solutions in their big data oil and gas workflows. PARADIGM and DDN Seismic data sets can be very large, with the raw recording data ranging into the 10s or 100s of Terabytes. The ability to efficiently view a particular seismic data set in different orders or domains, without creating multiple copies of that data set is a key design feature of the seismic file handling capabilities of the Echos system. Initial loading of raw seismic recording data is usually done in the order of which the data is recorded, based on sequential shot numbering and a secondary order associated with the receivers associated with each shot. However, this primary and secondary key ordering, shot-receiver ordered data may not be the most effective order for certain subsurface analysis techniques. The Common Depth Point (CDP) Method is a particular seismic data acquisition and processing technique that transforms field recordings from seismic surveys into pseudo-cross-sectional images of the earth's geologic layering beneath the surface locations. Such images are key to geophysical analysis used to pinpoint likely drilling locations for oil and gas reservoirs. Three micro-workloads are used in this study, which are components of the Paradigm Echos. All three involve heavy I/O, but in differing ways. Reading and writing a shot-receiver ordered data set in sequential order (SEQIN) efficiently is a critical I/O workload for conditioning of seismic data for further analysis. Another important I/O workload is reading a very large shotreceiver ordered data set in CDP order (STRIDE) which can result in file system access pattern that are challenging for 2014 DataDirect Networks. All Rights Reserved. 3
4 general-purpose storage systems. An un-tuned system can limit the read performance to just a few MB/sec. The third micro-workload (SEQOUT) performs sequential writes rather than reads. The seismic file handling codes underlying the Echos software are multi threaded, with different tasks for reading, processing, timing etc. In the Echos processing system, a logical seismic file consists of multiple actual files on the file systems that enable specialized handling of different logical elements that make up the seismic file. The SFA Advantage Currently supporting 70% of the Top10 supercomputers in the world, DDN s SFA technology combines 15 years of hardware optimization in High Performance Computing with sophisticated I/O software (SFAOS). The result is a highly efficient system that delivers the highest throughput, density and IOPs of any system on the market. The SFA product line is divided into the high-end SFA12KX and the mid-range SFA7700. Both run SFAOS, which leverages Storage Fusion Xcelerator (SFX) for hybrid deployments of traditional SAS drives with SSDs. SFX integrates application-centric intelligence with flash media to accelerate awkward read intensive workloads. The GS12K Appliance is the highest performing file system appliance in the world, with a single appliance delivering over 20GB/s to the clients. The system uses the SFAOS embedded technology to run complete parallel filesystem services within the controller. This approach removes the need for external file servers typically required for parallel file systems, allowing for far better single client performance than can be reached with scale-out NAS offerings. GS12K supports very large file systems with extreme density. A single appliance can start with just 60 drives, but be scaled up to 10PB in just two racks. If a larger file system is required, further appliances can be added into the single namespace. The GS12K also supports advanced Enterprise features such as windows support, snapshots, replication, HSM, and NFS/CIFS exports DataDirect Networks. All Rights Reserved. 4
5 GS12K Features Client Access and connectivity Native Linux and Windows Clients Highly Available NFS v3 Native Linux and Windows Clients Client Connectivity 16 x 10/40GbE Host Port or 16 x 56Gb In-finiBand Host Port Support for InfiniBand RDMA I/O Client to Gateway Load Balancing Many to One, and One to Many Client to Gateway Access Volume Scalability Up to 3.45PBs of usable capacity (840 drives) Optional Metadata accelerator drives Up to 3.45PBs of usable capacity (840 drives) Over 10PB+ of Aggregated Capacity Performance 20GB/s per GS12K System (read & write) Aggregated Performance up to 200GB/sec 100,000s of File Operations/sec Data Protection Data Protection Data: RAID 6 (8+2) Metadata: RAID Snapshots per file system DirectProtect Data Integrity Synchronous File System Replication Optimization of SFA with Paradigm DDN s history in deploying systems for seismic processing has included significant benchmarking and tuning with the largest oil and gas companies in the world. This report discusses the results of some particular major work with one such company where a large GS12K environment is deployed with 40Gigabit Ethernet. We present three sets of results, each of which are of significance to O&G seismic processing systems: 1. GS12KE baseline. Maximum data throughput of the GS12K in 40GE environment as measured by synthetic benchmarks 2. The potential for 40GE vs 10GE in heavy I/O seismic environments 3. Efficiency of Paradigm Echos I/O in large scale systems Data and Systems Centralized con figuration and Management The benchmark system comprised a single monitoring solution Online File System Growth GS12K system connected via 8 40GE ports to a Object-Based File Locking Monitoring and Event Notification System 40GE backbone. Each GS12K uses 5 SS8460 Disk Enclosures with between 300 and 400 NL-SAS drives for data. Data protection was provided via RAID 6. GRIDScaler metadata was held on a small dedicated set of drives separate from the data drives. The clients used varied in the test, but all were a mid-range clock speed Intel system connected, with either 10Gigabit Ethernet or 40Gigabit Ethernet. SAS 6Gb x DataDirect Networks. All Rights Reserved. 5
6 GS12KE Baseline The IOR benchmark ( is a popular parallel I/O benchmark used to characterise storage systems and file systems. IOR allows the benchmark engineer to run multiple threads, reading or writing from many clients simultaneously. A number of parameters are tuneable to avoid caching, alter the size of the data transfers issued from the IOR tasks and change the nature of the I/O topology. In the following case, we present results from executing IOR across 16 clients simultaneously. Each client is connected to the network with 40GE and runs between 8 and 32 threads per client, each thread read or writing with varying transfer sizes to a single shared file DataDirect Networks. All Rights Reserved. 6
7 Paradigm Echos Microbenchmarks The three I/O intensive benchmarks were compared, over time to assess the optimal value configuration. In particular, a single-socket system was compared with a dual-socket Intel system. Secondly, a system configured with dual-port 10GE was compared with a system fitted with 40Gigabit IO. The relative merits of the configurations in terms of I/O performance could then be balanced against other factors of CPU/memory bandwidth bottlenecks to arrive at the ideal system. For all benchmarks, we also show a dotted line indicating the throughput achieved by a single-threaded IOZONE sequential I/O benchmark on the systems. Note that with GRIDScaler, the single thread results are actually quite representative of maximum throughput as GRIDScaler delivers extremely strong single threaded I/O particularly for writes. In the results below for the write-benchmark, it is clear that the 10Gigabit node is achieving a good proportion of available bandwidth, with job counts above eight per node hitting the I/O ceiling. For 40Gigabit Ethernet the efficiency is still very strong, but of-course allows higher scaling. Only with 2 socket nodes, does the I/O performance really merit the additional cost of the 40G interface, where throughputs above 3GB/s are see to the application. Figure 1: ECHOs Write Benchmark showing the measured write performance when loading a single client with multiple copies of the I/O benchmark. The sequential read benchmark shows a similar overall pattern with the 2-socket node, clearly delivering benefits over a single socket node. The maximum read performance is achieved with just eight concurrent jobs per node and reaches very close to 4GB/s within 80% of the IOZONE measurements DataDirect Networks. All Rights Reserved. 7
8 Finally the Strided Read benchmarks were tested. In this case, the added dimension of a more onerous I/O task for the file system and storage subsystem has an impact on the efficiency of the application I/O for both 40G and 10G networked systems. In this case, the application benchmark still achieves over 2.5GB/s for the 40G connected client. The above throughput tests demonstrate strong scaling to around eight concurrent jobs before the I/O and network bottleneck significantly constrains the workload. In fact, for other reasons including CPU/memory bandwidth considerations, the chosen optimal value was three. Thus with three concurrent jobs per node we further investigated the number of nodes supported efficiently via a single GS12K appliance. It was found that a strong compromise involved 15 nodes each running three jobs per node. In this case, the measured throughput on the storage system indicated that the I/O benchmarks attained near the maximum throughput of ~20GB/s for Sequential Reads DataDirect Networks. All Rights Reserved. 8
9 Sequential Write Sequential Read Strided Read Measured throughput for 45 concurrent jobs across 15 nodes 13,155 MB/s 18,153 MB/s 6,756 MB/s One more preliminary test was performed to establish the impact of SSDs particularly on the Strided Read benchmark, where the I/O pattern is evidently more onerous on the storage system. To this end, we took advantage of the SFX read cache facility within the GS12K. A smaller test system was used and compared the performance of 100 spindles alone, with 100 spindles accelerated with 12 SSDs. Strided Reads No SFX 1639MB/s SFX Accelerated 5110 MB/s The smaller test system used 4 clients, with each running the 3 instances of the Strided Read benchmark. It is clear that the use of SFX read caching on the SSDs eliminates the disk contention that was the major bottleneck for the benchmark on spinning disk. This result is important as SFX allows a relatively small expenditure on a handful of SSDs to have a large impact on runtimes. Summary The requirement for increasing levels of I/O performance in large seismic processing environments is clear. The GS12K delivers very strong, scalable performance with a small footprint and enterprise features. Each GS12K appliance delivers 20GB/s to the clients and each single client can achieve over 4 GB/s. A good portion of the theoretical bandwidth is attainable by Paradigm Echos I/O intensive regions particularly alongside a strong 40GE interconnect. We show that synthetic benchmarks are a good indicator of Echos performance and can be used to assess the suitability of a good I/O subsystem in a seismic environment. In addition to providing a system that helps minimize the hardware environment, it was also of benefit to reduce the number of file systems. In such a large-scale production environment, there is an always a balance between the manageability of an extremely large file system, and the usability of multiple file systems. Today we provide file systems comprising of four GS12K appliances for production systems, each housing 400 drives. The economics at the procurement time drives the choice of disk capacity, but currently with 4TB drives, this would provide single file systems of over 5PBs. The key benefit of a true parallel file system over scale-out NAS is that very high single client performance can be delivered and sustained when many hundreds of clients are working concurrently with intensive seismic I/O. The single client I/O performance is particularly important because with increasing core counts per system, this I/O 2014 DataDirect Networks. All Rights Reserved. 9
10 quickly becomes the major bottleneck. If a client is limited to 10Gigabit Ethernet or the files system does not achieve near-wire speed access, the number of concurrent jobs that can be efficiently run per node decreases, hugely impacting the total cost of ownership of the system and increasing the time to value. This report concludes that the GS12K provides a highly efficient platform for the I/O-heavy aspects of Paradigm Echos, demonstrating an ability to scale to large file systems over 5PB with only four GS12K appliances. This can support real-world production workloads involving dense computation with high bandwidth networks. We also introduce SFX as a cost-effective way to gain benefits to application throughput without major investment on a large SSD estate. As evidenced with the testing and results presented in this paper, oil and gas companies can be confident in leveraging the Paradigm Echos and DDN solution to accelerate discovery and time to results for seismic processing in their own implementations. DDN About Us DataDirect Networks (DDN) is the world leader in massively scalable storage. Our data storage and processing solutions and professional services enable content-rich and high growth IT environments to achieve the highest levels of systems scalability, efficiency and simplicity. DDN enables enterprises to extract value and deliver business results from their information. Our customers include the world s leading online content and social networking providers, high performance cloud and grid computing, life sciences, media production, and security and intelligence organizations. Deployed in thousands of mission critical environments worldwide, DDN s solutions have been designed, engineered and proven in the world s most scalable data centers to ensure competitive business advantage for today s information powered enterprise. For more information, go to www. or call DataDirect Networks, Inc. All Rights Reserved. DDN, GS12K, SFA7700, SFA, SFX, Storage Fusion Architecture, Storage Fusion Xcelerator, is a trademark of DataDirect Networks. Other Names and Brands May Be Claimed as the Property of Others. Version-1 10/ DataDirect Networks. All Rights Reserved. 10
HadoopTM Analytics DDN
DDN Solution Brief Accelerate> HadoopTM Analytics with the SFA Big Data Platform Organizations that need to extract value from all data can leverage the award winning SFA platform to really accelerate
More informationWith DDN Big Data Storage
DDN Solution Brief Accelerate > ISR With DDN Big Data Storage The Way to Capture and Analyze the Growing Amount of Data Created by New Technologies 2012 DataDirect Networks. All Rights Reserved. The Big
More informationANY SURVEILLANCE, ANYWHERE, ANYTIME
ANY SURVEILLANCE, ANYWHERE, ANYTIME WHITEPAPER DDN Storage Powers Next Generation Video Surveillance Infrastructure INTRODUCTION Over the past decade, the world has seen tremendous growth in the use of
More informationModernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com
DDN Technical Brief Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. A Fundamentally Different Approach To Enterprise Analytics Architecture: A Scalable Unit
More informationIntroduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution
Arista 10 Gigabit Ethernet Switch Lab-Tested with Panasas ActiveStor Parallel Storage System Delivers Best Results for High-Performance and Low Latency for Scale-Out Cloud Storage Applications Introduction
More informationIntegrated Grid Solutions. and Greenplum
EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving
More informationScaling Objectivity Database Performance with Panasas Scale-Out NAS Storage
White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage
More informationDriving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
More informationMAKING THE BUSINESS CASE
MAKING THE BUSINESS CASE LUSTRE FILE SYSTEMS ARE POISED TO PENETRATE COMMERCIAL MARKETS table of contents + Considerations in Building the.... 1... 3.... 4 A TechTarget White Paper by Long the de facto
More informationWOS Cloud. ddn.com. Personal Storage for the Enterprise. DDN Solution Brief
DDN Solution Brief Personal Storage for the Enterprise WOS Cloud Secure, Shared Drop-in File Access for Enterprise Users, Anytime and Anywhere 2011 DataDirect Networks. All Rights Reserved DDN WOS Cloud
More informationFlash Memory Arrays Enabling the Virtualized Data Center. July 2010
Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,
More informationECMWF HPC Workshop: Accelerating Data Management
October 2012 ECMWF HPC Workshop: Accelerating Data Management Massively-Scalable Platforms and Solutions Engineered for the Big Data and Cloud Era Glenn Wright Systems Architect, DDN Data-Driven Paradigm
More informationMaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products
MaxDeploy Ready Hyper- Converged Virtualization Solution With SanDisk Fusion iomemory products MaxDeploy Ready products are configured and tested for support with Maxta software- defined storage and with
More informationScala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
More informationCisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database
Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a
More informationAccelerate > Converged Storage Infrastructure. DDN Case Study. ddn.com. 2013 DataDirect Networks. All Rights Reserved
DDN Case Study Accelerate > Converged Storage Infrastructure 2013 DataDirect Networks. All Rights Reserved The University of Florida s (ICBR) offers access to cutting-edge technologies designed to enable
More informationPanasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory
Customer Success Story Los Alamos National Laboratory Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory June 2010 Highlights First Petaflop Supercomputer
More informationObject storage in Cloud Computing and Embedded Processing
Object storage in Cloud Computing and Embedded Processing Jan Jitze Krol Systems Engineer DDN We Accelerate Information Insight DDN is a Leader in Massively Scalable Platforms and Solutions for Big Data
More informationPerformance in a Gluster System. Versions 3.1.x
Performance in a Gluster System Versions 3.1.x TABLE OF CONTENTS Table of Contents... 2 List of Figures... 3 1.0 Introduction to Gluster... 4 2.0 Gluster view of Performance... 5 2.1 Good performance across
More informationCluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer
Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer Stan Posey, MSc and Bill Loewe, PhD Panasas Inc., Fremont, CA, USA Paul Calleja, PhD University of Cambridge,
More informationBenchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
More informationANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter
WHITEPAPER ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter THE BIG DATA CHALLENGE AND OPPORTUNITY The.proliferation,.management.and.analysis.of.intelligence.data.is.a.fast.growing.concern.
More informationAccelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software
WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications
More informationIsilon IQ Scale-out NAS for High-Performance Applications
Isilon IQ Scale-out NAS for High-Performance Applications Optimizing Performance with Isilon IQ Storage By Shai Harmelin, Sr. Solutions Architect An Isilon Systems Technical Whitepaper July 2009 ISILON
More informationBlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything
BlueArc unified network storage systems 7th TF-Storage Meeting Scale Bigger, Store Smarter, Accelerate Everything BlueArc s Heritage Private Company, founded in 1998 Headquarters in San Jose, CA Highest
More informationEMC XTREMIO EXECUTIVE OVERVIEW
EMC XTREMIO EXECUTIVE OVERVIEW COMPANY BACKGROUND XtremIO develops enterprise data storage systems based completely on random access media such as flash solid-state drives (SSDs). By leveraging the underlying
More informationNetapp @ 10th TF-Storage Meeting
Netapp @ 10th TF-Storage Meeting Wojciech Janusz, Netapp Poland Bogusz Błaszkiewicz, Netapp Poland Ljubljana, 2012.02.20 Agenda Data Ontap Cluster-Mode pnfs E-Series NetApp Confidential - Internal Use
More informationHADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director brett.weninger@adurant.com Dave Smelker, Managing Principal dave.smelker@adurant.com
More informationSolving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
More informationEXPLORATION TECHNOLOGY REQUIRES A RADICAL CHANGE IN DATA ANALYSIS
EXPLORATION TECHNOLOGY REQUIRES A RADICAL CHANGE IN DATA ANALYSIS EMC Isilon solutions for oil and gas EMC PERSPECTIVE TABLE OF CONTENTS INTRODUCTION: THE HUNT FOR MORE RESOURCES... 3 KEEPING PACE WITH
More informationwww.xenon.com.au STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING
www.xenon.com.au STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING GPU COMPUTING VISUALISATION XENON Accelerating Exploration Mineral, oil and gas exploration is an expensive and challenging
More informationwww.thinkparq.com www.beegfs.com
www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
More informationZadara Storage Cloud A whitepaper. @ZadaraStorage
Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers
More informationIOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org
IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator
More informationAchieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
More informationSoftware-defined Storage Architecture for Analytics Computing
Software-defined Storage Architecture for Analytics Computing Arati Joshi Performance Engineering Colin Eldridge File System Engineering Carlos Carrero Product Management June 2015 Reference Architecture
More informationHow To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory)
WHITE PAPER Oracle NoSQL Database and SanDisk Offer Cost-Effective Extreme Performance for Big Data 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Abstract... 3 What Is Big Data?...
More informationApplication Performance for High Performance Computing Environments
Application Performance for High Performance Computing Environments Leveraging the strengths of Computationally intensive applications With high performance scale out file serving In data storage modules
More informationDirect Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era
Enterprise Strategy Group Getting to the bigger truth. White Paper Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era Apeiron introduces NVMe-based storage innovation designed
More informationCloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
More informationHadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
More informationSoftware-defined Storage at the Speed of Flash
TECHNICAL BRIEF: SOFTWARE-DEFINED STORAGE AT THE SPEED OF... FLASH..................................... Intel SSD Data Center P3700 Series and Symantec Storage Foundation with Flexible Storage Sharing
More informationMicrosoft SQL Server 2014 Fast Track
Microsoft SQL Server 2014 Fast Track 34-TB Certified Data Warehouse 103-TB Maximum User Data Tegile Systems Solution Review 2U Design: Featuring Tegile T3800 All-Flash Storage Array http:// www.tegile.com/solutiuons/sql
More informationCommoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre
Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor
More informationGeospatial Imaging Cloud Storage Capturing the World at Scale with WOS TM. ddn.com. DDN Whitepaper. 2011 DataDirect Networks. All Rights Reserved.
DDN Whitepaper Geospatial Imaging Cloud Storage Capturing the World at Scale with WOS TM Table of Contents Growth and Complexity Challenges for Geospatial Imaging 3 New Solutions to Drive Insight, Simplicity
More informationALPS Supercomputing System A Scalable Supercomputer with Flexible Services
ALPS Supercomputing System A Scalable Supercomputer with Flexible Services 1 Abstract Supercomputing is moving from the realm of abstract to mainstream with more and more applications and research being
More informationMaginatics Cloud Storage Platform for Elastic NAS Workloads
Maginatics Cloud Storage Platform for Elastic NAS Workloads Optimized for Cloud Maginatics Cloud Storage Platform () is the first solution optimized for the cloud. It provides lower cost, easier administration,
More informationEMC SOLUTION FOR SPLUNK
EMC SOLUTION FOR SPLUNK Splunk validation using all-flash EMC XtremIO and EMC Isilon scale-out NAS ABSTRACT This white paper provides details on the validation of functionality and performance of Splunk
More informationEssentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS
Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS M ost storage vendors now offer all-flash storage arrays, and many modern organizations recognize the need for these highperformance
More informationMilestone Solution Partner IT Infrastructure Components Certification Summary
Milestone Solution Partner IT Infrastructure Components Certification Summary Dell FS8600 NAS Storage 12-1-2014 Table of Contents Introduction:... 2 Dell Storage Architecture:... 3 Certified Products:...
More informationIEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV
IEEE Mass Storage Conference Vendor Reception Lake Tahoe, NV 11 Manager May 04, 2010 Joe Rotiroti Client Systems IBM, Federal 484 433 9756 cell 845 491 5227 fax rotiroti@us.ibm.com How is GPFS different?
More informationUnderstanding Enterprise NAS
Anjan Dave, Principal Storage Engineer LSI Corporation Author: Anjan Dave, Principal Storage Engineer, LSI Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA
More informationHPC Advisory Council
HPC Advisory Council September 2012, Malaga CHRIS WEEDEN SYSTEMS ENGINEER WHO IS PANASAS? Panasas is a high performance storage vendor founded by Dr Garth Gibson Panasas delivers a fully supported, turnkey,
More informationDeploying Affordable, High Performance Hybrid Flash Storage for Clustered SQL Server
Deploying Affordable, High Performance Hybrid Flash Storage for Clustered SQL Server Flash storage adoption has increased in recent years, as organizations have deployed it to support business applications.
More informationIBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads
89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com @EdisonGroupInc 212.367.7400 IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads A Competitive Test and Evaluation Report
More informationPanasas: High Performance Storage for the Engineering Workflow
9. LS-DYNA Forum, Bamberg 2010 IT / Performance Panasas: High Performance Storage for the Engineering Workflow E. Jassaud, W. Szoecs Panasas / transtec AG 2010 Copyright by DYNAmore GmbH N - I - 9 High-Performance
More informationPlatfora Big Data Analytics
Platfora Big Data Analytics ISV Partner Solution Case Study and Cisco Unified Computing System Platfora, the leading enterprise big data analytics platform built natively on Hadoop and Spark, delivers
More informationOptimizing Large Arrays with StoneFly Storage Concentrators
Optimizing Large Arrays with StoneFly Storage Concentrators All trademark names are the property of their respective companies. This publication contains opinions of which are subject to change from time
More informationPivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview
Pivot3 Desktop Virtualization Appliances vstac VDI Technology Overview February 2012 Pivot3 Desktop Virtualization Technology Overview Table of Contents Executive Summary... 3 The Pivot3 VDI Appliance...
More informationDeploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk
WHITE PAPER Deploying Flash- Accelerated Hadoop with InfiniFlash from SanDisk 951 SanDisk Drive, Milpitas, CA 95035 2015 SanDisk Corporation. All rights reserved. www.sandisk.com Table of Contents Introduction
More informationioscale: The Holy Grail for Hyperscale
ioscale: The Holy Grail for Hyperscale The New World of Hyperscale Hyperscale describes new cloud computing deployments where hundreds or thousands of distributed servers support millions of remote, often
More informationPOWER ALL GLOBAL FILE SYSTEM (PGFS)
POWER ALL GLOBAL FILE SYSTEM (PGFS) Defining next generation of global storage grid Power All Networks Ltd. Technical Whitepaper April 2008, version 1.01 Table of Content 1. Introduction.. 3 2. Paradigm
More informationHow To Get The Most Out Of A Large Data Set
DDN Solution Brief Overcoming > The Big Data Technology Hurdle Turning Data into Answers with DDN & Vertica 20 Networks. All Rights Reserved. Executive Summary Networks and Vertica have collaborated to
More informationCollaborative Research Infrastructure Deployments. ddn.com. Accelerate > DDN Case Study
DDN Case Study Accelerate > Collaborative Research Infrastructure Deployments University College London Transforms Research Collaboration and Data Preservation with Scalable Cloud Object Storage Appliance
More information(Scale Out NAS System)
For Unlimited Capacity & Performance Clustered NAS System (Scale Out NAS System) Copyright 2010 by Netclips, Ltd. All rights reserved -0- 1 2 3 4 5 NAS Storage Trend Scale-Out NAS Solution Scaleway Advantages
More informationNetapp HPC Solution for Lustre. Rich Fenton (fenton@netapp.com) UK Solutions Architect
Netapp HPC Solution for Lustre Rich Fenton (fenton@netapp.com) UK Solutions Architect Agenda NetApp Introduction Introducing the E-Series Platform Why E-Series for Lustre? Modular Scale-out Capacity Density
More informationNetApp High-Performance Computing Solution for Lustre: Solution Guide
Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5
More informationHigh Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology
High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology Evaluation report prepared under contract with Brocade Executive Summary As CIOs
More informationLSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment
LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment Evaluation report prepared under contract with LSI Corporation Introduction Interest in solid-state storage (SSS) is high, and
More informationDirect NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle
Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server
More informationHigh Performance Server SAN using Micron M500DC SSDs and Sanbolic Software
High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software White Paper Overview The Micron M500DC SSD was designed after months of close work with major data center service providers and
More informationHow To Speed Up A Flash Flash Storage System With The Hyperq Memory Router
HyperQ Hybrid Flash Storage Made Easy White Paper Parsec Labs, LLC. 7101 Northland Circle North, Suite 105 Brooklyn Park, MN 55428 USA 1-763-219-8811 www.parseclabs.com info@parseclabs.com sales@parseclabs.com
More informationThe Ultimate in Scale-Out Storage for HPC and Big Data
Node Inventory Health and Active Filesystem Throughput Monitoring Asset Utilization and Capacity Statistics Manager brings to life powerful, intuitive, context-aware real-time monitoring and proactive
More informationEverything you need to know about flash storage performance
Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices
More informationTHE EMC ISILON STORY. Big Data In The Enterprise. Copyright 2012 EMC Corporation. All rights reserved.
THE EMC ISILON STORY Big Data In The Enterprise 2012 1 Big Data In The Enterprise Isilon Overview Isilon Technology Summary 2 What is Big Data? 3 The Big Data Challenge File Shares 90 and Archives 80 Bioinformatics
More informationWOS. High Performance Object Storage
Datasheet WOS High Performance Object Storage The Big Data explosion brings both challenges and opportunities to businesses across all industry verticals. Providers of online services are building infrastructures
More informationFlash Memory Technology in Enterprise Storage
NETAPP WHITE PAPER Flash Memory Technology in Enterprise Storage Flexible Choices to Optimize Performance Mark Woods and Amit Shah, NetApp November 2008 WP-7061-1008 EXECUTIVE SUMMARY Solid state drives
More informationSMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
More informationAn Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine
An Oracle White Paper May 2011 Exadata Smart Flash Cache and the Oracle Exadata Database Machine Exadata Smart Flash Cache... 2 Oracle Database 11g: The First Flash Optimized Database... 2 Exadata Smart
More informationData management challenges in todays Healthcare and Life Sciences ecosystems
Data management challenges in todays Healthcare and Life Sciences ecosystems Jose L. Alvarez Principal Engineer, WW Director Life Sciences jose.alvarez@seagate.com Evolution of Data Sets in Healthcare
More informationA Survey of Shared File Systems
Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...
More informationOracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers
Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers Masood Ahmed EMEA Infrastructure Solutions Oracle/SAP Relationship Overview First SAP R/3 release
More informationAppro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
More informationSolving Agencies Big Data Challenges: PED for On-the-Fly Decisions
White Paper Solving Agencies Big Data Challenges: PED for On-the-Fly Decisions Carina Veksler, NetApp March 2012 WP-7158 ABSTRACT With the growing volumes of rich sensor data and imagery used today to
More informationData Center Solutions
Data Center Solutions Systems, software and hardware solutions you can trust With over 25 years of storage innovation, SanDisk is a global flash technology leader. At SanDisk, we re expanding the possibilities
More informationData Center Storage Solutions
Data Center Storage Solutions Enterprise software, appliance and hardware solutions you can trust When it comes to storage, most enterprises seek the same things: predictable performance, trusted reliability
More informationEasier - Faster - Better
Highest reliability, availability and serviceability ClusterStor gets you productive fast with robust professional service offerings available as part of solution delivery, including quality controlled
More informationMellanox Accelerated Storage Solutions
Mellanox Accelerated Storage Solutions Moving Data Efficiently In an era of exponential data growth, storage infrastructures are being pushed to the limits of their capacity and data delivery capabilities.
More informationNEXSAN NST STORAGE FOR THE VIRTUAL DESKTOP
NST STORAGE FOR THE VIRTUAL DESKTOP Nexsan s innovative product, the NST5000, is a hybrid storage system with unified protocols and highly dense storage for a combination of great performance, low cost,
More informationBlock based, file-based, combination. Component based, solution based
The Wide Spread Role of 10-Gigabit Ethernet in Storage This paper provides an overview of SAN and NAS storage solutions, highlights the ubiquitous role of 10 Gigabit Ethernet in these solutions, and illustrates
More informationData Center Performance Insurance
Data Center Performance Insurance How NFS Caching Guarantees Rapid Response Times During Peak Workloads November 2010 2 Saving Millions By Making It Easier And Faster Every year slow data centers and application
More informationHow to Choose your Red Hat Enterprise Linux Filesystem
How to Choose your Red Hat Enterprise Linux Filesystem EXECUTIVE SUMMARY Choosing the Red Hat Enterprise Linux filesystem that is appropriate for your application is often a non-trivial decision due to
More informationIntroduction to Gluster. Versions 3.0.x
Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster
More informationBig data management with IBM General Parallel File System
Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers
More informationColgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP
selects SAP HANA to improve the speed of business analytics with IBM and SAP Founded in 1806, is a global consumer products company which sells nearly $17 billion annually in personal care, home care,
More informationVMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014
VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup
More informationMaximum performance, minimal risk for data warehousing
SYSTEM X SERVERS SOLUTION BRIEF Maximum performance, minimal risk for data warehousing Microsoft Data Warehouse Fast Track for SQL Server 2014 on System x3850 X6 (95TB) The rapid growth of technology has
More informationGPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
More information