The Lustre File System. Eric Barton Lead Engineer, Lustre Group Sun Microsystems
|
|
- Antonia Ray
- 8 years ago
- Views:
Transcription
1 The Lustre File System Eric Barton Lead Engineer, Lustre Group Sun Microsystems 1
2 Lustre Today What is Lustre Deployments Community Topics Lustre Development Industry Trends Scalability Improvements
3 Lustre File System The world's fastest, most scalable file system Parallel shared POSIX file system Scalable High performance Petabytes of storage Tens of thousands of clients Coherent Single namespace Strict concurrency control Heterogeneous networking High availability GPL Open source Multi-platform Multi-vendor
4 Lustre File System Major components Client Client Client Client Client Client Client Client MGS configuration MDS MDS namespace MDS data
5 Lustre Networking Simple Message Queues RDMA Active - Get/Put Passive Attach Asynchronous Events Error handling Unlink Layered LNET / LND Multiple Networks Routers RPC Queued requests RDMA bulk RDMA reply Recovery Resend Replay
6 A Lustre Cluster Metadata Servers (MDS) I/O Servers () MDS 1 (active) MDS 2 (standby) Multiple Networks 1 Commodity Storage Servers TCP/IP QSNet 2 Lustre Clients 10 s - 10,000 s Myrinet InfiniBand iwarp Cray Seastar 3 Shared storage enables failover Router 4 5 = failover 6 Enterprise-Class Storage Arrays & SAN Fabrics 7
7 Lustre Today Lustre is the leading HPC file system > 7 of Top 10 > Over 40% of Top100 Demonstrated scalability and performance > 190GB/sec IO > 26,000 clients > Many systems with over 1,000 nodes
8 Livermore Blue Gene/L SCF 3.5 PB storage; 52 GB/s I/O throughput 131,072 processor cores TACC Ranger 1.73 PB storage; 40GB/s I/O throughput 62,976 processor cores Sandia Red Storm 340 TB Storage; 50GB/s I/O throughput 12,960 multi-core compute sockets ORNL Jaguar 10.5PB storage; 240 GB/s I/O throughput goal 265,708 processor cores
9 Center-wide File System Spider will provide a shared, parallel file system for all systems Based on Lustre file system Demonstrated bandwidth of over 190 GB/s Over 10 PB of RAID-6 Capacity 13,440 1 TB SATA Drives 192 Storage servers 3 TeraBytes of memory Available from all systems via our high-performance scalable I/O network Over 3,000 InfiniBand ports Over 3 miles of cables Scales as storage grows Undergoing system checkout with deployment expected in summer 2009
10 Future LCF Infrastructure Everest Powerwall Remote Visualization Cluster End-to-End Cluster Application Development Cluster Data Archive 25 PB SION 192x 48x 192x XT5 Login XT4 Spider
11 Lustre Success - Media Customer challenges > Eliminate data storage bottlenecks resulting from scalability issues NFS can't handle > Increase system performance and reliability Lustre value > Doubled data storage at three times less cost of compelling solutions > The ability to provide a single file system namespace to its production artists > Easy-to-install open source software with great flexibility on storage and server hardware While we were working on The Golden Compass, we faced the most intensive I/O requirements on any project to date. Lustre played a vital role in helping us to deliver this project. Daire Byrne, senior systems integrator, Framestore
12 Lustre success - Telecommunications Customer challenges > Provide scalable service > Ensure continuous availability > Control costs NBC broadcast 2008 Summer Olympics live online over Level 3 network using Lustre Lustre value > The ability to scale easily > Works well with commodity equipment from multiple vendors > High performance and stability With Lustre, we can achieve that balancing act of maintaining a reliable network with lesscostly equipment. It allows us to replace servers and expand the network quickly and easily - Kenneth Brookman, Level 3 Communications
13 Lustre success - Energy Customer challenges > Process huge and growing volumes of data > Keep hardware costs manageable > Scale existing cluster easily Lustre value > Ability to handle exponential growth in data > Capability to scale computer clusters easily > Reduced hardware costs > Reduced maintenance costs More Success
14 Open Source Community Lustre OEM Partners
15 Open Source Community Resources Web News and information Operations Manual Detailed technical documentation Mailing lists > General/operational issues > Architecture and features Bugzilla Defect tracking and patch database CVS repository Lustre Internals training material
16 HPC Trends Processor performance / RAM growing faster than I/O Relative number of I/O devices must grow to compensate Storage component reliability not increasing with capacity > Failure is not an option it s guaranteed Trend to shared file systems Multiple compute clusters Direct access from specialized systems Storage scalability critical
17 DARPA HPCS Capacity 1 trillion files per file system 10 billion files per directory 100 PB system capacity 1 PB single file size >30k client nodes 100,000 open files Reliability End-to-end data integrity No performance impact during RAID build Performance 40,000 file creates/sec > Single client node 30GB/sec streaming data > Single client node 240GB/sec aggregate I/O > File per process > Shared file
18 Lustre and the Future Continued focus on extreme HPC Capacity > Exabytes of storage > Trillions of files > Many client clusters each with 100,000's of clients Performance > TB's/sec of aggregate I/O > 100,000's of aggregate metadata ops/sec Community Driven Tools and Interfaces > Management and Performance Analysis
19 HPC Center of the Future Capability 500,000 Nodes Capacity 1 250,000 Nodes Capacity 2 150,000 Nodes Capacity 3 50,000 Nodes Test 25,000 Nodes Viz 1 Viz 2 WAN Access Shared Storage Network 10 TB/sec User Data 1000 MDTs Metadata 25 MDS s HPSS Archive Lustre Storage Cluster
20 Lustre Scalability Definition Performance / capacity grows nearly linearly with hardware Component failure does not have a disproportionate impact on availability Requirements Scalable I/O & MD performance Expanded component size/count limits Increased robustness to component failure Overhead grows sub-linearly with system size Timely failure detection & recovery
21 Lustre Scaling
22 Architectural Improvements Clustered Metadata (CMD) 10s 100s of metadata servers Distributed inodes > Files local to parent directory entry / subdirs may be non-local Distributed directories > Hashing Striping Distributed Operation Resilience/Recovery > Uncommon HPC workload - Cross-directory rename > Short term - Sequenced cross-mds ops > Longer term - Transactional - ACID - Non-blocking - deeper pipelines - Hard - cascading aborts, synch ops
23 Epochs Global Oldest Volatile Epoch Reduction Network Oldest Epoch Current Globally Known Oldest Volatile Epoch Newest Epoch Stable Unstable Committed Uncommitted Server 1 Updates Server 2 Server 3 Operations Client 1 Client 2 Local Oldest Volatile Epochs Redo
24 Architectural Improvements Fault Detection Today RPC timeout > Timeouts must scale O(n) to distinguish death / congestion Pinger > No aggregation across clients or servers > O(n) ping overhead Routed Networks > Router failure can be confused with end-to-end peer failure Fully automatic failover scales with slowest time constant > Many 10s of minutes on large clusters > Failover could be much faster if useless waiting eliminated
25 Architectural Improvements Scalable Health Network Burden of monitoring clients distributed not replicated > ORNL 35,000 clients, 192 s, 7 OSTs/ Fault-tolerant status reduction/broadcast network > Servers and LNET routers LNET high-priority small message support > Health network stays responsive Prompt, reliable detection > Time constants in seconds > Failed servers, clients and routers > Recovering servers and routers Interface with existing RAS infrastructure Receive and deliver status notification
26 Health Monitoring Network Primary Health Monitor Failover Health Monitor Client
27 Architectural Improvements Metadata Writeback Cache Avoids unnecessary server communications > Operations logged/cached locally > Performance of a local file system when uncontended Aggregated distributed operations > Server updates batched and tranferred using bulk protocols (RDMA) > Reduced network and service overhead Sub-Tree Locking > Lock aggregation a single lock protects a whole subtree > Reduce lock traffic and server load
28 Architectural Improvements Current - Flat Communications model Stateful client/server connection required for coherence and performance Every client connects to every server O(n) lock conflict resolution Future - Hierarchical Communications Model Aggregate connections, locking, I/O, metadata ops Caching clients > Aggregate local processes (cores) > I/O Forwarders scale another 32x or more Caching Proxies > Aggregate whole clusters > Implicit Broadcast - scalable conflict resolution
29 Hierarchical Communications Lustre Storage Cluster MDS MDS MDS MDS MDS MDS Proxy Cluster Proxy Cluster Proxy Server WBC Client Proxy Server Proxy Server WBC Client Proxy Server Proxy Server WBC Client Proxy Server User Proc. Lustre Client IO Forwarding Server I/O Forwarder WBC Client IO Forwarding Server I/O Forwarder User Proc. IO Forwarding Client IO Forwarding Client IO Forwarding Client User Proc. User Proc. User Proc. User Proc. User Proc. User Proc. User Proc. User Proc. Proxy Server WBC Client Proxy Server WAN / Security Domain WBC Client User Proc. Luster Comm System Calls
30 ZFS End-to-end data integrity Checksums in block pointers Ditto blocks Transactional mirroring/raid Remove ldiskfs size limits Immense Capacity (128 bit) No limits on files, dirents etc COW Transactional Snapshots
31 Performance Improvements SMP Scaling Improve MDS performance / small message handling CPU affinity Finer granularity locking # Client Nodes RPC Trhoughput RPC Througput Total client processes Total client processes
32 Load (Im)Balance Request Queue Depth Time Server #
33 Network Request Scheduler Much larger working set than disk elevator Higher level information - client, object, offset, job/rank Prototype Initial development on simulator Scheduling strategies - quanta, offset, fairness etc. Testing at ORNL pending Future Exchange global information - gang scheduling QoS - Real time / Bandwidth reservation (min/max)
34 Metadata Protocol Improvements Size on MDT (SOM) Avoid multiple RPCs for attributes derived from OSTs OSTs remain definitive while file open Compute on close and cache on MDT Readdir+ Aggregation > Directory I/O > Getattrs > Locking
35 Lustre Scalability Attribute Today Future Number of Clients 10,000s Flat comms model 1,000,000s Hierarchical comms model Server Capacity Ext3 8TB ZFS - Petabytes Metadata Performance Single MDS CMD SMP scaling Recovery Time RPC timeout - O(n) Health Network - O(log n)
36 THANK YOU Eric Barton 36
New Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
More informationSun Storage Perspective & Lustre Architecture. Dr. Peter Braam VP Sun Microsystems
Sun Storage Perspective & Lustre Architecture Dr. Peter Braam VP Sun Microsystems Agenda Future of Storage Sun s vision Lustre - vendor neutral architecture roadmap Sun s view on storage introduction The
More informationA Scalable Health Network For Lustre
Lustre User Group Orlando Fl April 2011 A Scalable Health Network For Lustre Eric Barton CTO Whamcloud, Inc LNET Fault Detection Today Based on LND timeout Independent of Lustre timeout Token buildup if
More informationAn Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing
An Alternative Storage Solution for MapReduce Eric Lomascolo Director, Solutions Marketing MapReduce Breaks the Problem Down Data Analysis Distributes processing work (Map) across compute nodes and accumulates
More informationLustre Networking BY PETER J. BRAAM
Lustre Networking BY PETER J. BRAAM A WHITE PAPER FROM CLUSTER FILE SYSTEMS, INC. APRIL 2007 Audience Architects of HPC clusters Abstract This paper provides architects of HPC clusters with information
More informationNetApp High-Performance Computing Solution for Lustre: Solution Guide
Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5
More informationArchitecting a High Performance Storage System
WHITE PAPER Intel Enterprise Edition for Lustre* Software High Performance Data Division Architecting a High Performance Storage System January 2014 Contents Introduction... 1 A Systematic Approach to
More informationCray DVS: Data Virtualization Service
Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with
More informationLUSTRE FILE SYSTEM High-Performance Storage Architecture and Scalable Cluster File System White Paper December 2007. Abstract
LUSTRE FILE SYSTEM High-Performance Storage Architecture and Scalable Cluster File System White Paper December 2007 Abstract This paper provides basic information about the Lustre file system. Chapter
More informationHigh Performance Computing OpenStack Options. September 22, 2015
High Performance Computing OpenStack PRESENTATION TITLE GOES HERE Options September 22, 2015 Today s Presenters Glyn Bowden, SNIA Cloud Storage Initiative Board HP Helion Professional Services Alex McDonald,
More informationLessons learned from parallel file system operation
Lessons learned from parallel file system operation Roland Laifer STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association
More informationCurrent Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
More informationPerformance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
More informationPOWER ALL GLOBAL FILE SYSTEM (PGFS)
POWER ALL GLOBAL FILE SYSTEM (PGFS) Defining next generation of global storage grid Power All Networks Ltd. Technical Whitepaper April 2008, version 1.01 Table of Content 1. Introduction.. 3 2. Paradigm
More informationCommoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre
Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor
More informationCOSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
More informationParallele Dateisysteme für Linux und Solaris. Roland Rambau Principal Engineer GSE Sun Microsystems GmbH
Parallele Dateisysteme für Linux und Solaris Roland Rambau Principal Engineer GSE Sun Microsystems GmbH 1 Agenda kurze Einführung QFS Lustre pnfs ( Sorry... ) Roland.Rambau@Sun.Com Sun Proprietary/Confidential
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
More informationScala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
More informationLustre: A Scalable, High-Performance File System Cluster File Systems, Inc.
Lustre: A Scalable, High-Performance File System Cluster File Systems, Inc. Abstract: Today's network-oriented computing environments require high-performance, network-aware file systems that can satisfy
More informationDriving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
More informationBig data management with IBM General Parallel File System
Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers
More information(Scale Out NAS System)
For Unlimited Capacity & Performance Clustered NAS System (Scale Out NAS System) Copyright 2010 by Netclips, Ltd. All rights reserved -0- 1 2 3 4 5 NAS Storage Trend Scale-Out NAS Solution Scaleway Advantages
More informationBlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything
BlueArc unified network storage systems 7th TF-Storage Meeting Scale Bigger, Store Smarter, Accelerate Everything BlueArc s Heritage Private Company, founded in 1998 Headquarters in San Jose, CA Highest
More informationCOSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel Computation Parallel I/O (I) I/O basics Spring 2008 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network
More informationCluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
More informationDirect NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle
Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server
More informationHPC Update: Engagement Model
HPC Update: Engagement Model MIKE VILDIBILL Director, Strategic Engagements Sun Microsystems mikev@sun.com Our Strategy Building a Comprehensive HPC Portfolio that Delivers Differentiated Customer Value
More informationIntroduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
More informationSun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
More informationNetwork Attached Storage. Jinfeng Yang Oct/19/2015
Network Attached Storage Jinfeng Yang Oct/19/2015 Outline Part A 1. What is the Network Attached Storage (NAS)? 2. What are the applications of NAS? 3. The benefits of NAS. 4. NAS s performance (Reliability
More informationHow to Choose your Red Hat Enterprise Linux Filesystem
How to Choose your Red Hat Enterprise Linux Filesystem EXECUTIVE SUMMARY Choosing the Red Hat Enterprise Linux filesystem that is appropriate for your application is often a non-trivial decision due to
More informationDesign and Evolution of the Apache Hadoop File System(HDFS)
Design and Evolution of the Apache Hadoop File System(HDFS) Dhruba Borthakur Engineer@Facebook Committer@Apache HDFS SDC, Sept 19 2011 Outline Introduction Yet another file-system, why? Goals of Hadoop
More informationStorage Challenges for Petascale Systems
Storage Challenges for Petascale Systems Dilip D. Kandlur Director, Storage Systems Research IBM Research Division 2004 IBM Corporation Outline Storage Technology Trends Implications for high performance
More informationBlobSeer: Towards efficient data storage management on large-scale, distributed systems
: Towards efficient data storage management on large-scale, distributed systems Bogdan Nicolae University of Rennes 1, France KerData Team, INRIA Rennes Bretagne-Atlantique PhD Advisors: Gabriel Antoniu
More informationQuantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
More informationDistributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
More informationEnterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011 Executive Summary Large enterprise Hyper-V deployments with a large number
More informationFebruary, 2015 Bill Loewe
February, 2015 Bill Loewe Agenda System Metadata, a growing issue Parallel System - Lustre Overview Metadata and Distributed Namespace Test setup and implementation for metadata testing Scaling Metadata
More informationApplication Performance for High Performance Computing Environments
Application Performance for High Performance Computing Environments Leveraging the strengths of Computationally intensive applications With high performance scale out file serving In data storage modules
More informationwww.thinkparq.com www.beegfs.com
www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
More informationIntroduction to Gluster. Versions 3.0.x
Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster
More informationSAN Conceptual and Design Basics
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer
More informationGPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"
GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID
More informationHadoop: Embracing future hardware
Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop
More informationSolving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
More informationFlash Performance for Oracle RAC with PCIe Shared Storage A Revolutionary Oracle RAC Architecture
Flash Performance for Oracle RAC with PCIe Shared Storage Authored by: Estuate & Virident HGST Table of Contents Introduction... 1 RAC Share Everything Architecture... 1 Oracle RAC on FlashMAX PCIe SSDs...
More informationHighly-Available Distributed Storage. UF HPC Center Research Computing University of Florida
Highly-Available Distributed Storage UF HPC Center Research Computing University of Florida Storage is Boring Slow, troublesome, albatross around the neck of high-performance computing UF Research Computing
More informationIBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand.
IBM Global Technology Services September 2007 NAS systems scale out to meet Page 2 Contents 2 Introduction 2 Understanding the traditional NAS role 3 Gaining NAS benefits 4 NAS shortcomings in enterprise
More informationWrite a technical report Present your results Write a workshop/conference paper (optional) Could be a real system, simulation and/or theoretical
Identify a problem Review approaches to the problem Propose a novel approach to the problem Define, design, prototype an implementation to evaluate your approach Could be a real system, simulation and/or
More informationDistributed File System Choices: Red Hat Storage, GFS2 & pnfs
Distributed File System Choices: Red Hat Storage, GFS2 & pnfs Ric Wheeler Architect & Senior Manager, Red Hat June 27, 2012 Overview Distributed file system basics Red Hat distributed file systems Performance
More informationHIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010
White Paper HIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010 Abstract This white paper demonstrates key functionality demonstrated in a lab environment
More informationBeyond Embarrassingly Parallel Big Data. William Gropp www.cs.illinois.edu/~wgropp
Beyond Embarrassingly Parallel Big Data William Gropp www.cs.illinois.edu/~wgropp Messages Big is big Data driven is an important area, but not all data driven problems are big data (despite current hype).
More informationScaling Objectivity Database Performance with Panasas Scale-Out NAS Storage
White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage
More informationCeph. A file system a little bit different. Udo Seidel
Ceph A file system a little bit different Udo Seidel Ceph what? So-called parallel distributed cluster file system Started as part of PhD studies at UCSC Public announcement in 2006 at 7 th OSDI File system
More informationEOFS Workshop Paris Sept, 2011. Lustre at exascale. Eric Barton. CTO Whamcloud, Inc. eeb@whamcloud.com. 2011 Whamcloud, Inc.
EOFS Workshop Paris Sept, 2011 Lustre at exascale Eric Barton CTO Whamcloud, Inc. eeb@whamcloud.com Agenda Forces at work in exascale I/O Technology drivers I/O requirements Software engineering issues
More informationIBM System x GPFS Storage Server
IBM System x GPFS Storage Server Schöne Aussicht en für HPC Speicher ZKI-Arbeitskreis Paderborn, 15.03.2013 Karsten Kutzer Client Technical Architect Technical Computing IBM Systems & Technology Group
More informationSciDAC Petascale Data Storage Institute
SciDAC Petascale Data Storage Institute Advanced Scientific Computing Advisory Committee Meeting October 29 2008, Gaithersburg MD Garth Gibson Carnegie Mellon University and Panasas Inc. SciDAC Petascale
More informationBusiness-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000
Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000 Clear the way for new business opportunities. Unlock the power of data. Overcoming storage limitations Unpredictable data growth
More informationFour Reasons To Start Working With NFSv4.1 Now
Four Reasons To Start Working With NFSv4.1 Now PRESENTATION TITLE GOES HERE Presented by: Alex McDonald Hosted by: Gilles Chekroun Ethernet Storage Forum Members The SNIA Ethernet Storage Forum (ESF) focuses
More informationMilestone Solution Partner IT Infrastructure MTP Certification Report Scality RING Software-Defined Storage 11-16-2015
Milestone Solution Partner IT Infrastructure MTP Certification Report Scality RING Software-Defined Storage 11-16-2015 Table of Contents Introduction... 4 Certified Products... 4 Key Findings... 5 Solution
More informationPanasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF
Panasas at the RCF HEPiX at SLAC Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory Centralized File Service Single, facility-wide namespace for files. Uniform, facility-wide
More informationRed Hat Cluster Suite
Red Hat Cluster Suite HP User Society / DECUS 17. Mai 2006 Joachim Schröder Red Hat GmbH Two Key Industry Trends Clustering (scale-out) is happening 20% of all servers shipped will be clustered by 2006.
More informationMonitoring Tools for Large Scale Systems
Monitoring Tools for Large Scale Systems Ross Miller, Jason Hill, David A. Dillow, Raghul Gunasekaran, Galen Shipman, Don Maxwell Oak Ridge Leadership Computing Facility, Oak Ridge National Laboratory
More informationEnabling High performance Big Data platform with RDMA
Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery
More informationPanasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory
Customer Success Story Los Alamos National Laboratory Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory June 2010 Highlights First Petaflop Supercomputer
More informationHigh Availability Databases based on Oracle 10g RAC on Linux
High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN, June 2006 Luca Canali, CERN IT Outline Goals Architecture of an HA DB Service Deployment at the CERN Physics Database
More informationVMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014
VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup
More informationioscale: The Holy Grail for Hyperscale
ioscale: The Holy Grail for Hyperscale The New World of Hyperscale Hyperscale describes new cloud computing deployments where hundreds or thousands of distributed servers support millions of remote, often
More informationDiagram 1: Islands of storage across a digital broadcast workflow
XOR MEDIA CLOUD AQUA Big Data and Traditional Storage The era of big data imposes new challenges on the storage technology industry. As companies accumulate massive amounts of data from video, sound, database,
More informationHPC Advisory Council
HPC Advisory Council September 2012, Malaga CHRIS WEEDEN SYSTEMS ENGINEER WHO IS PANASAS? Panasas is a high performance storage vendor founded by Dr Garth Gibson Panasas delivers a fully supported, turnkey,
More informationCloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
More informationAgenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
More informationWith DDN Big Data Storage
DDN Solution Brief Accelerate > ISR With DDN Big Data Storage The Way to Capture and Analyze the Growing Amount of Data Created by New Technologies 2012 DataDirect Networks. All Rights Reserved. The Big
More informationLustre * Filesystem for Cloud and Hadoop *
OpenFabrics Software User Group Workshop Lustre * Filesystem for Cloud and Hadoop * Robert Read, Intel Lustre * for Cloud and Hadoop * Brief Lustre History and Overview Using Lustre with Hadoop Intel Cloud
More informationWindows Server 2008 Essentials. Installation, Deployment and Management
Windows Server 2008 Essentials Installation, Deployment and Management Windows Server 2008 Essentials First Edition. This ebook is provided for personal use only. Unauthorized use, reproduction and/or
More informationHigh Availability with Windows Server 2012 Release Candidate
High Availability with Windows Server 2012 Release Candidate Windows Server 2012 Release Candidate (RC) delivers innovative new capabilities that enable you to build dynamic storage and availability solutions
More informationAppro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
More informationADVANCED NETWORK CONFIGURATION GUIDE
White Paper ADVANCED NETWORK CONFIGURATION GUIDE CONTENTS Introduction 1 Terminology 1 VLAN configuration 2 NIC Bonding configuration 3 Jumbo frame configuration 4 Other I/O high availability options 4
More informationDistributed File Systems
Distributed File Systems Mauro Fruet University of Trento - Italy 2011/12/19 Mauro Fruet (UniTN) Distributed File Systems 2011/12/19 1 / 39 Outline 1 Distributed File Systems 2 The Google File System (GFS)
More informationAchieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
More informationWHITE PAPER SCALABLE NETWORKED STORAGE. Convergence of SAN and NAS with HighRoad. www.wintercorp.com SPONSORED RESEARCH PROGRAM
WHITE PAPER W I N T E R C O R P O R A T I O N SCALABLE NETWORKED STORAGE Convergence of SAN and NAS with HighRoad www.wintercorp.com SPONSORED RESEARCH PROGRAM Scalable Networked Storage: Convergence of
More informationChapter 11 Distributed File Systems. Distributed File Systems
Chapter 11 Distributed File Systems Introduction Case studies NFS Coda 1 Distributed File Systems A distributed file system enables clients to access files stored on one or more remote file servers A file
More informationLustre failover experience
Lustre failover experience Lustre Administrators and Developers Workshop Paris 1 September 25, 2012 TOC Who we are Our Lustre experience: the environment Deployment Benchmarks What's next 2 Who we are
More informationDr Markus Hagenbuchner markus@uow.edu.au CSCI319. Distributed Systems
Dr Markus Hagenbuchner markus@uow.edu.au CSCI319 Distributed Systems CSCI319 Chapter 8 Page: 1 of 61 Fault Tolerance Study objectives: Understand the role of fault tolerance in Distributed Systems. Know
More informationOracle Maximum Availability Architecture with Exadata Database Machine. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska
Oracle Maximum Availability Architecture with Exadata Database Machine Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska MAA is Oracle s Availability Blueprint Oracle s MAA is a best practices
More informationNetwork File System (NFS) Pradipta De pradipta.de@sunykorea.ac.kr
Network File System (NFS) Pradipta De pradipta.de@sunykorea.ac.kr Today s Topic Network File System Type of Distributed file system NFS protocol NFS cache consistency issue CSE506: Ext Filesystem 2 NFS
More informationDistributed Software Development with Perforce Perforce Consulting Guide
Distributed Software Development with Perforce Perforce Consulting Guide Get an overview of Perforce s simple and scalable software version management solution for supporting distributed development teams.
More informationLustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.
Reference Architecture Designing High-Performance Storage Tiers Designing High-Performance Storage Tiers Intel Enterprise Edition for Lustre* software and Intel Non-Volatile Memory Express (NVMe) Storage
More informationEMC BACKUP MEETS BIG DATA
EMC BACKUP MEETS BIG DATA Strategies To Protect Greenplum, Isilon And Teradata Systems 1 Agenda Big Data: Overview, Backup and Recovery EMC Big Data Backup Strategy EMC Backup and Recovery Solutions for
More informationFeature Comparison. Windows Server 2008 R2 Hyper-V and Windows Server 2012 Hyper-V
Comparison and Contents Introduction... 4 More Secure Multitenancy... 5 Flexible Infrastructure... 9 Scale, Performance, and Density... 13 High Availability... 18 Processor and Memory Support... 24 Network...
More informationSTORAGE CENTER WITH NAS STORAGE CENTER DATASHEET
STORAGE CENTER WITH STORAGE CENTER DATASHEET THE BENEFITS OF UNIFIED AND STORAGE Combining block and file-level data into a centralized storage platform simplifies management and reduces overall storage
More informationCan High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
More informationHigh Performance Computing Specialists. ZFS Storage as a Solution for Big Data and Flexibility
High Performance Computing Specialists ZFS Storage as a Solution for Big Data and Flexibility Introducing VA Technologies UK Based System Integrator Specialising in High Performance ZFS Storage Partner
More informationHigh Performance Server SAN using Micron M500DC SSDs and Sanbolic Software
High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software White Paper Overview The Micron M500DC SSD was designed after months of close work with major data center service providers and
More informationHADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director brett.weninger@adurant.com Dave Smelker, Managing Principal dave.smelker@adurant.com
More informationBigdata High Availability (HA) Architecture
Bigdata High Availability (HA) Architecture Introduction This whitepaper describes an HA architecture based on a shared nothing design. Each node uses commodity hardware and has its own local resources
More informationBig + Fast + Safe + Simple = Lowest Technical Risk
Big + Fast + Safe + Simple = Lowest Technical Risk The Synergy of Greenplum and Isilon Architecture in HP Environments Steffen Thuemmel (Isilon) Andreas Scherbaum (Greenplum) 1 Our problem 2 What is Big
More informationObject storage in Cloud Computing and Embedded Processing
Object storage in Cloud Computing and Embedded Processing Jan Jitze Krol Systems Engineer DDN We Accelerate Information Insight DDN is a Leader in Massively Scalable Platforms and Solutions for Big Data
More informationHDFS Architecture Guide
by Dhruba Borthakur Table of contents 1 Introduction... 3 2 Assumptions and Goals... 3 2.1 Hardware Failure... 3 2.2 Streaming Data Access...3 2.3 Large Data Sets... 3 2.4 Simple Coherency Model...3 2.5
More information