THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC



Similar documents
Sun Constellation System: The Open Petascale Computing Architecture

HPC Update: Engagement Model

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Cluster Implementation and Management; Scheduling

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

New Storage System Solutions

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Scala Storage Scale-Out Clustered Storage White Paper

Boost Database Performance with the Cisco UCS Storage Accelerator

ARCHITECTING COST-EFFECTIVE, SCALABLE ORACLE DATA WAREHOUSES

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

HPC Advisory Council

Sun Microsystems Special Promotions for Education and Research January 9, 2007

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

Clusters: Mainstream Technology for CAE

Optimizing Storage for Better TCO in Oracle Environments. Part 1: Management INFOSTOR. Executive Brief

HUAWEI Tecal E6000 Blade Server

<Insert Picture Here> Refreshing Your Data Protection Environment with Next-Generation Architectures

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

præsentation oktober 2011

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Parallele Dateisysteme für Linux und Solaris. Roland Rambau Principal Engineer GSE Sun Microsystems GmbH

E4 UNIFIED STORAGE powered by Syneto

FLOW-3D Performance Benchmark and Profiling. September 2012

CUTTING-EDGE SOLUTIONS FOR TODAY AND TOMORROW. Dell PowerEdge M-Series Blade Servers

Netapp HPC Solution for Lustre. Rich Fenton UK Solutions Architect

Improved LS-DYNA Performance on Sun Servers

Sun Storage Perspective & Lustre Architecture. Dr. Peter Braam VP Sun Microsystems

Reference Architectures for Repositories and Preservation Archiving

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

PRIMERGY server-based High Performance Computing solutions

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture

Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

Private cloud computing advances

Red Hat Enterprise Linux 6. Stanislav Polášek ELOS Technologies

Panasas: High Performance Storage for the Engineering Workflow

Cisco Unified Computing System and EMC VNXe3300 Unified Storage System

Easier - Faster - Better

IBM System x family brochure

巨 量 資 料 分 層 儲 存 解 決 方 案

STORAGE CENTER WITH NAS STORAGE CENTER DATASHEET

VTrak SATA RAID Storage System

Integrated Grid Solutions. and Greenplum

LS DYNA Performance Benchmarks and Profiling. January 2009

Pentaho High-Performance Big Data Reference Configurations using Cisco Unified Computing System

How To Build A Cisco Ukcsob420 M3 Blade Server

Versity All rights reserved.

Big data management with IBM General Parallel File System

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

Hadoop on the Gordon Data Intensive Cluster

February, 2015 Bill Loewe

Virtual Compute Appliance Frequently Asked Questions

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

Quantum StorNext. Product Brief: Distributed LAN Client

Storage Switzerland White Paper Storage Infrastructures for Big Data Workflows

Cray DVS: Data Virtualization Service

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

Oracle Maximum Availability Architecture with Exadata Database Machine. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

Tier 2 Nearline. As archives grow, Echo grows. Dynamically, cost-effectively and massively. What is nearline? Transfer to Tape

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

Building Clusters for Gromacs and other HPC applications

Data management challenges in todays Healthcare and Life Sciences ecosystems

ECLIPSE Performance Benchmarks and Profiling. January 2009

MESOS CB220. Cluster-in-a-Box. Network Storage Appliance. A Simple and Smart Way to Converged Storage with QCT MESOS CB220

The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.

HUAWEI Tecal E9000 Converged Infrastructure Blade Server

The safer, easier way to help you pass any IT exams. Exam : Storage Sales V2. Title : Version : Demo 1 / 5

Scaling from Workstation to Cluster for Compute-Intensive Applications

Dell s SAP HANA Appliance

Unified Computing Systems

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

Moving Virtual Storage to the Cloud

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Data Center Architecture Overview

(Scale Out NAS System)

Transcription:

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC The Right Data, in the Right Place, at the Right Time José Martins Storage Practice Sun Microsystems 1

Agenda Sun s strategy and commitment to the HPC or technical computing market Storage challenges we hear from HPC customers Addressing these challenges with the Sun Storage and Archive Solution for HPC Next Steps 2

Advancements in HPC HPC is Now Everywhere FINANCE ENGINEERING RESEARCH / DEVELOPMENT Portfolio risk analysis and optimization, derivatives pricing, fraud detection... Simulate mechanical and electronic systems before building costly prototypes Simulate the effects and prognosis of chemotherapy, radiation, and surgery ENTERTAINMENT WEATHER AND CLIMATE OIL AND GAS Render photo-realistic scenes, customer materials, product illustrations and videos Weather and climate forecasting Analyze seismic data to find oil and gas and determine how to extract it in the most efficient way 3

Structured Unstructured The Solution for Large Scale Fixed Content Archival Fixed Data Dynamic Data Problems Fixed Content Market Object-based Storage Media production, E-Research data, digital work in progress ecad, repositories, national mcad, Office docs heritage artifacts, medical images, e-books,design docs, photo services Transactional systems, ERP, CRM data warehousing A new Storage Paradigm

Solution Overview

What We Have Heard: Storage Challenges for HPC High Performance on a Fixed Budget Explosive Growth with Limited Staff Lack of Tuning and Customization Limited Flexibility 6

Addressing these Challenges Sun Storage and Archive Solution for HPC A Comprehensive Environment That: Efficiently addresses all your HPC storage needs by > Integrating multiple storage types > Automating data life cycle and workflow Enables breakthrough economics: > Reduce three year TCO up to 44%* > Reduce energy consumption up to 24%* > Or increase cluster compute power up to 23%*, by reinvesting the savings in more cluster nodes *Based on comparative analysis with typical NAS products serving a compute grid

Sun Storage and Archive Solution for HPC The Right Data, in the Right Place, at the Right Time Key Functions High performance I/O for the compute cluster or grid Reliable and robust home directory space Automated migration of data to the most cost effective storage for the job Seamless user access to data that is moved to tape Continuous automated backup protection of data 8

Leveraging the Most Cost Effective Asset Online Storage Module (Tier 1) Disk Archive Module (Tier 2) Tape Archive Module (Tier 3) Perf Perf Perf Cost/GB Cost/GB Cost/GB Capacity Capacity Capacity

Closer Look

Solution Architecture and Modules Core Data Management Module: manages the environment, data locations and provides the policy interface Storage Access Module (optional): provides cluster working space or scratch space access Online Storage Module: stores cluster working data and user home directories Disk Archive Module: stores data that has been archived either by policy or by user action Tape Archive Module: stores data that has been archived on tape at a lower cost and power burden than the disk archive module Sun Customer Ready Scalable Storage Cluster (optional): very high performance scratch space module, utilizes Lustre and InfiniBand Sun storage; used in place of the Storage Access Module HPC Cluster or Grid Infiniband Network NFS Access Network Storage Access Module Core Data Management Module Scalable Storage Cluster Online Storage Module Disk Archive Module Tape Archive Module 11

Reference Configurations Node Count Guidelines Small Medium Large 32-128 for light I/O per node 32-64 for heavy I/O per node 64-512 for light I/O per node 32-256 for heavy I/O per node 512-1000s for light I/O per node 256-1000s for heavy I/O per node Aggregate Bandwidth Data Access NFS - Cluster I/O 100-400MB/sec NFS- Home Directories NFS - Cluster I/O 400MB 1GB/sec NFS- Home Directories IB - Cluster I/O 1-50+ GB/sec NOTE sizing information is provided as a starting guideline; work with your Sun systems engineer to configure for your specific needs NFS- Home Directories Name Space Combined Scratch and Archive Combined Scratch and Archive Separate Scratch and Archive connected by data mover

Small Configuration Departmental HPC needs Grid Compute Nodes Typically 32-128 nodes Provides up to ~ 400 MB/sec HPC cluster and users are served via NFS (10GE or GE) Core modules are clustered via QFS, add more to scale performance Core module controls data movement between the Online and Archive modules NFS Access Net Switch Core Data Management Module Cluster I/O Data Lifecycle Online Storage Module Disk Archive Module (Tier 1) (Tier 2) Tape Archive Module (Tier 3) 13

Medium Configuration Divisional HPC needs Grid Compute Nodes Typically 64-512 nodes Scales ~ 1 GB/sec Data Access modules are clustered via QFS, add more to scale performance NFS Access Net Data Access Module Cluster I/O HPC cluster or grid is served via NFS by the Data access modules Core Data Management Module Switch Core module controls data movement between the Online and Archive modules Data Lifecycle Online Storage Module (Tier 1) Disk Archive Module (Tier 2) Tape Archive Module (Tier 3) 14

Large Configuration Large scale HPC needs 1-50GB/sec and beyond InfiniBand Network Cluster I/O Users home directories are served via NFS Data movers bridge between the high performance cluster scratch space and archive space Core modules controls data movement between the Online and Archive modules Core Data Management Module Grid Compute Nodes Up to 1000s of nodes HPC cluster or grid is served via InfiniBand and Lustre Grid Compute Nodes SAN Data Movers Data Lifecycle Scalable Storage Cluster Online Storage Module (Tier 1) Disk Archive Module (Tier 2) Tape Archive Module (Tier 3) 15

Thumper Perfect Grid Computing Storage Nod Lustre Grid Deployment MDS MDS server Galaxy server Compute Server Lustre Clients server Compute Server Compute Server Compute Server Compute Server GigE or High Speed Network Fabric Infiniband, Myrinet Lustr OSS server server server Thum

Sun Customer Ready Scalable Storage Cluster Addresses all those challenges: > Fast, incredibly fast > Scalable to the nth degree > Industry-leading storage density and > > > > power-efficiency Exceptional price/performance Field proven design Customer specifics accomodated Configured, built, tested by Sun, ready to run off the pallet

Components and Connectivity Object Storage Servers Data Movers MetaData Servers Standard on large, else optional Infiniband Network Compute Cluster Archive / Permanent Storage Archive Network

Object Storage Server Sun Fire X4500 Server Compute 2 x Dual Core Opteron processors 16GB Memory Storage 48 SATA II drives 24TB raw capacity I/O 6 x SATA channels 2 dual port 4x IB HCAs Availability Manages storage targets, each object contains whole file or partial stripe, controls locking for it's own I/O Hot-swap/plug power, fans, disks Management Same management as other Galaxy servers O/S RHEL 4u4, Lustre OSS

Metadata Server Sun Fire X4200 Server Compute 2 AMD Opteron (200 series) Processors Dualcore 4 GB DDR1 RAM I/O 5x PCI-X slots 4x Gigabit Ethernet ports 1 x 73GB SAS disk drive 2 dual port 4x IB HCAs High Availability Metadata server manages file layout, metadata locking and access/security control Redundant hot-swap Power supplies and Fans, hot-swap disk drives Hardware RAID 0 & 1 Dual MDS configured for high availability Management IPMI 2.0; remote KVM, floppy/cdrom with dedicated 10/100 Ethernet port RHEL 4u4 with Lustre MDS software installed

Data Mover Sun Fire X4600 M2 Server Compute 8 Next-Generation AMD Opteron Processor 8000 Series (dual core) processors Support Multi-generation single, dual-core AMD Opteron 32 DIMM Slots (4 per socket) DDR2-667 16 GB RAM I/O 20GB/s (160Gb/s) bi-directional I/O 6 x PCI-Express slots (4 @ 8X, 2 @ 4X), 2 x PCI-X 4 x GigE standard, USB 2.0, Video, Serial ports 2 x 73 GB disk drives Availability Data mover managers the transfer of data between Lustre and SAM-QFS, proven policy-based archive Hot-swap disk; RAID 0 or 1 built-in 4 power supplies, 2+2 redundant, hot-swap Redundant, hot-swap fans Sun Cluster, MS Cluster support Management ILOM remote power on/off/status, browers+cli, IPMI 2.0, SNMP Remote KVM, Floppy/CDROM N1 System Manager, N1 Service Provisioning System O/S RHEL 4u4, Lustre client software and shared QFS client for Linux

Infiniband Switches Voltaire ISR 9024 10-20 Gbps performance for clusters and grids Ultra-low latency: under 140 nanoseconds Available bandwidth of up to 960 Gbps Powerful CPU to allow management of fabrics, as well as device management capabilities

Lustre Software Architecture Separate file system services allow complete parallelization of I/O functions - enabling massive I/O scalability Deployed in clusters > 256 nodes using IB or other low-latency interconnect Very high metadata and I/O performance > 8,000 file creations/sec in 1 dir, 1,200 nodes > Single clients/servers up to 2.5GB/sec. > Aggregate up to 11GB/sec Scalable to 1,000 s of nodes, petabytes per file system 200-300 commercially supported worldwide deployments including 7 Top10 Supercomputers (according to the November 07 Top500 list):

Scalable Configurations Small Large Small configuration Sun Rack 1000 42 Expansion Sun Rack 1000 42 Sun Rack 1000 42 X4600 X4600 > 48TB1, 1 IB switch Large Configuration > 144 TB1 > 2 IB switches > Data Mover X4200 X4200 X4200 X4200 6140 lite 6140 lite 1 Raw Data Size X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 X4500 ethernet switch IB switch ethernet switch IB switch IB switch ethernet switch IB switch IB switch Expansion Rack > Up to 1921TB > 2 IB switches > Data Mover X4500

Customer Examples

Tsubame Supercomputer Tokyo Institute of Technology #7 Fastest Supercomputer in the world

Tsubame By The Numbers 38.18 TFLOPS in 31 Days 100+ 77 100+ 100+ 77 100+ 100+ 77 100+ x4500 x4600 Thumper Galaxy 44 Thumper Galaxy Thumper Galaxy 4 Thumper Galaxy 4 Thumper Galaxy 4 11 11 11 Infiniband Infiniband Infiniband Tier 22Switch Infiniband Tier Switch Infiniband Tier 22 Switch Infiniband Tier Switch Tier 2 Tier 2Switch Switch Infiniband Tier 1 Switch Infiniband Tier 1 Switch 6 linked sub-clusters 8 Voltaire Infiniband switches 42 Sun Fire x4500 Data Servers 80+ racks 360 Clearspeed FP accelerators 655 Sun Fire x4600 Servers 10480 Opteron cores 21 TeraBytes RAM 1 PetaByte storage N1 Grid Engine N1 System Manager Lustre parallel file system Asia's fastest supercomputer

Our Customers Say It Best: "The Arctic Region Supercomputing Center needed a flexible, cost-effective, high-performance supercomputing environment in order to facilitate a wide range of scientific research, such as ocean and climate modeling, tsunami analysis, regional weather forecasting and applications requiring basic computational fluid dynamics. Sun's new AMD Opteron-powered products, with huge memory, large disk bandwidth and a fast cluster interconnect, comprise an overall cost savings and energy-efficient architecture, making them the ideal systems to drive our compute-intensive work. Frank Williams, Director The Arctic Region Supercomputing Center

Recent Sun Achievements in HPC Deployed Texas Advanced Computing Center: > World's largest supercomputer using general purpose hardware Designed the most open blade platform: > Simultaneous support for AMD, Intel and UltraSparc blades; Solaris, Linux and Windows OS Created the world's first open Acquired Cluster File petascale architecture: Systems, developer of Lustre: > The Sun Constellation System > The world's fastest and most solutions > Over 37% of the world's data scalable parallel file system Produced a complete solution Developed the most scalable portfolio for HPC: & efficient Tape Library > Clustering, visualization, portfolio: storage and software is storage on Sun libraries

The Sun Cluster Portfolio Open, Seamless and Comprehensive Access Visualization, Workstation, Thin Clients, Remote Developer Management Compilers, Workload, Debuggers, Systems and Optimization Cluster Tools, Management Libraries Sun Services OS Interconnect Linux or Solaris InfiniBand or Ethernet Storage Storage/ Archive Cluster Racks or Storage, Blades Backup, Variety of Archive, File CPU Systems, Architectures HSM Compute Networking Systems Sun Customer Ready

AMD (EDA) Global supplier of processors > Accelerate electronic chip designs AMD (EDA) (MCAE) Global supplier of processors Metaldyne Global designer of engine, driveline and chassis > Accelerate electronic chip designs products > Improved(MCAE) mechanical design designer and time toofmarket Metaldyne Global engine, driveline and products chassis University College London Medical Research establishment > Improved mechanical design and time to market > Improved visualization to advance biotechnology research University College London Medical Research establishment Paramount Pictures Media and entertainment company > Improved visualization to advance biotechnology research > Accelerate video rendering and animation brought Barnyard release date forward three months Media and entertainment company Paramount Pictures Accelerate rendering and animation brought Barnyard release date >Cedars Sinaivideo Medical Center forward three months > Developing treatments for Cancer & Heart disease Cedars Sinai servers Medical Center 4TB data daily > 400 SunFire generating >> Developing treatments for Cancer & Heart disease Saved $60K and two months deployment time by using CRS > 400 SunFire servers generating 4TB data daily > Saved $60K and two months deployment time by using CRS

Getting Started Today For more information: > sun.com/hpcdata Learn more at Radio HPC: > hpcradio.blogspot.com/ Learn more about Sun s HPC solutions: > sun.com/hpc Customers that have deployed Sun HPC solutions: > sun.com/servers/hpc/customer_references.jsp Contact Sun: > sun.com/servers/hpc/start.jsp

The Sun StorageTek 5800 will enable you to Preserve & Protect Data Assets For Long Term Manage More With Less Save Time & Reduce Cost & Complexity Complexity Efficiently Manage & Scale More Data Assets Reduce

Double Parity Protection and Intelligent Data Placement Algorithm Data object is broken into 5 data and 2 parity fragments Data is evenly spread across all disks 34

Virtual Views Example: Medical Imaging Doctor = Smith Patient = BrianParks Patient_sex = M Modality = MRI Disease = Dropping ST5320 on hand Vendor = GE Date = 20061001 Define for_doctor Area View = Left Hand = by Doctor, Patient, Modality, Date, Area+.jpg /for_doctor/smith/brianparks/mri/20061001/hand.jpg Define View for_researcher = by Patient_sex, Disease, Area, Date+.jpg /for_researcher/male/injury/hand/20061001.jpg Define View for_technician = by Vendor, Modality, Date+.jpg /for_technician/ge/mri/20061001.jpg 35

THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC The Right Data, in the Right Place, at the Right Time jose.martins@sun.com 36