Low-cost storage @PSNC

Similar documents
Building low cost disk storage with Ceph and OpenStack Swift

Scientific Computing Data Management Visions

StorPool Distributed Storage Software Technical Overview

præsentation oktober 2011

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Deploying Ceph with High Performance Networks, Architectures and benchmarks for Block Storage Solutions

Scaling from Datacenter to Client

Terms of Reference Microsoft Exchange and Domain Controller/ AD implementation

VTrak SATA RAID Storage System

SOP Common service PC File Server

Increasing Storage Performance

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

PSAM, NEC PCIe SSD Appliance for Microsoft SQL Server (Reference Architecture) September 11 th, 2014 NEC Corporation

Evaluation of Dell PowerEdge VRTX Shared PERC8 in Failover Scenario

Building All-Flash Software Defined Storages for Datacenters. Ji Hyuck Yun Storage Tech. Lab SK Telecom

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Solid State Storage in the Evolution of the Data Center

How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda

Cloud Computing and Amazon Web Services

Building Storage Clouds for Online Applications A Case for Optimized Object Storage

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

Fujitsu PRIMERGY Servers Portfolio

Accelerating Server Storage Performance on Lenovo ThinkServer

UCS M-Series Modular Servers

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Zadara Storage Cloud A

Parallels Cloud Storage

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

FLOW-3D Performance Benchmark and Profiling. September 2012

FUJITSU Enterprise Product & Solution Facts

F600Q 8Gb FC Storage Performance Report Date: 2012/10/30

Product Spotlight. A Look at the Future of Storage. Featuring SUSE Enterprise Storage. Where IT perceptions are reality

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Data storage services at CC-IN2P3

Software-defined Storage at the Speed of Flash

Cost Efficient VDI. XenDesktop 7 on Commodity Hardware

Data Protection Technologies: What comes after RAID? Vladimir Sapunenko, INFN-CNAF HEPiX Spring 2012 Workshop

Fujitsu PRIMERGY Servers Portfolio

nexsan NAS just got faster, easier and more affordable.

QUESTIONS & ANSWERS. ItB tender 72-09: IT Equipment. Elections Project

Alexandria Overview. Sept 4, 2015

Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array

HUAWEI TECHNOLOGIES CO., LTD. HUAWEI FusionServer X6800 Data Center Server

VDI Without Compromise with SimpliVity OmniStack and Citrix XenDesktop

Hadoop: Embracing future hardware

Motherboard- based Servers versus ATCA- based Servers

Microsoft Exchange Server 2003 Deployment Considerations

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

HUAWEI Tecal E6000 Blade Server

TCO Case Study Enterprise Mass Storage: Less Than A Penny Per GB Per Year

Flash Storage Optimizing Virtual Desktop Deployments

The Data Placement Challenge

UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Amazon Cloud Storage Options

IBM Spectrum Protect in the Cloud

Up to 4 PCI-E SSDs Four or Two Hot-Pluggable Nodes in 2U

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

THE SUMMARY. ARKSERIES - pg. 3. ULTRASERIES - pg. 5. EXTREMESERIES - pg. 9

Hortonworks Data Platform Reference Architecture

Investigation of storage options for scientific computing on Grid and Cloud facilities

Brainlab Node TM Technical Specifications

Dell Virtual Remote Desktop Reference Architecture. Technical White Paper Version 1.0

29/07/2010. Copyright 2010 Hewlett-Packard Development Company, L.P.

AirWave 7.7. Server Sizing Guide

Data Sheet FUJITSU Server PRIMERGY CX400 S2 Multi-Node Server Enclosure

How SSDs Fit in Different Data Center Applications

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment

NexentaStor Enterprise Backend for CLOUD. Marek Lubinski Marek Lubinski Sr VMware/Storage Engineer, LeaseWeb B.V.

MS Exchange Server Acceleration

MapR Enterprise Edition & Enterprise Database Edition

The Revival of Direct Attached Storage for Oracle Databases

LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE. Matt Kixmoeller, Pure Storage

Hadoop Size does Hadoop Summit 2013

Connecting Flash in Cloud Storage

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

POSIX and Object Distributed Storage Systems

Hyperscale Use Cases for Scaling Out with Flash. David Olszewski

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

SALSA Flash-Optimized Software-Defined Storage

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, IBM Corporation

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Philip Lawrence. Senior Solution Architect Sun Microsystems UK & Ireland

Ultra-Scalable Storage Provides Low Cost Virtualization Solutions

Gladstone Health & Leisure Technical Services

Storage Architectures for Big Data in the Cloud

Distribution One Server Requirements

Flash Performance in Storage Systems. Bill Moore Chief Engineer, Storage Systems Sun Microsystems

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Building Storage Service in a Private Cloud

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Low-cost BYO Mass Storage Project. James Cizek Unix Systems Manager Academic Computing and Networking Services

Transcription:

Low-cost storage @PSNC Update for TF-Storage TF-Storage meeting @Uppsala, September 22nd, 2014

Agenda Motivations data center perspective Application / use-case Hardware components: bought some, will buy more Software decisions to be taken Deployment and use Tender and surprising results

Motivations We (thought we) make the solution low-cost: - We need to differentiate disk storage equipment classes we use (vendor-made arrays vs build-your-own-storage-cluster) - Need to push purchase costs down - Need to limit TCO down, e.g. avoid paying expensive support Online capacity needs grow constantly: - Everything to be online user expect that - Even archives are often used as online repository Vendor lock-in: - Can be tolerated in some cases + mission-critical systems - production VMs, services, mail, web etc. + HPC etc. where we push for throughput/iops - Low cost storage must be free from vendor lock-in + non-critical applications

Other storage systems @PSNC (storage hell) HPC storage: - Scratch: + DDN 10k / 300TB / FC connectivity + + Lustre, 12 OSD servers - Home: + 4x NetApp E5500 (2x360TB + 2x480TB) + SSD (EF540) + GPFS + cnfs: on top of 6 servers + SSD was needed! (latency, backups, I/O waiters ) General-purpose storage: - Storage 4 VMs and any other aplications: + 2x NetApp E5400 120 drives (360TB) + 1x NetApp E5500 90 drives (360TB) + IOPS may be a problem - Archival storage + 3,5 PBs of tape storage + 10PBs in partnering institutions + 200 TB cache + HSM solution (IBM TSM for Space Management) + note people ARE TRYING TO USE IT AS A DISK!

Applications and use-cases Application type: - Sync & share - Backups / DR / Archives - Media hosting / content delivery - VMs Target aims: - user base? - internal / external both: users and systems (VMs) - university staff / students first staff, then students - capacity? - PBs - number of files / objects? - several thousands per user - hundreds of millions per storage system

Hardware components and configs Server platforms: - Storage nodes: 16x 12-drives pan-cakes / pizza-boxes WHY: + small impact of putting the node down (failure, maintenance) + flexibility to build islands if we find it feasible at the end - Options considered: - Options dropped: + Quanta s: S100-11D + SuperMicro s 36/72-disk servers + SuperMicro SuperServer + OpenCompute storage vault (FatTwin) F617H6-FTL+ (not available in small quantities) SuperMicro SuperServer F617H6-FTL+

Hardware components and configs Server platforms: - Redundant PSUs and fans some traps and tricks + by requiring dual PSUs we killed some interesting options + Quanta was not available with 2x PSUs at that time + Traps: > single PSU => 2 large failure zones!!! => loss of flexibility > an extra PSU takes space inside the server > PSUs hot swappable > FANs redundant but not hot swappable + procedure: power down the node - SuperMicro FatWin: + a common PSUs for all 4 nodes within a chassis + Fan failure impacts 4 nodes!

Hardware components and configs Server configuration CPU & RAM: + 2 CPU/s node +not the cheapest option + high CPU to disks 2 / 12, + flexibility related to CPU power + we may run apps/vms on the second CPU + real life numbers will verify our decision + architecture: typical x86_64, Xeon E5-2620v2 + 80Watts, 2,1 GHz, 6 cores/cpu, 1 + 12 cores/node, 1 core/osd (Ceph) + RAM: 64 GB + enough for Ceph/OpenStack OSDs/Storage Nodes? + spare capacity for VMs / apps / Docker images Computing/application servers configuration CPU & RAM: + 2 CPU same as storage nodes + RAM: 256 GB

Hardware components and configs Server configuration disks: - HDD: + 10x 4TB / drive per node + 2 bays / node used for SSDs (less HDD capacity per node, more IOPS) + 160 drives / initial confg (640 TB raw capacity) + Hitachi UltraStar 7k400 (SATA) + heavy duty/enterprise + there are different opinions + not that big price difference + system to run in remote location + possibly the next time we will go for cheap rubbish - SSDs: + 2x SSD: Intel DC S3700 240 GB + XFS / ext3/4 journaling + additionally: 1x DOM (Disk On Module) operating system + logs - On-board SATA/SAS controllers: + Intel C602 etc, + hardware / hybrid RAID will not be used

Hardware components and configs Network: Network interfaces: - 10Gbit for front-end and back-end production traffic - Typical 10Gbit Ethernet (not the LLE) - 1Gbit for management/booting/monitoring - 2x Chelsio T520-SO-CR / node - dual PCI-E card - There was an option for on-board 10GbitE BUT: + power consumption of copper option much higher! + in this case we had more freedom with purchase cost, less freedom with maintenance costs Network switch a part of the overall purchase - 10Gbit for both front-end and back-end

Hardware components and configs

Hardware components and configs Installation / deployment: - Initial configuration in temporary data center: + built using container technology within 0,5 year + fully covered with systems (cooling, power, etc.) - Distant location within the city: + we will be experimenting with latency impact Data center B: ============== 20x IT racks 1 MW power 400 kw cooling

Hardware components and configs Installation / deployment: - Initial configuration moved to target data center - 2015 - Additional servers put to the new building Data center C: ============== 3 level 2x HPC room 813m² 2x 208 rack Power: 16 MW, Systems: UPS+Gen. VESDA, Hi-Fog, DLC

Hardware components and configs Testing / Usage: - Deployment / configuration tools + to be determined suggestions welcome - Monitoring: nagios + ganglia

Procurement stuff Tender: - Started in July, open in September Spec: - We expect 16 nodes fitting 16U - 10 HDDs + 2SSD + 1DOM / node explicitly - CPUs (clock, cores, architecture) explicitly - Redundant power to node (dedicated OR shared PSU) => 16x1U (e.g. Quanta) OR 4x4U (SM FatWin) allowed - In addition to storage servers: 4x computing servers Winner: - SuperMicro FatWin (through Action/Vector integrators)

Procurement stuff Pricing: Higher than expected Possible causes: + configuration: > a lot of CPUs: 2x CPU, 64 GB RAM / node (+20%) > E5 CPUs (E3 supports only 32GB RAM) > enterprise HDDs, SSDs (also good ones) (+20-30%?) > wanted performance, will see if we really got it + not enough push on vendor?

Software: what and why - Initial decision: Object stores + Why: > promises fault tolerance, cheap scalability, > support Erasure Coding - Implementation: + Ceph / RADOS OR OpenStack Swift + Performance comparison tests: > Ceph: RBD: 150-200MB/s, RADOS-GW: 50-100MB/s > Swift: 50-100 MB/s > Hardware: 2x Intel Xeon E5345, 2.33 GHz, 16 GB RAM 1x SSD Intrepid drive, 60GB for objects storage - Alternatives? - GlusterFS - - I feel a bit uncomfortable without FS

Roadmap - 4Q2014 installation and benchmarking of initial config - 1Q2015: + Final decision on Software platform + Reporting to TF-Storage (February 2014) + 2nd round tender to start - 2Q2015: + Tender opening, agreements, ordering + 2nd part of equipment installation & benchmarking - Usage decision to be made - experiments needed + We will divide final configuration into separate clusters + OR put all into one cluster

2nd round options Target: - 2 full racks of storage servers - 2x 40x (10-12)x 4TB = 3200-3840 TB RAW capacity Configuration options: Option 1: go cheaper - Regular drives, no SSD? - Less CPUs, RAM Option 2: go cooler ;) - Open Compute Storage Vault + OpenRack etc. (costs?) Option 3: go low-power ;) - ARM/Atom CPUs (performance?)

Lessons - It s not that cheap! (but good thing is we have space CPU RAM for VMs) - Not enough courage / experience to really relax requirements vs hardware - We will see the real costs in some years (now we may only try to model them, which is worth to do collaboration welcome) - A lot of platforms do not support which might be deeply hidden in the docs (take care) + double PSUs / CPUs option + hot-swap disks + hot-swap fans - Tendering takes time (kind of obvious )