IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez



Similar documents
Building an efficient and inexpensive PACS system. OsiriX - dcm4chee - JPEG2000

Scientific Computing Data Management Visions

Scala Storage Scale-Out Clustered Storage White Paper

Data Movement and Storage. Drew Dolgert and previous contributors

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect


How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda

Sawmill Log Analyzer Best Practices!! Page 1 of 6. Sawmill Log Analyzer Best Practices

Cluster Implementation and Management; Scheduling

PRIMERGY server-based High Performance Computing solutions

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Scalable Cloud Computing Solutions for Next Generation Sequencing Data

Hardware Configuration Guide

Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012

Sun Constellation System: The Open Petascale Computing Architecture

Terminal Server Software and Hardware Requirements. Terminal Server. Software and Hardware Requirements. Datacolor Match Pigment Datacolor Tools

Next Generation Operating Systems

Certification Document bluechip STORAGEline R54300s NAS-Server 03/06/2014. bluechip STORAGEline R54300s NAS-Server system

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

Hyper-V over SMB: Remote File Storage Support in Windows Server 2012 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Windows Server Performance Monitoring

Lustre SMB Gateway. Integrating Lustre with Windows

Flexible Scalable Hardware independent. Solutions for Long Term Archiving

EMC ISILON AND ELEMENTAL SERVER

Enabling High performance Big Data platform with RDMA

SMB Direct for SQL Server and Private Cloud

HPC on AWS. Hiroshi Kobayashi, Dev./Lab. IT System HGST Japan, Ltd. Jun 3, 2015

Hadoop on the Gordon Data Intensive Cluster

EMC SOLUTION FOR SPLUNK

(Scale Out NAS System)

Test Report Newtech Supremacy II NAS 05/31/2012. Newtech Supremacy II NAS storage system

Mellanox Accelerated Storage Solutions

Big Fast Data Hadoop acceleration with Flash. June 2013

Post-production Video Editing Solution Guide with Microsoft SMB 3 File Serving AssuredSAN 4000

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

White paper. QNAP Turbo NAS with SSD Cache

DSS. Diskpool and cloud storage benchmarks used in IT-DSS. Data & Storage Services. Geoffray ADDE

Mellanox Academy Online Training (E-learning)

E4 UNIFIED STORAGE powered by Syneto

IMPLEMENTING GREEN IT

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

Backup architectures in the modern data center. Author: Edmond van As Competa IT b.v.

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Solid State Storage in Massive Data Environments Erik Eyberg

Architecting High-Speed Data Streaming Systems. Sujit Basu

Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, IBM Corporation

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Big Data and Cloud Computing for GHRSST

General Pipeline System Setup Information

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

What is the real cost of Commercial Cloud provisioning? Thursday, 20 June 13 Lukasz Kreczko - DICE 1

I3: Maximizing Packet Capture Performance. Andrew Brown

Storage Solutions for Bioinformatics

CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012

POSIX and Object Distributed Storage Systems

ioscale: The Holy Grail for Hyperscale

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Low-cost BYO Mass Storage Project. James Cizek Unix Systems Manager Academic Computing and Networking Services

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

How To Back Up A Computer To A Backup On A Hard Drive On A Microsoft Macbook (Or Ipad) With A Backup From A Flash Drive To A Flash Memory (Or A Flash) On A Flash (Or Macbook) On

Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

PARALLELS CLOUD STORAGE

Architecting a High Performance Storage System

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

CyberStore WSS. Multi Award Winning. Broadberry. CyberStore WSS. Windows Storage Server 2012 Appliances. Powering these organisations

Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture

New Storage System Solutions

Low-cost

EMC DATA DOMAIN OPERATING SYSTEM

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

EMC DATA DOMAIN OPERATING SYSTEM

ntier Verde Simply Affordable File Storage

UNIFIED HYBRID STORAGE. Performance, Availability and Scale for Any SAN and NAS Workload in Your Environment

Samsung Portable SSD. Branded Product Marketing Team, Memory Business

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Current Status of FEFS for the K computer

Minimum Hardware Configurations for EMC Documentum Archive Services for SAP Practical Sizing Guide

How to Choose your Red Hat Enterprise Linux Filesystem

Building Clusters for Gromacs and other HPC applications

IBM System x SAP HANA

An Affordable Commodity Network Attached Storage Solution for Biological Research Environments.

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Accelerating Server Storage Performance on Lenovo ThinkServer

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

NetApp High-Performance Computing Solution for Lustre: Solution Guide

Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging

Deploying and managing a Visualization Onera

Building low cost disk storage with Ceph and OpenStack Swift

Transcription:

IT of SPIM Data Storage and Compression EMBO Course - August 27th Jeff Oegema, Peter Steinbach, Oscar Gonzalez 1

Talk Outline Introduction and the IT Team SPIM Data Flow Capture, Compression, and the Data Volume Problem Transfer, Network and Storage Infrastructure Planning for SPIM 2

People Involved - IT Staff Peter Steinbach (steinbac@mpi-cbg.de) Scientific / HPC Software Development Data Streaming Library Compression and HPC Algorithm Development Ian Henry (henry@mpi-cbg.de) Scientific Computing Leader Scientific and Project Coordination Collaboration Management Oscar Gonzalez (ogonzale@mpi-cbg.de) HPC Administrator Cluster Interaction and Queuing High-Performance Storage / Lustre Network Benchmarking and Performance Tuning Matt Boes (boes@mpi-cbg.de) Infrastructure Team Leader Network Design and Development Fileserver Design and Development Jeff Oegema (joegema@mpi-cbg.de) IT Coordinator Overall Project Coordination External Collaboration Management 3

SPIM Dataflow Camera Camera Capture Computer(s) Capture Computer(s) or Dedicated Local Storage Storage System Data Transfer Cluster for Processing Zeiss Windows Software for Processing Data Transfer Longer Term Storage Review Archival Presentation Capture Processing Usage and Archiving 4

SPIM Dataflow Camera Camera Capture Computer(s) Capture Computer(s) or Dedicated Local Storage Storage System Data Transfer Cluster for Processing Zeiss Windows Software for Processing Data Transfer Longer Term Storage Review Archival Presentation Capture Processing Usage and Archiving 5

Zeiss Lightsheet Z.1 2 x CamLink SPIM System Lightsheet Z.1 Capture Computer Capture Stream 150 MB / sec Storage Computer 8 / 32 TB Storage Drain to network / external storage 10Gb Ethernet 6

The Potential Deluge Developmental SPIM - Camera Potential - 138 TB / Day Estimated CERN data production / day 82 TB 82 TB - Estimated CERN Data Production / Day 50 GB Single Lightsheet Z1 Capture - 1 day 13 TB 13 TB Confocal Microscope Capture - 1 day 5 50 50 GB - Confocal - 1 day 7

The Potential Deluge - Multiple Lightsheet Z.1 150 MB / sec x 4 = 600 MB / sec ~52 TB / day ~364 TB / week Our entire online disk storage (fileserver) currently is 700 TB 8

The Potential Deluge - Future Tech 800 MB / sec x 2 cameras = 1.6 GB / sec ~138 TB / day Almost a PB per week CERN produces 30 PB of data annually from LHC experiments* * http://home.web.cern.ch/about/computing 9

SPIM Dataflow Camera Camera Capture Computer(s) Capture Computer(s) or Dedicated Local Storage Storage System Data Transfer Cluster for Processing Zeiss Windows Software for Processing Data Transfer Longer Term Storage Review Archival Presentation Capture Processing Usage and Archiving 10

Transfer Volumes and Times Data Volume / Time 1 Gbit 10 Gbit 150 MB / sec 9 GB / minute 540 GB / hour ~13 TB / day 1.5 sec.15 sec 90 sec 9 sec 1.5 hours 9 minutes 1.5 days 3.6 hours This assumes approximately theoretical maximum line speed - which never happens. Typically we see 60%. 11

First : Get the Data Off Camera Camera Capture Host 10 Gb/s 1.25 GB/s Remote Storage Storage network mounted drives (ex. SMB) secure network file transfer (scp/sftp/rsync) unencrypted file transfer (ex. ftp) pro simple secure fast con OS dependent encryption may slow transfer insecure 12

Operating Systems & Networking Percentage of Line Speed Extensive Network Streaming Tests Win7, Windows Server 2008 R2 10 Gbit/s fiber network same hardware No disk i/o involved 100 95 90 85 80 Windows Linux 13

Networks are ç a shared resource 14

Network File Transfer is necessary ( capture host becomes unusable / full ) protocols are important to keep in mind network is a shared http://erindriver.travellerspoint.com/148/ resource 15

Second : Bottlenecks again Camera Camera Capture Host Storage Spinning disk based storage Large Volume Comparatively cheap SSD based storage Fast Small Volume Expensive SPIM Operation Capture Transfer Capture Transfer Capture Transfer 16

Second : Get it Small Camera Camera Capture Host Storage cropping (only keep what you need) fusion + deconvolution (n stacks become 1) compression Reduce data volume before any network or disk 17

Compression : Demonstration Please zip a SPIM dataset of your choice How long did it take? How small is the compressed data? 18

Compression : Fast Capture Camera Camera Storage 19

Compression : Noise? SPIM data contains a lot of noise 20

Compression : Denoised 21

Compression : Denoised Original 22 Denoised

Compression : Sqeazy pipeline standard compression algorithms fast soon to be open-sourced currently: 3x lossless compression 10x lossy compression 23 Loic Royer initiated by: Martin Weigert

SPIM Dataflow Camera Camera Capture Computer(s) Capture Computer(s) or Dedicated Local Storage Storage System Data Transfer Cluster for Processing Zeiss Windows Software for Processing Data Transfer Longer Term Storage Review Archival Presentation Capture Processing Usage and Archiving 24

HPC for SPIM Lots of image data 25

HPC for SPIM Lots of image data CPU intensive 26

HPC for SPIM Lots of image data CPU intensive High memory footprint 27

HPC for SPIM Lots of image data CPU intensive High memory footprint High I/O 28

HPC for SPIM Lots of image data CPU intensive High memory footprint High I/O GPUs are promising 29

HPC for SPIM Lots of image data CPU intensive High memory footprint High I/O GPUs are promising 30

31

Cluster Architecture Head node Job management Cluster monitoring Clients Export node 10 GbE Worker nodes 40x: * 12 cores * 128 GB RAM * 1 TB HDD 4x: * 12 cores * 128 GB RAM * 1 TB HDD * GPU Disk server 200 TB net space (RAID 6) 10 GB/s InfiniBand non-blocking 400-600 MB/s (up to saturation) 32

MPI-CBG Cluster Storage 10 GbE 10 GbE 10 GbE Export Node Export Node Export Node Lustre Storage 200 TB Total Space Capable of 10 GB/s of up to 600 MB/s streams Also serves the cluster processing nodes Temporary storage Processing Nodes 33

Resource Usage The cluster was made available on Feb 2013 Total number of jobs done: 6,852,661 Average throughput: 462 jobs/h CPU time consumed: 151y 46d 10h 59m 12s Average CPU time: 11m 35s 34

Lessons Learned Cluster design is very important - think before you buy I/O is critical to move data in and out of the cluster I/O is VERY critical to access data from the cluster Storage requirements are huge, both inside and outside the cluster GPU resources might be useful but you need enough to make it practical 35

Workstations If/when a cluster is not an option, check what your WS can do. Example Data PC 12 cores 128 GB 4x 2TB (RAID 5) Pros: Rather cheap Fine for small datasets Convenient for data visualisation Cons: Limited computing resources Limited storage capacity and bandwidth 36

SPIM Dataflow Camera Camera Capture Computer(s) Capture Computer(s) or Dedicated Local Storage Storage System Data Transfer Cluster for Processing Zeiss Windows Software for Processing Data Transfer Longer Term Storage Review Archival Presentation Capture Processing Usage and Archiving 37

Current Infrastructure 10 Gpbs 8 Gpbs Fileserver Visible network-wide High capacity (900 TB) Robust (no SPOF) Backed up daily Tape library Second copies Long term storage 38

Future Plans... Scale-out NAS Highly scalable Very high capacity (10 PB) Very high bandwidth Very robust (no SPOF) 39

Taking it Home - External Drives 4TB transfer USB 3.0 (600 MB / sec) Protocol 1.85 hours USB 2.0 (60 MB / sec) 18.5 hours WD Black (130.4 MB / sec) Hitachi Deskstar (102.95 / MB sec) Samsung SSD - 1 TB (550 MB /sec) Drive Speed 8.52 hours 10.79 hours 2.02 hours The limitation is the slower of the two 40

Bottlenecks at Each Stage Capture Network Processing Storage Single Drive RAID Array 1 Gbit 10 Gbit Single Machine HPC Fileserver New Fileserver or Data System Bottlenecks can be addressed but the pipeline can t be made infinitely wide Experiment Design and Data Management become extremely important Compression can help but the issue remains 41

IT Planning for SPIM (or things to think about before I capture ) What is the practical output of your SPIM setup? How long are you planning on capturing at a time? What processing do you need to do on your data? How fast do you need to complete the processing? What is the data you will consider primary data for publication? How will you present your data to the world or turn it into movies or results more easily shared? 42

Discussion and Questions 43