Data Storage and Data Transfer. Dan Hitchcock Acting Associate Director, Advanced Scientific Computing Research

Similar documents
Briefing for NFAC February 5, Sharlene Weatherwax Associate Director Office of Biological and Environmental Research U.S. Department of Energy

Data Requirements from NERSC Requirements Reviews

Network Middleware Solutions

Analisi di un servizio SRM: StoRM

Measurement of BeStMan Scalability

The Evolution of Research and Education Networks and their Essential Role in Modern Science

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

Science DMZs Understanding their role in high-performance data transfers

ESnet Support for WAN Data Movement

Deploying distributed network monitoring mesh

How To Write A Data Mining Program On A Large Data Set

Optimizing Data Management at the Advanced Light Source with a Science DMZ

Next-Generation Networking for Science

Exploration of adaptive network transfer for 100 Gbps networks Climate100: Scaling the Earth System Grid to 100Gbps Network

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

Mass Storage at GridKa

The GRID and the Linux Farm at the RCF

GridFTP GUI: An Easy and Efficient Way to Transfer Data in Grid

TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research

Scalla/xrootd. Andrew Hanushevsky, SLAC. SLAC National Accelerator Laboratory Stanford University 19-May-09. ANL Tier3(g,w) Meeting

Improving Scientific Outcomes at the APS with a Science DMZ

irods at CC-IN2P3: managing petabytes of data

Migrating NASA Archives to Disk: Challenges and Opportunities. NASA Langley Research Center Chris Harris June 2, 2015

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Solution for private cloud computing

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

The dcache Storage Element

Addressing research data challenges at the. University of Colorado Boulder

Installing and Configuring Windows Server Module Overview 14/05/2013. Lesson 1: Planning Windows Server 2008 Installation.

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

BIG data big problems big opportunities Rudolf Dimper Head of Technical Infrastructure Division ESRF

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

Accelerating Experimental Elementary Particle Physics with the Gordon Supercomputer. Frank Würthwein Rick Wagner August 5th, 2013

Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting.

Intel HPC Distribution for Apache Hadoop* Software including Intel Enterprise Edition for Lustre* Software. SC13, November, 2013

Data Management Plan Requirements

Resume. Wenjing. Date of birth: June 11th, 1982 Nationality: Chinese Phone number: Cell phone:

The Science DMZ. Eli Dart, Network Engineer Joe Metzger, Network Engineer ESnet Engineering Group. LHCOPN / LHCONE meeting. Internet2, Washington DC

Workflow Requirements (Dec. 12, 2006)

Network & HEP Computing in China. Gongxing SUN CJK Workshop & CFI

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

BMC Recovery Manager for Databases: Benchmark Study Performed at Sun Laboratories

The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets

In-Situ Bitmaps Generation and Efficient Data Analysis based on Bitmaps. Yu Su, Yi Wang, Gagan Agrawal The Ohio State University

Next Generation Networking for Science Program Update. NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014

DOE National Laboratories Improvement Council s White Paper on Management of Major Research Facility Construction Projects

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

PLUMgrid Toolbox: Tools to Install, Operate and Monitor Your Virtual Network Infrastructure

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

A Physics Approach to Big Data. Adam Kocoloski, PhD CTO Cloudant

Exascale Computing Project (ECP) Update

Big Data Analytics. for the Exploitation of the CERN Accelerator Complex. Antonio Romero Marín

dcache, list of topics

Mission Need Statement for the Next Generation High Performance Production Computing System Project (NERSC-8)

POSIX and Object Distributed Storage Systems

dcache, Software for Big Data

The Ultra-scale Visualization Climate Data Analysis Tools (UV-CDAT): A Vision for Large-Scale Climate Data

Globus for Data Management

McAfee Global Threat Intelligence File Reputation Service. Best Practices Guide for McAfee VirusScan Enterprise Software

Cluster, Grid, Cloud Concepts

Optimizing Large Data Transfers over 100Gbps Wide Area Networks

Modern Financial Markets and Data Intensive Science: Leveraging 35 Years of Federal Research

ASCAC Facilities Statement

Patrick Fuhrmann. The DESY Storage Cloud

globus online Cloud-based services for (reproducible) science Ian Foster Computation Institute University of Chicago and Argonne National Laboratory

Software Scalability Issues in Large Clusters

Next Generation Tier 1 Storage

DataMover: Robust Terabyte-Scale Multi-file Replication over Wide-Area Networks

CISCO WIDE AREA APPLICATION SERVICES (WAAS) OPTIMIZATIONS FOR EMC AVAMAR

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

Distributed Computing for CEPC. YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep.

Scala Storage Scale-Out Clustered Storage White Paper

Cloud Computing for Research Roger Barga Cloud Computing Futures, Microsoft Research

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

Data Center Performance Insurance

Data storage services at CC-IN2P3

Windows Server 2008 R2 Essentials

Workload Characterization and Analysis of Storage and Bandwidth Needs of LEAD Workspace

SCI Briefing: A Review of the New Hitachi Unified Storage and Hitachi NAS Platform 4000 Series. Silverton Consulting, Inc.

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Big data management with IBM General Parallel File System

Architecture & Experience

HyperQ Remote Office White Paper

Florida Site Report. US CMS Tier-2 Facilities Workshop. April 7, Bockjoo Kim University of Florida

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

Gfarm: Present Status and Future Evolution

Summer Student Project Report

Basic Energy Sciences Network Requirements Review

PACE Predictive Analytics Center of San Diego Supercomputer Center, UCSD. Natasha Balac, Ph.D.

High Performance Storage System. Overview

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities

ANI Network Testbed Update

Dynamic Extension of a Virtualized Cluster by using Cloud Resources CHEP 2012

Storage Resource Managers: Middleware Components for Grid Storage

Network Awareness in the Open Science Grid

Biomile Network - Design and Implementation

Accelerating and Simplifying Apache

EMC Unified Storage for Microsoft SQL Server 2008

Transcription:

Data Storage and Data Transfer Dan Hitchcock Acting Associate Director, Advanced Scientific Computing Research

Office of Science Organization 2

Data to Enable Scientific Discovery DOE SC Investments in Data Driven Science support SC and DOE missions: Advanced Scientific Computing Research: Leadership Computers and NERSC Basic Energy Sciences: Light Sources, Neutron Sources, Biological and Environmental Research: Climate, Proteomics, Fusion Energy Sciences: Experiments and Simulation High Energy Physics: LHC, Nuclear Physics: TJNAF, RHIC, Data and Science closely coupled End to End perspective 3

Requirements Gathering ESnet and NERSC separately survey each SC program office every 2-3 years (two per year) to determine program s computing and networking requirements Workshops jointly organized by ASCR and Program Office Program Office chooses representatives Findings from the NERSC-BES Requirements workshop, February, 2010 Several large BES experiments coming online that will have large data requirements. For example, the Coherent X-ray Imaging instrument at SLAC is expected to generate 1.7 PB of data in 2012 and 10.4 PB in 2015 Other than these large experiments, all BES projects presented at the workshop had data needs that are currently being met and are projected to being met in the future based on our current data growth projections. 4

BES ESnet Requirements Workshop Highlights Instrument upgrades are driving dramatic changes in science process for BES facility users Current paradigm is to take data home on portable USB hard drives Current paradigm tractable because data sets are small, and end sites have poor network performance characteristics Instrument and facility upgrades are increasing data set sizes by up to 100x in the next few years (individual beamlines at some facilities, significant portion of NSLS-II at BNL in 2014) New paradigm: data sets too large, not portable, must be analyzed at facility and/or transferred to home institution Remote control will become more important as experiments become increasingly automated Science Data Network (SDN) has a role to play here (we have seen this coming) Need to vet operational models with facilities BES facilities and users are going to need a lot of help from ESnet, site networks, and tools providers ESnet already planning pilot projects (ALS, NSLS) 5

Other BES-ESnet Workshop Highlights Multiple scientists expressed a need for remote visualization this will increase as data sets increase in size (data too large to transfer home analyze data at remote institution remote visualization need) Significant knowledge gap in science community regarding appropriate tools, proper setup for data transfer, etc http://fasterdata.es.net is helping, but there is more to do ESnet working with facilities (network tuning talk at workshop generated significant interest from attendees) Cybersecurity is a major obstacle SSH/SCP are used for data transfer (installed by default and approved by cybersecurity) and are the worst performers for data transfer Appropriate tools (e.g. GridFTP) need to be deployable in order to be useful if cybersecurity prevents their deployment, they do not exist as useful scientific tools Portals and other data services (e.g. NERSC Science Gateways) were of interest to several participants 6

ESnet Backbone Connecting Facilities Router node 10G link 7

Data Transfer Tools The Center for Enabling Distributed Petascale Science (CEDPS) provides secure, high-bandwidth fire-andforget file transfer services CEDPS project team: ANL, LBNL, USC/ISI, FNAL, UW-Madison Serving scientists and resource providers from: LBNL / NERSC, ORNL, ANL, GFDL, TeraGrid, Europe, Australia Data Source Data moved securely; Errors handled transparently CEDPS Globus.org service Researcher initiates transfer via simple command Data Destination Researcher informed of status/completion Accomplishments Transferred without human intervention, automatically recovering from many network and systems failures: STAR: 300GB of data between BNL and NERSC/PDSF SCEC: 600,000 4MB files from TeraGrid resources to OSG CEDPS: 20TB among 12 sites (including DOE labs, TeraGrid, Australia) in under 15 hours Powered by: 8

Uniformity of Interface Compatibility of SRMs SRM/DP M Unixbased Disk Pools client/user applications SRM/ dcache dcache SRM/ CASTOR CASTOR CCLRC RAL SRM/ StoRM GPFS SRM/ BeStMan Unixbased Disk Pools Implementation Bestman/NFS 24 dcache 15 Bestman/HDFS 4 Bestman/XROOTD 4 Bestman/GPFS 2 Bestman/Lustre 2 Bestman/Reddnet 1 # of Sites Uniform data movement between very different configurations and disk cache sizes. Unix-based Disk Pools include: NFS, GPFS, HDFS, Lustre, XROOTD, Rednet Data transferred from/to sites in Europe, North America, and Asia. 51 Institutions use inter-operating SRM implementations. Clients can access different mass storage and file systems through a uniform SRM interface. 9

FastBit - Efficient Search Technology for Data Driven Science Problem Quickly find records satisfying a set of user-specified conditions in a large, complex data set Example: High-energy physics data find a few thousand events based on conditions on energy level and number of particles in billions of collision events, with hundreds of variables Solution Developed new indexing techniques and a new compression method for the indexes, achieved 10-100 fold speedup compared with existing methods Efficient software implementation: available open source from http://sdm.lbl.gov/fastbit/ (1000s of downloads), received a R&D 100 Award Impact WAH compression research begins Laser Wakefield Particle Accelerator data analysis: FastBit acts as an efficient back-end for a visual analytics system, providing information for identifying and tracking particles Combustion data analysis: FastBit identifies ignition kernels based on user specified conditions and tracks evolution of the regions Testimonial FastBit is at least 10x, in many situations 100x, faster than current commercial database technologies Senior Software Engineer, Yahoo! Inc WAH published Binning & Multi-level Encoding for Additional Speedup WAH Patent STAR Combustion Grid Collector Won Best Paper @ ISC Query Driven Vis Theoretical Analysis Published BioSolveIT Use Begins Open Source Software Released Spatial Join Yahoo! Use (Fusion) Begins 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 10