Addressing research data challenges at the. University of Colorado Boulder



Similar documents
Globus and the Centralized Research Data Infrastructure at CU Boulder

perfsonar Overview Jason Zurawski, ESnet Southern Partnerships for Advanced Networking November 3 rd 2015

SDN for Science Networks

NUIT Tech Talk: Trends in Research Data Mobility

Campus Network Design Science DMZ

SwitchOn Workshop São Paulo October 15-16, 2015

perfsonar Multi-Domain Monitoring Service Deployment and Support: The LHC-OPN Use Case

Science DMZs Understanding their role in high-performance data transfers

PetaLibrary Storage Service MOU

The Science DMZ: A network design pattern for data-intensive science 1

Optimizing Data Management at the Advanced Light Source with a Science DMZ

Achieving the Science DMZ

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Data Center SDN. ONF SDN Solutions Showcase Theme Demonstrations SDN SOLUTIONS SHOWCASE

Campus Bridging Made Easy via Globus Services

Introduction to Supercomputing with Janus

The Science DMZ. Eli Dart, Network Engineer Joe Metzger, Network Engineer ESnet Engineering Group. LHCOPN / LHCONE meeting. Internet2, Washington DC

Network Awareness in the Open Science Grid

The Science DMZ: A Network Design Pattern for Data-Intensive Science

Campus Research Network Overview

Fundamentals of Data Movement Hardware

Data Movement and Storage. Drew Dolgert and previous contributors

Integrating a heterogeneous and shared Linux cluster into grids

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

ScotGrid. Bolting the door. Network Based Security Mechanisms. David Crooks, Mark Mitchell on behalf of ScotGrid Glasgow

The Science DMZ: A Network Design Pa8ern for Data- Intensive Science

Campus Bridging Made Easy via Globus Services

Storage Systems: 2014 and beyond. Jason Hick! Storage Systems Group!! NERSC User Group Meeting! February 6, 2014

Remote PC Guide Series - Volume 1

LHCONE Site Connections

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Datacenter Operating Systems

Cornell University Center for Advanced Computing

Lecture 02b Cloud Computing II

A Possible Approach for Big Data Access to Support Climate Science

Introduction & Motivation

NERSC File Systems and How to Use Them

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Globus Research Data Management: Endpoint Configuration and Deployment. Steve Tuecke Vas Vasiliadis

Science DMZ Security

Best Practices for Research Data Management. October 30, 2014

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Cloud Optimize Your IT

Introduc)on & Mo)va)on

Virtualization Infrastructure at Karlsruhe

Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009

Virtualization, SDN and NFV

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

GridFTP GUI: An Easy and Efficient Way to Transfer Data in Grid

Tier3 Network Issues. Richard Carlson May 19, 2009

How To Build A Research Platform

A Reliable and Fast Data Transfer for Grid Systems Using a Dynamic Firewall Configuration

Cluster Implementation and Management; Scheduling

Lessons learned from parallel file system operation

Why Software Defined Networking (SDN)? Boyan Sotirov

Wrangler: A New Generation of Data-intensive Supercomputing. Christopher Jordan, Siva Kulasekaran, Niall Gaffney

Software Defined Networking for big-data science

Cornell University Center for Advanced Computing A Sustainable Business Model for Advanced Research Computing

SDN AND SECURITY: Why Take Over the Hosts When You Can Take Over the Network

EMERGING AND ENABLING GLOBAL, NATIONAL, AND REGIONAL NETWORK INFRASTRUCTURE TO SUPPORT RESEARCH & EDUCATION

EXINDA NETWORKS. Deployment Topologies

REDCENTRIC INFRASTRUCTURE AS A SERVICE SERVICE DEFINITION

University of Utah backbone is fully redundant with one or more 10Gb/s connecting each distribution node to a redundant core which connects to a

Networking Topology For Your System

Hadoop on the Gordon Data Intensive Cluster

Software-Defined Networking for the Data Center. Dr. Peer Hasselmeyer NEC Laboratories Europe

The BIG Data Era has. your storage! Bratislava, Slovakia, 21st March 2013

Data Management Best Practices

Conference. Smart Future Networks THE NEXT EVOLUTION OF THE INTERNET FROM INTERNET OF THINGS TO INTERNET OF EVERYTHING

IPv6 Traffic Analysis and Storage

OpenFlow: Load Balancing in enterprise networks using Floodlight Controller

D1.2 Network Load Balancing

WHITE PAPER: customize. Best Practice for NDMP Backup Veritas NetBackup. Paul Cummings. January Confidence in a connected world.

on the Road to Storage and Server Efficiency Grant Leathers

Integration of Network Performance Monitoring Data at FTS3

High Performance Computing OpenStack Options. September 22, 2015

IP Telephony Management

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

Network Virtualiza/on on Internet2. Eric Boyd Senior Director for Strategic Projects

Steroid OpenFlow Service: Seamless Network Service Delivery in Software Defined Networks

for my computation? Stefano Cozzini Which infrastructure Which infrastructure Democrito and SISSA/eLAB - Trieste

GARUDA - NKN Partner's Meet 2015 Big data networks and TCP

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Traditional v/s CONVRGD

THE REVOLUTION TOWARDS SOFTWARE- DEFINED NETWORKING

VMWARE WHITE PAPER 1

10th TF-Storage Meeting

Network Performance Optimisation and Load Balancing. Wulf Thannhaeuser

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

SAN RFP Questions and Answers

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Big Data and Cloud Computing for GHRSST

Next-Generation Networking for Science

Hosting Solutions Made Simple. Managed Services - Overview and Pricing

An Integrated CyberSecurity Approach for HEP Grids. Workshop Report.

Solution for private cloud computing

Experiences with Dynamic Circuit Creation in a Regional Network Testbed

Distributed IaaS Clouds and 100G Networking for HEP applications

Transcription:

Addressing research data challenges at the University of Colorado Boulder Thomas Hauser Director Research Computing University of Colorado Boulder thomas.hauser@colorado.edu Research Data Challenges Research data can be big in different ways [1] Volume that challenges our computing, storage and network infrastructure Lasting significance, e.g. clinical trial, environmental data Descriptive challenges, e.g. experimental setup Research Computing (RC) services to address research data challenges Partner with faculty and researchers RC-DMZ for large data transfers PetaLibrary storage infrastructure Large Scale Compute Research Data Services Collaboration between RC and the CU-Boulder Libraries Under development [1] C. Lynch, Big data: How do your data grow?, Nature, vol. 455, no. 7209, pp. 28 29, Sep. 2008. 1

Data intensive projects at CU- Boulder National Snow and Ice Data Center makes cryospheric and other data accessible and useful to researchers around the world High Energy Physics (HEP) group runs an Open Science Grid (OSG) site as part of the Large Hadron Collider (LHC) experiment BioFrontiers institute that collaborates with the Anschutz Medical Center in Denver Researchers in the Department of Computer collect and data-mine data from social networks such as Twitter to model and understand how social networks are used in emergencies Scientific Data Movement The Task: Large Data Transfer End to End Disk, Network Card, OS, Application Protocols, LAN, WAN Topologically and Physically complex (e.g. multi-domain) The Concerns Machine/OS/Protocol Tuning (e.g. TCP as the typical choice in this space) HEP is prime example that a collaboration and end to end tuning and performance debugging is necessary 2

Science DMZ Concept involves some important players: Architectural Split Enterprise vs Science use Migration of big data off of the LAN (mutually benefits the regular users too) Security and Networking Paradigm shift learn to trust things vs untrust of everything Router filters are faster than firewalls Using the Right Tools Monitoring of the network perfsonar Dynamic allocation of bandwidth DYNES, OSCARS, OpenFlow Proper data movement applications SCP = Bad, GridFTP = Good Well tuned servers (hardware, software, and protocol stack) CU-Boulder s Science DMZ: Current and future RC-DMZ: Collaboration with OIT Networking group and RC Current RC-DMZ Campus wide dedicated 10 gig Dedicated uplinks Single path single point of failures Next year (CC NIE) Upgrade border routers to be 100G and OpenFlow capability Add redundant paths and more paths between key sites using Arista switches and MLAG Dedicated perfsonar and BRO nodes 3

GridFTP and Globus online Gridftp resources (Globus Online) Four Dell PowerEdge R710s as GridFTP servers Dedicated 10Gb ethernet per node External access via science DMZ colorado#gridftp Internal access via dedicated private vlans colorado#jila, colorado#nsidc --data-interface <vlan> 4

Projects Enabled by RC-DMZ DYNES High speed access to central research data storage Globus Online Enabling different groups to have data driven high speed workflows applications HEP openscience grid node BioFrontier Institute: Moving large amounts of data between CU-Boulder and CU-Anschutz campuses NSIDIC: sharing of data JILA: Transfer to and from XSEDE resources Performance numbers Data transferred from colorado#gridftp Data transferred to colorado#gridftp 122.5 TB 21.6 TB Peak transfer rate between distinct endpoints 2.9 Gb/s Peak transfer rate to/from Janus (disk) Peak transfer rate to/from Janus (memory) 5.9 Gb/s 9.5 Gb/s 5

Storage at CU-Boulder Storage Resources 80 TB of mirrored and snapshoted project spaces on IBM Nseries (netapp) Janus Supercomputer 16,416 Westmere cores 850 TB of Lustre scratch NSF funded petalibrary project Grant provides the infrastructure Researcher pays for the medium Development of a long-term business model Global File System DDN with GPFS Capacity of several PB Performance tier Science data cloud storage Data Management Support Data management task force Members: RC, Libraries, Faculty Report to the campus leadership and faculty by October Set of recommendations to make support of Data Management a priority for the campus Research Data Services (RDS) Collaboration between CU-Boulder Libraries and RC Data.colorado.edu Basic services with existing staffing Develop a business model Build full data management services 6

Contributions Conan Moore, OIT Networking Engineer Daniel Milroy, RC System Administrator Jazcek Braden, RC Senior System Administrator Kimberly Stacey, RC Data Manager Jason Zurawski, Internet 2 Questions? 7