ESnet Support for WAN Data Movement
|
|
|
- Elfrieda Shana Ward
- 10 years ago
- Views:
Transcription
1 ESnet Support for WAN Data Movement Eli Dart, Network Engineer ESnet Science Engagement Group Joint Facilities User Forum on Data Intensive Computing Oakland, CA June 16, 2014
2 Outline ESnet overview Support for data movement specific engineering concerns Effect of packet loss Science DMZ model Data movement rates and volumes Current network ESnet5 Network enhancements What are others doing?
3 ESnet by the Numbers High-speed national network, optimized for DOE science missions: high performance networking facility for DOE/SC connecting 40 labs, plants and facilities with > networks $32.6M in FY14, 42FTE older than commercial Internet, growing twice as fast $62M ARRA grant for G upgrade: transition to new era of optical networking fiber assets + access to spectrum shared with Internet2 world s first G network at continental scale first connections were to DOE supercomputer centers Culture of urgency: 4 awards in past 3 years R&D in FY13 5 out of 5 for customer satisfaction in last review
4 ESnet Is A Science Network First We specifically engineer the network to support data-intensive science High-speed, high-capability connections to sites, facilities, and other networks G to ANL, LBNL, FNAL, NERSC, ORNL G to BNL, LANL, LLNL, SLAC coming soon G connections to major peering exchanges Test and measurement instrumentation to ensure performance Engagement with science experiments, facilities, and sites to help get the maximum benefit from the infrastructure ESnet s capabilities are derived from the requirements of the science 6/16/14 4
5 Science Data Transferred Each Month by the Energy Sciences Network 15.5 Petabytes (March 2013) Bytes Transferred Jan-90 Feb-94 Apr-98 Jun-02 Aug-06 Oct- Month
6 17 16 Science Data Transferred Each Month by the Energy Sciences Network Petabytes (March 2013) Bytes Transferred Jan-90 Feb-94 Apr-98 Jun-02 Aug-06 Oct- Month
7 ESnet is not the Commercial Internet
8 Elephant Flows Place Great Demands on Networks è Physical pipe that leaks water at rate of.0046% by volume. è Result % of water transferred. è Network pipe that drops packets at rate of.0046%. è Result % of data transferred, slowly, at <<5% opemal speed. essentially fixed determined by speed of light With proper engineering, we can minimize packet loss. Assumptions: Gbps TCP flow, 80ms RTT. See Eli Dart, Lauren Rotman, Brian Tierney, Mary Hester, and Jason Zurawski. The Science DMZ: A Network Design Pattern for Data-Intensive Science. In Proceedings of the IEEE/ACM Annual SuperComputing Conference (SC13), Denver CO, 2013.
9 A small amount of packet loss makes a huge difference in TCP performance Local (LAN) Metro Area With loss, high performance beyond metro distances is essentially impossible Regional Continental International Measured (TCP Reno) Measured (HTCP) Theoretical (TCP Reno) Measured (no loss) 6/16/14
10 Working With TCP In Practice Far easier to support TCP than to fix TCP People have been trying to fix TCP for years limited success Like it or not we re stuck with TCP in the general case Pragmatically speaking, we must accommodate TCP Sufficient bandwidth to avoid congestion Zero packet loss Verifiable infrastructure Networks are complex Must be able to locate problems quickly Small footprint is a huge win small number of devices so that problem isolation is tractable HPC facilities must also do this the other end must also do this
11 Putting A Solution Together Effective support for TCP-based data transfer Design for correct, consistent, high-performance operation Design for ease of troubleshooting Easy adoption is critical Large laboratories and universities have extensive IT deployments Drastic change is prohibitively difficult Cybersecurity defensible without compromising performance Borrow ideas from traditional network security Traditional DMZ separate enclave at network perimeter ( Demilitarized Zone ) Specific location for external-facing services Clean separation from internal network Do the same thing for science Science DMZ 6/16/14 11
12 The Science DMZ Design Pattern Dedicated Systems for Data Transfer Data Transfer Node Network Science DMZ Performance perfsonar High performance Architecture Dedicated network Testing & location for high-speed Measurement Configured specifically for data transfer Proper tools data resources Appropriate security Easy to deploy - no need to redesign the whole network Enables fault isolation Verify correct operation Widely deployed in ESnet and other networks, as well as sites and facilities 6/16/14 12
13 Science DMZ Implications Many Science DMZs have been or are being deployed DOE laboratories and facilities Supercomputer centers (e.g. XSEDE) Many tens of universities NSF CC-NIE and CC*IIE programs (over $40M and still counting) LHC infrastructure International Australia (~$50M program) Europe Brazil This is how many of your users will move data to/from your systems Data transfer nodes Globus High-speed connections from Science DMZ to research networks 6/16/14 13
14 Sample Data Transfer Rates This table available at:" " 6/16/14 J. Zurawski 14
15 User Requirements What have users asked for? 2Gbps, 5Gbps, Gbps per experiment Light sources Microscopy 1PB/week data replication cosmological simulations Petascale data replication climate science To from a capacity perspective, the network part is done 1PB/week = 14Gbps in round numbers Light source workflows are bursty lots of them fit in G Effective use of the network is now gated on two things End system resources, particularly storage (but tools too) Squeaky-clean network path (zero packet loss) 6/16/14 15
16 Current ESnet5 Capabilities G connections to DOE/ASCR computing facilities (ANL, NERSC, ORNL) G connections to additional labs (BNL, FNAL, LBNL, LLNL) G connections to more sites soon (LANL, SLAC) G connections to major peering exchanges MANLAN (New York) OmniPOP (Chicago) Starlight (Chicago) WIX (Washington, DC) Pacific Wave coming soon (US West Coast) G connections to major science networks (CENIC, Internet2) Performance Assurance using perfsonar (we watch for performance problems and we fix them) Multiple G Data Transfer Nodes for end-to-end testing of disk to disk transfers Current G core runs at 5% to 20% load, with bursts to 40% 6/16/14 16
17 ESnet5 Map SEAT PNNL SNLL LLNL 1 1 ANL CLEV PPPL GFDL PU Physics BNL SDSC 1 LASV ALBU CHAT 1
18 EEX Future Expansion of the ESnet network into Europe Drivers: Improve the quality and quantity of transit to Europe for: All currently supported ESnet programs All US LHC science Proposed plan: extend ESnet5 geography Hubs at London, Amsterdam and 2 at CERN Links G Ring in Europe G T/A - WASH-CERN, AOFA-LOND, NEWY-LOND 40G T/A - BOST-AMS Peering connections to GEANT, Netherlight & London Open
19 One Possible EEX Design 6/16/14 19
20 Implications of EEX for ESnet constituents To first order, the ESnet G core would be extended into Europe Significant upgrade to existing capability Current capacity is showing its limitations 3xGE from New York to Amsterdam 3xGE from Washington to Frankfurt EEX plans include diverse GE with 40G failover Significant improvement in operational model Current connectivity is shared between many players Different players at each end Very difficult to reach resolution on problems (months!) EEX would be run by ESnet as part of the ESnet Facility 6/16/14 20
21 Future Upgrades ESnet is already planning for core bandwidth upgrades Anticipation of LHC Run 2 in early 2015 Increased demand from HPC facilities, light sources, etc Remember the loss plot we have to stay ahead Upgrades will occur where they are needed SF Bay Area Chicago / New York / Washington (major LHC footprint) If you know of significant needs, please tell us ASAP If you are planning to support something, ESnet should be planning too If we don t know about it, we can t help you Some procurements take months It is ESnet s intent to stay ahead of demand we don t like being the bottleneck 6/16/14 21
22 Other Players NSF is currently soliciting proposals for IRNC (international connectivity) This funds much of the current international science connectivity for the US Expect increased bandwidth to Asia, Latin America in the coming years GEANT (Pan-European network) has G core now Many European NRENs (national networks) are G also We get to them via GEANT France Germany Netherlands UK EEX will provide robust connectivity Many, many science DMZs folks at the other end are really upping their game All of this comes to a HPC facility at its DTNs 6/16/14 22
23 Questions? Thanks! Eli Dart - [email protected]
Campus Network Design Science DMZ
Campus Network Design Science DMZ Dale Smith Network Startup Resource Center [email protected] The information in this document comes largely from work done by ESnet, the USA Energy Sciences Network see
Deploying distributed network monitoring mesh
Deploying distributed network monitoring mesh for LHC Tier-1 and Tier-2 sites Phil DeMar, Maxim Grigoriev Fermilab Joe Metzger, Brian Tierney ESnet Martin Swany University of Delaware Jeff Boote, Eric
Software Defined Networking for big-data science
Software Defined Networking for big-data science Eric Pouyoul Chin Guok Inder Monga (presenting) SRS presentation November 15 th, Supercomputing 2012 Acknowledgements Many folks at ESnet who helped with
The Science DMZ and the CIO: Data Intensive Science and the Enterprise
The Science DMZ and the CIO: Data Intensive Science and the Enterprise Eli Dart & Jason Zurawski ESnet Science Engagement Lawrence Berkeley National Laboratory RMCMOA Workshop @ Westnet Conference Tempe,
Software Defined Networking for big-data science
Software Defined Networking for big-data science Eric Pouyoul Chin Guok Inder Monga (presenting) TERENA Network Architects meeting, Copenhagen November 21 st, 2012 ESnet: World s Leading Science Network
Fundamentals of Data Movement Hardware
Fundamentals of Data Movement Hardware Jason Zurawski ESnet Science Engagement [email protected] CC-NIE PI Workshop April 30 th 2014 With contributions from S. Balasubramanian, G. Bell, E. Dart, M. Hester,
The Science DMZ: Introduction & Architecture
The Science DMZ: Introduction & Architecture Eli Dart, Lauren Rotman, Brian Tierney, Jason Zurawski,, Eric Pouyoul ESnet Science Engagement Operating Innovative Networks (OIN) Berkeley, CA Februrary 27
! Strategic!Plan! FY20143FY2023! ESnet! !!!! March!1,!2013!
StrategicPlan FY20143FY2023 ESnet March1,2013 ThisworkwassupportedbytheDirector,OfficeofScience,OfficeofAdvanced ScientificComputingResearchoftheU.S.DepartmentofEnergyunderContractNo.DECAC02C05CH11231.
ANI Network Testbed Update
ANI Network Testbed Update Brian Tierney, ESnet, Joint Techs, Columbus OH, July, 2010 ANI: Advanced Network Initiative Project Start Date: September, 2009 Funded by ARRA for 3 years Designed, built, and
Next-Generation Networking for Science
Next-Generation Networking for Science ASCAC Presentation March 23, 2011 Program Managers Richard Carlson Thomas Ndousse Presentation
Networking for Science + Software-Defined Networking (SDN): Hype vs. Hope
Networking for Science + Software-Defined Networking (SDN): Hype vs. Hope Inder Monga Chief Technologist and Area Lead HEAnet Conference 2013 Athlone, Ireland Outline Energy Sciences Network Networking
The Science DMZ. Eli Dart, Network Engineer Joe Metzger, Network Engineer ESnet Engineering Group. LHCOPN / LHCONE meeting. Internet2, Washington DC
The Science DMZ Eli Dart, Network Engineer Joe Metzger, Network Engineer ESnet Engineering Group LHCOPN / LHCONE meeting Internet2, Washington DC June 13 2011 Overview Science Needs Data Deluge, new science
LHCONE Site Connections
LHCONE Site Connections Michael O Connor [email protected] ESnet Network Engineering Asia Tier Center Forum on Networking Daejeon, South Korea September 23, 2015 Outline Introduction ESnet LHCONE Traffic Volumes
SC14 Remote I/O Pipeline Processing Demonstrtion
OpenFabrics Software User Group Workshop SC4 Remote I/O Pipeline Processing Demonstrtion #OFSUserGroup Dardo D Kleiner / CIPS, Corp. / [email protected] Linden Mercer / PSU / [email protected] Srinath Jayasundera
EVALUATING NETWORK BUFFER SIZE REQUIREMENTS
EVALUATING NETWORK BUFFER SIZE REQUIREMENTS for Very Large Data Transfers Michael Smitasin Lawrence Berkeley National Laboratory (LBNL) Brian Tierney Energy Sciences Network (ESnet) [ 2 ] Example Workflow
The Science DMZ: A network design pattern for data-intensive science 1
Scientific Programming 22 (2014) 173 185 173 DOI 10.3233/SPR-140382 IOS Press The Science DMZ: A network design pattern for data-intensive science 1 Eli Dart a,,laurenrotman a, Brian Tierney a, Mary Hester
Tier3 Network Issues. Richard Carlson May 19, 2009 [email protected]
Tier3 Network Issues Richard Carlson May 19, 2009 [email protected] Internet2 overview Member organization with a national backbone infrastructure Campus & Regional network members National and International
Integration of Network Performance Monitoring Data at FTS3
Integration of Network Performance Monitoring Data at FTS3 July-August 2013 Author: Rocío Rama Ballesteros Supervisor(s): Michail Salichos Alejandro Álvarez CERN openlab Summer Student Report 2013 Project
perfsonar: End-to-End Network Performance Verification
perfsonar: End-to-End Network Performance Verification Toby Wong Sr. Network Analyst, BCNET Ian Gable Technical Manager, Canada Overview 1. IntroducGons 2. Problem Statement/Example Scenario 3. Why perfsonar?
ENOS: a Network Opera/ng System for ESnet Testbed
ENOS: a Network Opera/ng System for ESnet Testbed Eric Pouyoul ([email protected]) Technology Exchange Cleveland, Ohio, September 2015 Is ESnet really developing Yet Another Network Opera:ng System (YANOS)?
Software-Defined Networks (SDN): Bridging the application-network divide
Software-Defined Networks (SDN): Bridging the application-network divide Inder Monga Chief Technologist and Area Lead, Engineering, Research and Software development CHEP 2013 Shared historical moment
Globus for Data Management
Globus for Data Management Computation Institute Rachana Ananthakrishnan ([email protected]) Data Management Challenges Transfers often take longer than expected based on available network capacities
Schedule 2e. Additional terms for Ethernet services
1. SERVICE DESCRIPTION The Interoute Ethernet Clear and Ethernet Flex Services offer Circuits carrying telecommunications traffic between Network Termination Points on the Interoute Ethernet Network and/
A High-Performance Storage and Ultra-High-Speed File Transfer Solution
A High-Performance Storage and Ultra-High-Speed File Transfer Solution Storage Platforms with Aspera Abstract A growing number of organizations in media and entertainment, life sciences, high-performance
Briefing for NFAC February 5, 2014. Sharlene Weatherwax Associate Director Office of Biological and Environmental Research U.S. Department of Energy
Briefing for NFAC February 5, 2014 Sharlene Weatherwax Associate Director Office of Biological and Environmental Research U.S. Department of Energy Outline of the Presentation DOE and the National Labs--background
Network monitoring with perfsonar. Duncan Rand Imperial College London
Network monitoring with perfsonar Duncan Rand Imperial College London A little history: Gridmon Cast your minds back to GridPP16 at QMUL about 6 years ago; we saw a not too dissimilar network monitoring
Best Practices in Legal IT. How to share data and protect critical assets across the WAN
Best Practices in Legal IT How to share data and protect critical assets across the WAN Agenda Requirements for Data Center outsourcing Timothy Titus, Director of Managed Network Services Overcoming WAN
Network performance monitoring Insight into perfsonar
Network performance monitoring Insight into perfsonar Szymon Trocha, Poznań Supercomputing and Networking Center E-infrastructure Autumn Workshops, Chisinau, Moldova 9 September 2014 Agenda! Network performance
addition to upgrading connectivity between the PoPs to 100Gbps, GPN is pursuing additional collocation space in Kansas City and is in the pilot stage
Cyberinfrastructure Plan GPN GPN connects research and education networks in 6 states to the Internet2 backbone and to ESnet. GPN provides redundant connectivity for several other affiliated campuses and
PORTrockIT. Spectrum Protect : faster WAN replication and backups with PORTrockIT
1 PORTrockIT 2 Executive summary IBM Spectrum Protect, previously known as IBM Tivoli Storage Manager or TSM, is the cornerstone of many large companies data protection strategies, offering a wide range
CENIC and CALREN An Update
CENIC and CALREN An Update John Silvester Chair of the CENIC Board Vice-Provost for Scholarly Technology, University of Southern California 5 o Workshop RNP2 Gramado, Brasil May 13 th, 2004 CENIC History
Exascale Computing Project (ECP) Update
Exascale Computing Project (ECP) Update Presented to ASCAC Paul Messina Project Director Stephen Lee Deputy Director Washington, D.C. April 4, 2016 ECP is in the process of making the transition from an
Four Ways High-Speed Data Transfer Can Transform Oil and Gas WHITE PAPER
Transform Oil and Gas WHITE PAPER TABLE OF CONTENTS Overview Four Ways to Accelerate the Acquisition of Remote Sensing Data Maximize HPC Utilization Simplify and Optimize Data Distribution Improve Business
The Next Generation Internet Program. Mari Maeda ITO
The Next Generation Internet Program Mari Maeda ITO Today s Internet Traffic Makeup other ICMP HTTP Today s Internet Flow Size Distribution Comparison of 97 to 99 20000 Number of transfers 18000 16000
OnTimeDetect: Offline and Online Network Anomaly Notification Tool
OnTimeDetect: Offline and Online Network Anomaly Notification Tool Prasad Calyam, Ph.D. [email protected] Other Team Members: Jialu Pu, Weiping Mandrawa Network Tools Tutorial Session, Internet2 Spring Member
Advanced High. Architecture. www.service-now.com
Advanced High Availability Architecture www.service-now.com Advanced High Availability Architecture WHITE PAPER Advanced High Availability Architecture Overview This document describes the ServiceNow Advanced
Software Defined Networking for big-data science
Software Defined Networking for big-data science Architectural models from campus to the WAN Inder Monga, Eric Pouyoul, Chin Guok Energy Sciences Network Lawrence Berkeley Lab Berkeley, California {inder,
GÉANT MD-VPN Service Description High performance IP-VPN connectivity to support European Research and Education
GÉANT MD-VPN Service Description High performance IP-VPN connectivity to support European Research and Education Issue Date: 12 June 2015 1 1 GÉANT MD-VPN service description The
Los Nettos and Translight/Pacific Wave. SC 08 November 16-20, 2008
Los Nettos and Translight/Pacific Wave SC 08 November 16-20, 2008 20 years of networking The Los Nettos Regional Network was founded in 1988 by leading internet visionaries, Los Nettos continues to provide
Intelsat Update. David Ball Regional Vice President Asia-Pacific. New 2006 Template - 1
Intelsat Update David Ball Regional Vice President Asia-Pacific New 2006 Template - 1 Intelsat is Continuing to grow and perform Robust fleet investment program underway 11 satellites currently in development
NSF IRNC Program International Deployment and Experimental Efforts with SDN in 2013
NSF IRNC Program International Deployment and Experimental Efforts with SDN in 2013 The National Science Foundation (NSF) International Research Network Connections (IRNC) program is enabling the development
CISCO WIDE AREA APPLICATION SERVICES (WAAS) OPTIMIZATIONS FOR EMC AVAMAR
PERFORMANCE BRIEF CISCO WIDE AREA APPLICATION SERVICES (WAAS) OPTIMIZATIONS FOR EMC AVAMAR INTRODUCTION Enterprise organizations face numerous challenges when delivering applications and protecting critical
LEVEL 3 DATA CENTER SERVICES IN SAN FRANCISCO HOUSE POWER-INTENSIVE APPLICATIONS AND SUPPORT BUSINESS CONTINUITY
LEVEL 3 DATA CENTER SERVICES IN SAN FRANCISCO HOUSE POWER-INTENSIVE APPLICATIONS AND SUPPORT BUSINESS CONTINUITY THE SCALE TO SUPPORT GLOBAL ENTERPRISE IT THE SCALE TO SUPPORT FLEXIBLE ENTERPRISE IT It
Making a Case for Including WAN Optimization in your Global SharePoint Deployment
Making a Case for Including WAN Optimization in your Global SharePoint Deployment Written by: Mauro Cardarelli Mauro Cardarelli is co-author of "Essential SharePoint 2007 -Delivering High Impact Collaboration"
DeltaV System Cyber-Security
January 2013 Page 1 This paper describes the system philosophy and guidelines for keeping your DeltaV System secure from Cyber attacks. www.deltav.com January 2013 Page 2 Table of Contents Introduction...
Achieving Performance Isolation with Lightweight Co-Kernels
Achieving Performance Isolation with Lightweight Co-Kernels Jiannan Ouyang, Brian Kocoloski, John Lange The Prognostic Lab @ University of Pittsburgh Kevin Pedretti Sandia National Laboratories HPDC 2015
4 High-speed Transmission and Interoperability
4 High-speed Transmission and Interoperability Technology 4-1 Transport Protocols for Fast Long-Distance Networks: Comparison of Their Performances in JGN KUMAZOE Kazumi, KOUYAMA Katsushi, HORI Yoshiaki,
Brochure More information from http://www.researchandmarkets.com/reports/3050461/
Brochure More information from http://www.researchandmarkets.com/reports/3050461/ Global Cloud Computing Market Service (Software, Infrastructure, Platform)- Strategy, Overview, Size, Share, Global Trends,
Data Center Solutions
Data Center Solutions New Data Center Challenges Require New Solutions Data Center Architecture. Inside and Out. Data centers are mission-critical facilities. A silo-based approach to designing, deploying
Executive Summary. Internet Traffic and Capacity
Executive Summary New applications and services have combined with strong broadband subscriber growth worldwide to fuel demand for Internet capacity. Despite predictions from some quarters that Internet
