JASMIN and CEMS: An Overview



Similar documents
JASMIN/CEMS Services and Facilities

CEDA Storage. Dr Matt Pritchard. Centre for Environmental Data Archival (CEDA)

Parallel Processing using the LOTUS cluster

JASMIN Cloud ESGF and UV- CDAT Conference December 2014 STFC / Stephen Kill

Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS

JASMIN: the Joint Analysis System for big data. Bryan Lawrence

CMIP6 Data Management at DKRZ

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Virtualisation Cloud Computing at the RAL Tier 1. Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013

Big Data and Cloud Computing for GHRSST

Experiences and challenges in the development of the JASMIN cloud service for the environmental science community

Experience of Data Transfer to the Tier-1 from a DIRAC Perspective

Module 6: Parallel processing large data

NASA's Strategy and Activities in Server Side Analytics

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

Alternative Deployment Models for Cloud Computing in HPC Applications. Society of HPC Professionals November 9, 2011 Steve Hebert, Nimbix

Big Data Infrastructures for Processing Sentinel Data

The OPTIRAD Platform: Cloud-hosted IPython Notebooks for collaborative EO Data Analysis and Processing

SURFsara HPC Cloud Workshop

The Hartree Centre helps businesses unlock the potential of HPC

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Overview of HPC Resources at Vanderbilt

Interactive Data Visualization with Focus on Climate Research

Introduction to ACENET Accelerating Discovery with Computational Research May, 2015

SURFsara Data Services

Kriterien für ein PetaFlop System

Managing a local Galaxy Instance. Anushka Brownley / Adam Kraut BioTeam Inc.

On Demand Satellite Image Processing

Performance Analysis of a Numerical Weather Prediction Application in Microsoft Azure

PES. Batch virtualization and Cloud computing. Part 1: Batch virtualization. Batch virtualization and Cloud computing

Metadata for Data Discovery: The NERC Data Catalogue Service. Steve Donegan

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

GTC Presentation March 19, Copyright 2012 Penguin Computing, Inc. All rights reserved

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Hadoop on the Gordon Data Intensive Cluster

Steven Newhouse, Head of Technical Services

Cluster, Grid, Cloud Concepts

LSKA 2010 Survey Report Job Scheduler

Microsoft Private Cloud Fast Track Reference Architecture

An Oracle Technical White Paper November Oracle Solaris 11 Network Virtualization and Network Resource Management

SURFsara HPC Cloud Workshop

Code and Process Migration! Motivation!

Satellite'&'NASA'Data'Intro'

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

Emerging Technology for the Next Decade

Data Analytics at NERSC. Joaquin Correa NERSC Data and Analytics Services

Cloud Computing. Alex Crawford Ben Johnstone

Enabling Technologies for Distributed Computing

Scala Storage Scale-Out Clustered Storage White Paper

Big + Fast + Safe + Simple = Lowest Technical Risk

Solution for private cloud computing

Berkeley Research Computing. Town Hall Meeting Savio Overview

Cloud Computing through Virtualization and HPC technologies

Silviu Panica, Marian Neagul, Daniela Zaharie and Dana Petcu (Romania)

- An Essential Building Block for Stable and Reliable Compute Clusters

Research Technologies Data Storage for HPC

vcloud Virtual Private Cloud Fulfilling the promise of cloud computing A Resource Pool of Compute, Storage and a Host of Network Capabilities

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

Getting Started with HC Exchange Module

PRIMERGY server-based High Performance Computing solutions

Making a Smooth Transition to a Hybrid Cloud with Microsoft Cloud OS

Parallels Cloud Storage

An HPC Application Deployment Model on Azure Cloud for SMEs

HPC Wales Skills Academy Course Catalogue 2015

Neptune. A Domain Specific Language for Deploying HPC Software on Cloud Platforms. Chris Bunch Navraj Chohan Chandra Krintz Khawaja Shams

Parallel Software usage on UK National HPC Facilities : How well have applications kept up with increasingly parallel hardware?

Sun Constellation System: The Open Petascale Computing Architecture

NASA s Big Data Challenges in Climate Science

Benchmarking OPeNDAP services for modern ESM data workloads

Integrated Grid Solutions. and Greenplum

Cornell University Center for Advanced Computing

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Building Storage Service in a Private Cloud

DSS. Diskpool and cloud storage benchmarks used in IT-DSS. Data & Storage Services. Geoffray ADDE

Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise

HPC Cloud. Focus on your research. Floris Sluiter Project leader SARA

Microsoft Private Cloud

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Advancing Applications Performance With InfiniBand

Parallel Visualization of Petascale Simulation Results from GROMACS, NAMD and CP2K on IBM Blue Gene/P using VisIt Visualization Toolkit

Transcription:

JASMIN and CEMS: An Overview Alison Pamment On behalf of the JASMIN team (STFC:SCD, STFC/NERC:CEDA, NERC:NCAS, NERC:NCEO)

Setting the scene Overview of JASMIN/CEMS Funding Who runs it Purpose and usage Recent expansion

http://www.jasmin.ac.uk

JASMIN: Joint Analysis System J is for Joint Jointly delivered by STFC: CEDA (RAL Space) and SCD. Joint users (initially): Entire NERC community & Met Office Joint users (target): Industry (data users & service providers) Europe (wider environ. academia) S is for System 10m investment at RAL #1 in the world for big data analysis capability? A is for Analysis Private (Data) Cloud Compute Service Web Service Provision For Atmospheric Science Earth Observation Environmental Genomics and more. Opportunities JASMIN is a collaboration platform! within NERC (who are the main investor) with UKSA (& the S.A. Catapult via CEMS*) with EPSRC (joined up national e- infrastructure) with industry (as users, cloud providers, SMEs) (*CEMS - :the facility for Climate and Environmental Monitoring from Space)

Funded through JASMIN CEMS

STFC Centre for Environmental Data Archival Exist: to support environmental science, further environmental data archival practices, and develop and deploy new technologies to enhance access to data. Curation and Facilitation Curation: Four Data Centres British Atmospheric Data Centre NERC Earth Observation Data Centre IPCC Data Distribution Centre UK Solar System Data Centre (BADC, NEODC, IPCC-DDC, UKSSDC) Over 23,000 registered users! + active research in curation practices! Facilitation: Users by discipline Data Management for scientists (planning, formats, ingestion, vocabularies, MIP support, ground segment advice etc) Data Acquisition (archiving 3 rd party data for community use) JASMIN Support (Group Workspaces, JASMIN Analysis Platform, Cloud Services, Parallelisation)

CEDA Evolution Early 90's 2008 2014

Hosted by STFC Scientific Computing Department Computing Expertise across length scales from processes within atoms to environmental modelling Applications development and support, Compute and data facilities and services Research and Training Numerical Analysis High Performance Computing Emerald GPU cluster for Oxford, UCL, Southampton, Bristol. SCARF HPC for RAL Hartree: Blue Joule bluegene HPC Hartree: Blue Wonder idataplex HPC JASMIN: NERC super data cluster Data Services STFC: Facility Archives (ISIS, Diamond) LHC: UK Hub (Tier 1 archive) BBSRC: Institutes data archive MRC: Data Support Service NERC: CEDA backup and JASMIN elastic tape Close working partnership with industry NW - VE C CF M S

Who is allowed access? Primary communities and special cases addressing environmental challenges, delivering new science, driving UK innovation, economic growth and societal wellbeing http://www.jasmin.ac.uk/jasmin-users/who-can-use-jasmin/

Allocation of resources Application and decision-making Any project/group can apply to have a: Group Workspace(s) Dedicated VM(s) Virtual Organisation Access is decided based on research priorities Collaborations involving *.ac.uk are given higher priority Scale is also important small projects can be more diverse if using little resource Moving to a consortium model for each domain: Atmos Sci, Hydrology, BioInformatics, etc.

CEDA data storage & services Curated data archive Archive management services Archive access services (HTTP, FTP, Helpdesk,...) Data intensive scientific computing Global / regional datasets & models High spatial, temporal resolution Private cloud JASMIN functions Flexible access to high-volume & complex data for climate & earth observation communities Online workspaces Services for sharing & collaboration

Processing big data: the issues Parallel processing in the Environmental Sciences has historically focussed on highly-parallel models Data analysis was typically run sequentially because: It was a smaller problem It didn t have parallel resources available The software/scientists were not equipped to work in parallel Now we generate enormous datasets (e.g. UPSCALE - 300 Tb): Processing big data requires a parallel approach Platforms, tools, and programmers are becoming better equipped

JASMIN Use cases Processing large volume EO datasets to produce: Essential Climate Variables Long term global climate-quality datasets Data validation & intercomparisons Evaluation of models relying on the required datasets (EO datasets, in situ and simulations) being in the same place

JASMIN Use cases User access to 5 th Coupled Model Intercomparison Project (CMIP5) Large volumes of data from best climate models Greater throughput required Large model analysis facility Workspaces for scientific users. Climate modellers need 100s of Tb of disk space, with high-speed connectivity UPSCALE project We would never have been able to store, nor analyse that volume of data, without the existence of the [JASMIN] service UPSCALE spokescientist 250 Tb in 1 year PRACE supercomputing facility in Germany (HERMIT) Being shipped to RAL at present To be analysed by Met Office as soon as available Deployment of VMs running custom scientific software, colocated with data Outputs migrated to long term archive (BADC)

JASMIN Use Case: UPSCALE Picture courtesy of P-L Vidale & R. Schiemann, NCAS) Ocean temperatures (in colour going from blue=cold to violet=warm) are shown in the background, while clouds (B/W scale) and precipitation (colour) are shown in the foreground. Over land, snow cover is shown in white. 25 km resolution model run The largest ever PRACE computational project, led by the UK, dependent on CEDA-BADC to provide the data links and data analysis environment!

JASMIN Phase 1 hardware JASMIN is configured as a storage and analysis environment. Two types of compute: virtual/cloud environment batch compute Both sets of compute connected to 5 PB of parallel fast disk

Phase 1 JASMIN Now (Phase 2) hardware 5+7+1 = 13 PB Disk Batch Compute Hosted Compute Phase 1 Phase 2 From 228 cores to 3328 cores Phase 1 and Phase 2 co-exist (will be joined as part of Phase 3) Phase 2

Network: internal 12 x 40Gb Mellanox switches 1 connection to each bottom-ofrack switches Complete redundant mesh RAL site has 40Gb connection to JANET/internet using same 40Gb connections! 204 x 40Gb cables provides bandwidth of over 1 Terbyte / sec internal to JASMIN2 Phase 3 connects JASMIN1 to JASMIN2 via yellow 56Gbit cables

JASMIN/CEMS system architecture

VM Types Login [1] Transfer [1..*] Analysis [1..*] Project-specific [*] jasmin-login1.ceda.ac.uk; acts as a gateway to other JASMIN nodes; only one; no functionality. jasmin-xfer1.ceda.ac.uk; for copying data in/out; currently SCP & RSYNC; GridFTP; readwrite to GWS. jasmin-sci[12].ceda.ac.uk; for general scientific analysis; common software build; access to GWSs and archive. *.ceda.ac.uk requested by specific projects/users; ROOT access for trusted partners; read-write access to GWS. NOTE: CEMS equivalents also exist of these VM types but they are fundamentally the same.

General purpose scientific analysis VMs Scalable create multiple VMs as required SSH access via JASMIN login node Supported set of common software (including Iris and IDL) Repeatable build process developed in a managed way Process for requesting software installs/upgrades Read-access to BADC & NEODC archives (for authorised users) Write-access to cache disk Read/write-access to GWS (for authorised users) Home directories Currently we have: jasmin-sci1, jasmin-sci2 and cemssci1

Dedicated (project-specific) VMs Projects can request VMs for their own specific use: Logins limited to collaborators only Automated registration process (authorised by manager) Specification designed with manager (limited O/S) Can mount relevant Group Workspaces Root access allowed Can install packages from JASMIN Analysis Platform and/or your own requirements

Example Dedicated VMs upscale-mo1.ceda.ac.uk upscale-nerc1.ceda.ac.uk precis1.ceda.ac.uk jasmin-name1.ceda.ac.uk, jasmin-name2.ceda.ac.uk nemo1.ceda.ac.uk um-conv1.ceda.ac.uk leicester_proc02.cems.rl.ac.uk -> leicester-proc12.cems.rl.ac.uk edinburgh-proc01.cems.rl.ac.uk -> edinburgh-proc09.cems.rl.ac.uk and edinburghproc12.cems.rl.ac.uk ecmwf-svc1.ceda.ac.uk mms1.cems.rl.ac.uk mohc-test.ceda.ac.uk pml-proc01.cems.rl.ac.uk -> pml-proc04.cems.rl.ac.uk ukcp-batch1.ceda.ac.uk ukmo-01.cems.rl.ac.uk -> ukmo-03.cems.rl.ac.uk

Group Workspaces (GWS) For projects that want to: Share a LARGE network-accessible disk. Allow access from a number of institutions. Pull/push data from/to external sites to a common location. Move data from MONSooN/ARCHER to share with others. Use to store third-party data sets required by project. Process/analyse the data using VMs/LOTUS. Process/analyse the data in conjunction with other archived or group workspace datasets. Work up data ready for the BADC/NEODC archives. From 2 Tbytes to 380 Tbytes in volume (so far). Share data through HTTP access (restricted or public).

Example Group Workspaces JASMIN - Future Weather (NCAS) JASMIN - GLOCON JASMIN - GRIP JASMIN - HIGEM JASMIN - JULES Big Data JASMIN - MO-GC2 (Met Office Global Coupled Model Evaluation runs) JASMIN - NAME JASMIN - NCAS-CMS JASMIN - NEMO JASMIN - PRECIS JASMIN - RAINGAIN JASMIN - TIFF JASMIN - UKCA JASMIN - UPSCALE CEMS - Cloud-ECV (CCI/NCEO cloud products) CEMS - ESACCI Phase 1 Ocean Colour CEMS - ESACCI SST

The JASMIN Analysis Platform a re-usable, re-deployable bundle of common tools These tools can be installed anywhere inside JASMIN with one or two commands!

What JAP Provides Standard Analysis Tools: NetCDF4, HDF5, Grib Operators: NCO, CDO Python Stack Numpy, SciPy, Matplotlib IRIS, cf-python, cdat_lite Ipython GDAL, GEOS NCAR Graphics, NCL R, octave Parallelisation and Workflow: Python MPI bindings Jug (simple Python task scheduling) IPython notebook IPython-parallel JASMIN Community Inter-comparison Suite

Community Intercomparison Suite (CIS) (CIS = Component of JAP) Time-series Line plots Scatter plots Dataset Format AERONET Text MODIS HDF Global plots Curtain plots CALIOP CloudSAT HDF HDF AMSRE HDF TRMM HDF CCI aerosol & cloud NetCDF Overlay plots Histograms SEVIRI Flight campaign data NetCDF RAF Models NetCDF

CIS Co-location cis col <variable>:<source file> <sampling file>:colocator=lin -o <new file> cis plot <variable>:<new file> <variable>:<sampling file> --type comparativescatter \ --logx --xlabel 'Observations AOT 675nm' --xmin 1.e-3 --xmax 10 \ --logy --ylabel 'Model AOT 670nm' --ymin 1.e-3 --ymax 10 Model gives global output every 3 hours for a full month Source Collocation Observations are daytime site measurements, every 15 min for a full month Sampling

LOTUS compute cluster JASMIN Scientific Analysis Servers are limited in resources LOTUS is bigger! High performance computing facility Cluster of both physical and virtual machines High speed access to Panasas storage BADC and NEODC Archive Group workspaces Users home directories Scratch area (read/write) /tmp http://www.ceda.ac.uk/help/users-guide/lotus/

LOTUS Hardware Model Processor Cores Memory 194 x Viglen HX525T2i Intel Xeon E5-2650 v2 Ivy Bridge 16 128GB 14 x Viglen HX545T4i Intel Xeon E5-2650 v2 Ivy Bridge 16 512GB 6 x Dell R620 Intel Xeon E5-2660 Sandy Bridge 16 128GB 8 x Dell R610 Intel Xeon X5690 Westmere 12 48GB 3 x Dell R610 Intel Xeon X5675 Westmere 12 96GB 1 x Dell R815 AMD Opteron 48 256GB 226 physical hosts 3556 cores Intel/AMD processors 17 large memory hosts Mix of generations/specs

LOTUS Configuration jasmin-sci1/2.ceda.ac.uk Queues lotus.jc.rl.ac.uk Head node project VMs lotus lotus2 lotus-smp atsr-repro CN CN CN CN CN CN CN CN CN Compute nodes (physical hosts or virtual machines)

Job Submission Jobs are submitted using the LSF scheduler Resources are allocated as they become available Fair share of resources between users

JASMIN links

JASMIN Dedicated Network Connections Link RAL - ARCHER RAL MONSooN (Met Office) RAL - KNMI/WUR RAL - University of Leeds Notes 2 x 1Gbit/s lightpath to data component of the UK National Supercomputing Service in Edinburgh. 1 Gbit/s lightpath to UK Met Office supercomputer in Exeter. Dynamic lightpath via SURFnet to KNMI (Dutch Met Office) and Wageningen University (NL) Lightpath to enable efficient connection of JASMIN-North satellite node.

High Performance Transfer Server As well as the standard xfer VMs there is a dedicated server set up to achieve fast transfers for special cases. The High Performance Data transfer service is a machine with optimised network connectivity to permit rapid transfer of data in and out of JASMIN. The service allows transfers into the Group Workspaces using rsync, scp and bbcp http://www.jasmin.ac.uk/services/high-performance-data-transfer/

Elastic Tape A tape storage system is currently in its testing phase Projects can apply for volumes of tape storage to optimise how they store data between highperformance and tape: If it is not being used for significant period of time then it is cost-effective to push data onto tape

JASMIN Phase 2 Cloud Infrastructure firewall External Cloud Providers Users: Individuals and Organisations Public IP Science Analysis Science Compute VM Analysis Cluster 0 VM VM 0 JASMIN Managed Cloud JASMIN Unmanaged Cloud Web App VM GWS MyOrganisation-JVO Science Data transfers Database Analysis Science Science VM VM Analysis Analysis 0 VM VM 0 GWS CEMS-JVO-2 Science Analysis Science Science VM Analysis Analysis 0 VM VM 0 Data xfer VM Science Analysis Science Science VM Analysis Analysis 0 VM VM 0 Login VM Science Analysis Science Science VM Analysis Analysis 0 VM VM 0 Data xfer VM GWS Project1-JVO GWS MyProject-JVO GWS CEMS-JVO-1 lotus.jc.rl.ac.uk lotus.jc.rl.ac.uk lotus.jc.rl.ac.uk lotus.jc.rl.ac.uk Batch processing cluster Data Centre Archive /badc /neodc Key: General-purpose resources JVO - JASMIN Virtual Organisations Data centre resources

Finding out more JASMIN documentation pages http://www.jasmin.ac.uk Technical details (paper explaining JASMIN Phase 1 set up) doi:10.1109/bigdata.2013.6691556 JASMIN Phase 2 launch presentations http://jasmin.ac.uk/what-is-jasmin/jasmin-launch-event/ CEMS and the Satellite Applications Catapult https://sa.catapult.org.uk/cems