CMS Computing Model: Notes for a discussion with Super-B

Similar documents
The CMS analysis chain in a distributed environment

A multi-dimensional view on information retrieval of CMS data

The Compact Muon Solenoid Experiment. Conference Report. Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

No file left behind - monitoring transfer latencies in PhEDEx

Scalable stochastic tracing of distributed data management events

CMS Monte Carlo production in the WLCG computing Grid

John Kennedy LMU München DESY HH seminar 18/06/2007

ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC

Improvement Options for LHC Mass Storage and Data Management

Tier-1 Services for Tier-2 Regional Centres


ATLAS Software and Computing Week April 4-8, 2011 General News

The Data Quality Monitoring Software for the CMS experiment at the LHC

The CMS Tier0 goes Cloud and Grid for LHC Run 2. Dirk Hufnagel (FNAL) for CMS Computing

LHC Databases on the Grid: Achievements and Open Issues * A.V. Vaniachine. Argonne National Laboratory 9700 S Cass Ave, Argonne, IL, 60439, USA

Grid Computing in Aachen

Client/Server Grid applications to manage complex workflows

EDG Project: Database Management Services

Status and Evolution of ATLAS Workload Management System PanDA

Big Data and Storage Management at the Large Hadron Collider

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

Computing at the HL-LHC

Data Management Plan (DMP) for Particle Physics Experiments prepared for the 2015 Consolidated Grants Round. Detailed Version

Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil

ATLAS grid computing model and usage

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan

Evolution of Database Replication Technologies for WLCG

File Transfer Software and Service SC3

Data Management. Issues

PhEDEx. Physics Experiment Data Export. Chia-Ming, Kuo National Central University, Taiwan ( a site admin/user rather than developer)

Werkzeuge zur verteilten Analyse im ATLAS-Experiment

An objective comparison test of workload management systems

A Physics Approach to Big Data. Adam Kocoloski, PhD CTO Cloudant

High Availability Databases based on Oracle 10g RAC on Linux

Relational databases for conditions data and event selection in ATLAS

BNL Contribution to ATLAS

Software, Computing and Analysis Models at CDF and D0

WM Technical Evolution Group Report

CMS: Challenges in Advanced Computing Techniques (Big Data, Data reduction, Data Analytics)

CMS Dashboard of Grid Activity

Running a typical ROOT HEP analysis on Hadoop/MapReduce. Stefano Alberto Russo Michele Pinamonti Marina Cobal

ATLAS GridKa T1/T2 Status

Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007

PoS(EGICF12-EMITC2)110

Online data handling with Lustre at the CMS experiment

ATLAS job monitoring in the Dashboard Framework

NT1: An example for future EISCAT_3D data centre and archiving?

MAD2: A Scalable High-Throughput Exact Deduplication Approach for Network Backup Services

ANALYSIS FUNCTIONAL AND STRESS TESTING

BaBar and ROOT data storage. Peter Elmer BaBar Princeton University ROOT Oct. 2002

Distributed Database Access in the LHC Computing Grid with CORAL

LHCb activities at PIC

Energy Efficient Storage - Multi- Tier Strategies For Retaining Data

Computing Model for SuperBelle

Data Storage Accounting and Verification at LHC experiments

The glite File Transfer Service

The glite File Transfer Service

Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary

Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC

Storage Virtualization. Andreas Joachim Peters CERN IT-DSS

DSS. The Data Storage Services (DSS) Strategy at CERN. Jakub T. Moscicki. (Input from J. Iven, M. Lamanna A. Pace, A. Peters and A.

Report on WorkLoad Management activities

Tier0 plans and security and backup policy proposals

Realization of Inventory Databases and Object-Relational Mapping for the Common Information Model

HappyFace for CMS Tier-1 local job monitoring

Werkzeuge zur verteilen Analyse im ATLAS-Experiment

HowThe Most Effective Way To Get Rich Fast!

The Future of Information

The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets

Tools and Services for the Long Term Preservation and Access of Digital Archives

Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF)

Rapid web development using AJAX and Python

Intro to AWS: Storage Services

A New Tool for Monitoring CMS Tier 3 LHC Data Analysis Centers

Big Data Needs High Energy Physics especially the LHC. Richard P Mount SLAC National Accelerator Laboratory June 27, 2013

Database Services for CERN

Objectivity Data Migration

Understanding Enterprise NAS

CMS Software Deployment on OSG

Top-Quark Studies at CMS

OLTP Meets Bigdata, Challenges, Options, and Future Saibabu Devabhaktuni

Data Grids. Lidan Wang April 5, 2007

SUPPORT FOR CMS EXPERIMENT AT TIER1 CENTER IN GERMANY

Service Challenge Tests of the LCG Grid

LCG POOL, Distributed Database Deployment and Oracle

The LHCb Software and Computing NSS/IEEE workshop Ph. Charpentier, CERN

A SIMULATION STUDY FOR T0/T1 DATA REPLICATION AND PRODUCTION ACTIVITIES. Iosif C. Legrand *

Hitachi Storage Solution for Cloud Computing

A TECHNICAL WHITE PAPER ATTUNITY VISIBILITY

Protect Data... in the Cloud

Software-defined Storage Architecture for Analytics Computing

Scala Storage Scale-Out Clustered Storage White Paper

The LHC Open Network Environment Kars Ohrenberg DESY Computing Seminar Hamburg,

The LCG Distributed Database Infrastructure

Accelerating Experimental Elementary Particle Physics with the Gordon Supercomputer. Frank Würthwein Rick Wagner August 5th, 2013

Scalable Multi-Node Event Logging System for Ba Bar

IBM Smart Business Storage Cloud

Grid Data Management. Raj Kettimuthu

Database Monitoring Requirements. Salvatore Di Guida (CERN) On behalf of the CMS DB group

Transcription:

CMS Computing Model: Notes for a discussion with Super-B Claudio Grandi [ CMS Tier-1 sites coordinator - INFN-Bologna ] Daniele Bonacorsi [ CMS Facilities Ops coordinator - University of Bologna ] 1

Outline The CMS distributed computing system from guiding principles to architectural design Workflows (and actors) in CMS computing Computing Tiers A glance to Data Management (DM) and Workload Management (WM) components The realization of the CMS Computing Model in a Grid-enabled world Implementation of production-level systems on the Grid Data Distribution, MonteCarlo (MC) production, Data Analysis Computing challenges - Worldwide LCG challenges, and experiment-specific challenges 2

The CMS Computing Model The CMS computing system relies on a distributed infrastructure of Grid resources, services and toolkits distributed system to cope with computing requirements for storage, processing and analysis of data provided by LHC experiments building blocks provided by Worldwide LHC Computing Grid [WLCG] - CMS builds application layers able to interface with few - at most - different Grid flavors (LCGn, EGEE, OSG, NorduGrid,...) Several steps: CMS Computing Model document (CERN-LHCC-2004-035) CMS C-TDR released (CERN-LHCC-2005-023) - in preparation for the first year of LHC running - not blueprint, but baseline targets (+ development strategies) - hierarchy of computing tiers using WLCG services and tools - focus on Tiers role, functionality and responsibility Now partially old already? - ECoM group - To consider Evolution of Computing Model from Startup to Steady State (ECoM) - To digest and include the lessons learned 3

A Tiered architecture T0 (CERN): Accepts data from DAQ Prompt reconstruction Data archive and distribution to T1 s CAF (CERN Analysis Facility for CMS): Access to full RAW datasets Focused on latency-critical activities (detector diagnostics, trigger performance services, derivation of Al/Ca constants) Provide some CMS central services (e.g. store conditions and calibrations) >50 T2 7 T1 centers and >50 T2 centers (and a growing nb of T3 s ) See next slide 4

Towards a mesh model T1 5

T1/T2 roles CMS T1 functions Scheduled data-reprocessing and dataintensive analysis tasks: - later-pass reco, AOD extraction, skimming, Data archiving (real+mc): - custody of raw+reco & subsequently produced data Disk storage management: - fast cache to MSS, buffer for data transfer, Data distribution: - data serving to Tier-2 s for analysis Data Analysis - 5-10% of all processing is RAW data analysis, via special role CMS T2 functions 50% user data analysis - Data processing for calib/align tasks and detector studies - Proficient data access via CMS+ WLCG services 50% MC event prod - both fast and detailed Import skimmed datasets from T1s Export MC data to T1s 6

A data-driven baseline Baseline system with minimal functionality for first physics Keep it simple! Use Grid services as much as possible + add CMS-specific services if/where needed Optimize for the common case - for read access (most data is write-once, read-many) - for organized bulk processing, but without limiting single user Decouple parts of the system - Minimize job dependencies + site-local information remain site-local T0-T1 s activities driven by data placement Data is partitioned by the experiment as a whole All data is placed at a site through explicit CMS policy - do not move around in response to job submission Leads to very structured usage of Tier-0 and Tier-1 - T0 and T1 are resources for the whole esperiment - activities and functionality are largely predictable since nearly entirely specified - i.e. organized mass processing and custodial storage Unpredictable computing essentially restricted to data analysis at T2s T2s are the place where more flexible, user driven activities can occur Very significant computing resources and good data access are needed 7

Data organization CMS expects to produce large amounts of data (evts) O(PB)/yr Event data are in files average file size is kept reasonably large ( GB) - avoid scaling issues with storage systems and catalogues when dealing with too many small files - file merging also implemented and widely used in production activities O(10^6) files/year Files are grouped in fileblocks group files in blocks (1-10 TB) for bulk data management reasons - exists as a result of either MC production or data movement 10^3 fileblocks/yr Fileblocks are grouped in datasets Datasets are large (100 TB) or small (0.1 TB) - Dataset definition is physics-driven (size as well) 8

Data types Data types/volumes as input parameters for the model RAW Triggered evts recorded by DAQ - 2 copies: 1 at T0 and 1 spread over T1s RECO Reconstructed objects with their associated hits - Detailed output of the detector reco: track candidates, hits, cells for calib - 1 copy spread over T1s (together with associated RAW) AOD (Analysis Object Data) Main analysis format: objects + minimal hit info - Summary of the reco evt for common analyses: particles id, jets, - Whole set copied to each T1, large fraction copied to T2 TAG Fast selection info - Relevant info for fast evt selection in AOD Plus MC in ~ N:1 (N>=1) ratio with data 9

10

CMS Data Management Provide tools to discover, access and transfer event data in a distributed computing environment Track and replicate data with a granularity of file blocks Minimize the load on catalogues The logical components: DBS (Dataset Bookkeeping system) - DBS provides the means to define, discover and use CMS event data DLS (Dataset Location Service) - DLS provides the means to locate replicas of data in the distributed system local file catalogue solutions - A trivial file catalogue as a baseline solution PhEDEx (Physics Experiment Data Export) - integration with most recent EGEE transfer services 11

CMS MC production system Current MC production system in production since 2006 Overcome previous inefficiencies + introduce new capabilities - less man-power consuming, better handling of Grid-sites unreliability, better use of resources, automatic retrials, better error report/handling Flexible and automated architecture - ProdManager (PM) (+ the policy piece) - manage the assignment of requests to 1+ ProdAgents and tracks the global completion of the task - ProdAgent (PA) - Job creation, submission and tracking, management of merges, failures, resubmissions, - It works with a set of resources (e.g. a Grid, a Site) 12

CMS MC production operations MC prod operations strategy changed: 2006-07 : 6 regional teams managed by a central manager 2008-09 : 1 central team (6 people) managing submissions in defined T1-regions 13

CMS Data Placement system Physics Experiment Data Export (PhEDEx) Large-scale reliable and scalable dataset/fileblock replication multi-hop routing following a transfer topology (T0-T1-T2-T3 s), data pre-stage from tape, monitoring, bookkeeping, priorities and policy,... In production since almost 2004 In the hall-of-fame in terms of mature and high-quality production services for LHC experiments Managing transfers of several TB/day PhEDEx integration with EGEE services glite File Transfer Service (FTS) - PhEDEx takes care of reliable, scalable CMS dataset replication (and more ) - FTS takes care of reliable point-to-point transfers of files 14

Passive! Monitoring! Mgmt at the! block level! Active! Find lowest-cost! replica! File!level operations" 15

100 TB/day 16

3 (!) orders of magnitude of improvement since 2004 CMS today has >600 (!) commissioned links 17

CMS Analysis resources *3456-"-'"3'/78'"-'9#$:;$<#=';3'4' >5;?4556'="-@$"?A@#='B;55#BC;3';:' B;<9AC3>':4B"5"C#-'!"#$%.' /012'' /*+'!"#"$%&'()"$*+,'-%#"'',".%$%/"&0' %11234/"5'%4%&0,),'' 6%1)&)7",' *+,-'!"#$%&'!"#$%&'!"#$%&'!"#$%&' ()"$*8'92:.374;'<%1)&)7",'%$"' -%&6'5"#2/"5'/2',):3&%724' -%&6'3,"$'%4%&0,),='>$):%$0''?",23$1"'62$'@4%&0,),'!"#$%('!"#$%('!"#$%('!"#$%('!"#$%('!"#$%)'!"#$%)' ()"$*A'92:.374;'<%1)&)7",'%$"' "47$"&0'124/$2&&"5'B0'/-"'.$2#)5)4;')4,7/3724'3,"5'62$' %4%&0,),''!"#$%)'!"#$%)'!"#$%)'!"#$%)'!"#$%)' 18

CMS Distributed Analysis on the Grid CMS Remote Analysis Builder (CRAB) Tool for job preparation, submission and monitoring Satisfies the needs of CMS users Better resource control usage via the CRAB Analysis Server!"#$%&'()*(+,"-.$!"#$%2+26.$ %/+00$1+2+$34"56(2.$ @=AB$ A(('..$2"$7C()+0$1+2+$!"#$% &'()'(% @=AB$ A(('..$2"$8.'4$ 1+2+$$ *+,-.+%% /0'(123% 9:'-26+00;$"62&62$2"$$ <'/&"4+4;$%&+('$ ='/"2'$ %2+>'?"62$ D)0'.$$ ='>).2'4'5$)-$ 1+2+$E+-+>'/'-2$ 4,5.+% 6!.57'% Since 2008 : >1k distinct CMS users 100k analysis jobs/day in reach 4,5.+% /0'(12% 762&62$2"$$ 8.'4$%&+('$ 19

20

21

ECoM A group to consider Evolution of Computing Model from Startup to Steady State re-examine the CMS Computing Model using various different use-cases that may occur during startup and before "steady state" is reached. revisit the utilization of resources, the pattern and distribution of data and exactly what happens where and when. several use-cases should be explored in order to understand what flexibility and agility is available ahead of the startup so as to be able to make mid-course corrections as required A work in progress. 22

Back-up 23

!"#$#%&'$!())%**%(+%+,$! -./'012'*$! 3'*&$455$67+01(+45%&8$9':7%9';$69()$!"#$4&$'40<$*%&'$%+$4$0(+1+7(7*$)(;'$! ='&'9)%+'$%6$&<'$*%&'$%*$7*4.5'$4+;$*&4.5'$! ><4&$%*$&'*&';?$! @(.$*7.)%**%(+$! A(045$*%&'$0(+B,7941(+$4+;$!"#$*(CD49'$%+*&45541(+$! =4&4$400'**$4+;$;4&4$*&4,'E(7&$69()$.4&0<$+(;'$&($*&(94,'$! FG4H'I$4+458*%*$/(.*$! J745%&8$(6$;4&4$&94+*6'9*$409(**$*%&'*$! ><4&$%*$)'4*79';?$! #%&'$424%54.%5%&8K$69401(+$(6$1)'$455$67+01(+45$&'*&*$%+$4$*%&'$49'$*700'**675$! @(.$L(.(&$'M0%'+08K$69401(+$(6$*700'**675$F64H'I$4+458*%*$/(.*$! A%+H$:745%&8K$+7).'9$(6$;4&4$&94+*6'9$5%+H*$D%&<$4+$400'N&4.5'$:745%&8$! ><4&$%*$0450754&';?$! O$,5(.45$'*1)4&(9$D<%0<$'PN9'**'*$<(D$,((;$4+;$*&4.5'$4$*%&'$%*$ 24

!"#$%$&%'#()'*+,"%' Site summary table Site ranking Site history -./'01!'!2"/'0,332%%2,(2(4'"/#3' 25

!"#$%&'$()*+$,-$.+-/0$! 1"$2-&+.3-$45",&5$(3-'/+$*'$()-$-6"5.7"'$"8$()-$3-5*&,*5*(9$"8$+*(-+$! :2;3-++*6-$3-+.5(+$*'$()-$5&+($9-&3$! <--=59$3-6*-#+$"8$()-$+*(-$3-&/*'-++$! >3"/.%7"'$(-&2+$%&'$,-?-3$;5&'$#)-3-$("$3.'$;3"/.%7"'+$! @.("2&7%&559$2&;$("$;3"/.%7"'$&'/$&'&59+*+$(""5+$0$ 1)-$ABC$C*(-$A"22*++*"'*'4$(-&2$ 26

!"#$!%&'%()$*&+$!,-./0&1$#234)$!"#$+%./'%$;4%()6/&7$!%&'(%$)'$ <%(.08)=$!"#$*:4%(0.%&'$!/&'(/8$+//.$!"#$!%&'(%$)'$!*+,-$$./&0'/(0&12$3/.456&1$/4%()6/&72$ )&)89707$!$!"#$(/&&3&1$!,-./0&1$)234)$5678$!$9&:,/(*1%$(%-,'%$)234)$!$"*3&$'*);<$-,&3',($*&+$*=*(-$!"#$)3'%)$>$!,-./0&1$9?.%(')$ 27