Tier 1 Services - CNAF to T1



Similar documents
CDFII Computing Status

Software, Computing and Analysis Models at CDF and D0

The CMS analysis chain in a distributed environment

The Grid-it: the Italian Grid Production infrastructure

An objective comparison test of workload management systems

A Web-based Portal to Access and Manage WNoDeS Virtualized Cloud Resources

RO-11-NIPNE, evolution, user support, site and software development. IFIN-HH, DFCTI, LHCb Romanian Team

CDF software distribution on the Grid using Parrot

Grid Computing in Aachen

Roberto Barbera. Centralized bookkeeping and monitoring in ALICE

The ENEA-EGEE site: Access to non-standard platforms

Integrating a heterogeneous and shared Linux cluster into grids

GRID workload management system and CMS fall production. Massimo Sgaravatto INFN Padova

The dcache Storage Element

Computing in High- Energy-Physics: How Virtualization meets the Grid

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Status and Evolution of ATLAS Workload Management System PanDA


The GENIUS Grid Portal

Distributed Computing for CEPC. YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep.

SSM6437 DESIGNING A WINDOWS SERVER 2008 APPLICATIONS INFRASTRUCTURE

The ENEA gateway approach providing EGEE/gLite access to unsupported platforms and operating systems

BaBar and ROOT data storage. Peter Elmer BaBar Princeton University ROOT Oct. 2002

INFN Testbed status report

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC

Measurement of BeStMan Scalability

LHC schedule: what does it imply for SRM deployment? CERN, July 2007

MSU Tier 3 Usage and Troubleshooting. James Koll

Using L&B to Monitor Torque Jobs across a National Grid

Analisi di un servizio SRM: StoRM

SPACI & EGEE LCG on IA64

Designing a Windows Server 2008 Applications Infrastructure

Course 6437A: Designing a Windows Server 2008 Applications Infrastructure

Cyberinfrastructure Education and Hands-on Training Using the CH3D-GTM Virtual Appliance on SURAGrid

Software installation and condition data distribution via CernVM File System in ATLAS

EDG Project: Database Management Services

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

FermiGrid Highly Available Grid Services

U-LITE Network Infrastructure

IT-INFN-CNAF Status Update

dcache - Managed Storage - LCG Storage Element - HSM optimizer Patrick Fuhrmann, DESY for the dcache Team

Security in OSG. Tuesday afternoon, 3:15pm. Igor Sfiligoi Member of the OSG Security team University of California San Diego

Plateforme de Calcul pour les Sciences du Vivant. SRB & glite. V. Breton.

Sun Grid Engine, a new scheduler for EGEE

A multi-dimensional view on information retrieval of CMS data

6437A: Designing a Windows Server 2008 Applications Infrastructure (3 Days)

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

GESTIONE DI APPLICAZIONI DI CALCOLO ETEROGENEE CON STRUMENTI DI CLOUD COMPUTING: ESPERIENZA DI INFN-TORINO

Florida Site Report. US CMS Tier-2 Facilities Workshop. April 7, Bockjoo Kim University of Florida

SEE-GRID-SCI. SEE-GRID-SCI USER FORUM 2009 Turkey, Istanbul December, 2009

High Availability Databases based on Oracle 10g RAC on Linux

Bringing Federated Identity to Grid Computing. Dave Dykstra CISRC16 April 6, 2016

The EDGeS project receives Community research funding

Roadmap for Applying Hadoop Distributed File System in Scientific Grid Computing

GridICE: monitoring the user/application activities on the grid

Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb

CERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN

Storage solutions for a. infrastructure. Giacinto DONVITO INFN-Bari. Workshop on Cloud Services for File Synchronisation and Sharing

Retirement of glite3.1 and unsupported glite 3.2

Configuring Security Features of Session Recording

Designing a Windows Server 2008 Applications Infrastructure

GRIDSEED: A Virtual Training Grid Infrastructure

Data Management. Issues

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

See-GRID Project and Business Model

Client/Server Grid applications to manage complex workflows

Service Challenge Tests of the LCG Grid

John Kennedy LMU München DESY HH seminar 18/06/2007

A Web Services Data Analysis Grid *

ATLAS job monitoring in the Dashboard Framework

Adding IaaS Clouds to the ATLAS Computing Grid

A Uniform Job Monitoring Service in Multiple Job Universes

HADOOP, a newly emerged Java-based software framework, Hadoop Distributed File System for the Grid

SAM-FS - Advanced Storage Management Solutions for High Performance Computing Environments

YAN, Tian. On behalf of distributed computing group. Institute of High Energy Physics (IHEP), CAS, China. CHEP-2015, Apr th, OIST, Okinawa

PES. Batch virtualization and Cloud computing. Part 1: Batch virtualization. Batch virtualization and Cloud computing

glite Job Management

Virtualization, Grid, Cloud: Integration Paths for Scientific Computing

Yale Software Library. PGP 9.6 for Windows

The GRID and the Linux Farm at the RCF

Single Sign-In User Centered Computing for High Energy Physics

Integration of Virtualized Worker Nodes in Standard-Batch-Systems CHEP 2009 Prague Oliver Oberst

AFS Usage and Backups using TiBS at Fermilab. Presented by Kevin Hill

Michael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week

Objectivity Data Migration

Assignment # 1 (Cloud Computing Security)

Data Management Plan (DMP) for Particle Physics Experiments prepared for the 2015 Consolidated Grants Round. Detailed Version

The Italian Grid Infrastructure (IGI) CGW08 Cracow. Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director

Understanding Disk Storage in Tivoli Storage Manager

The EU DataGrid Data Management

A Multilevel Secure MapReduce Framework for Cross-Domain Information Sharing in the Cloud

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

The glite Workload Management System

10th TF-Storage Meeting

LHC Databases on the Grid: Achievements and Open Issues * A.V. Vaniachine. Argonne National Laboratory 9700 S Cass Ave, Argonne, IL, 60439, USA

WM Technical Evolution Group Report

Solution Brief: Creating Avid Project Archives

OSG PUBLIC STORAGE. Tanya Levshina

Transcription:

CDF Report on Tier 1 Usage Donatella Lucchesi for the CDF Italian Computing Group INFN Padova Outline The CDF Computing Model Tier1 resources usage as today CDF portal for European GRID: lcgcaf People and Tier 1 support 1

CDF Computing Model: Analysis Data & MC flow User Desktop DH= Data access, catalog is SAM Remote Analysis System CDF=CAFs D0=Remote Farms Monte Carlo Production Robotic Tape Storage DH Disk cache DH Data processing Central Analysis System (Central Analysis Facility CAF CDFCNAF 2

CDF need to migrate to GRID Need to expand resources luminosity expect to increase by factor 8 Resource are in dedicated pools, limited expansion and need to be maintained Resource at large CDF has three projects: GlideCAF: : MC production on GRID, data processing at T1 Production systems: CNAF, San Diego, Fermilab, Lyon. LcgCAF: : MC production on GRID OSGCAF: : (USA) CDF priorities for computing outside Fermilab: Monte Carlo production Few selected site will become Physics centers CNAF the Italian Tier 1 chosen for B physics (partially already used as) 3

Current CDF Resources Access : GlideCAF Developed by S. Sarkar with I. Sfiligoi help Maintained @CNAF by Subir Sarkar Production since Sept. 2005, pure GlideCAF Dec. 2005 Portal Condor Gatekeeper Job outputs CNAF users disks or rcp home institution Tier 1 pool includes CDF resources 4

Current CDF Resources : Disk space Data Storage: o 65 TB in total; 40 TB used o Major B Physics datasets and skimmed data up to September 2005. New data import in progress Users disk space (icaf) : o 6.5 T only for Italians Data files are copied to wn, analyzed and the output stored icaf (Italians) or copied to own institution (non-italians) Grid storage element (SE): o 1.7 TB for Monte Carlo Production 5

Current CDF Resources : Disks Data disks Filesystem Size Used Avail Used% Mounted on /dev/gpfs_cdf1 11T 11T 644G 95% /storage/gpfs_cdf1 /dev/gpfs_cdf2 11T 11T 918G 92% /storage/gpfs_cdf2 /dev/gpfs_cdf3 7.6T 5.4T 2.2T 71% /storage/gpfs_cdf3 /dev/gpfs_cdf4 7.2T 427G 6.7T 6% /storage/gpfs_cdf4 /dev/gpfs_cdf5 7.2T 414G 6.7T 6% /storage/gpfs_cdf5 /dev/gpfs_cdf6 7.2T 544G 6.6T 8% /storage/gpfs_cdf6 /dev/gpfs_cdf7 7.2T 542G 6.6T 8% /storage/gpfs_cdf7 /dev/gpfs_cdf8 7.2T 559G 6.6T 8% /storage/gpfs_cdf8 Users disk space : 6.5 T only for Italians cdfdata01.cr.cnaf.infn.it:/cdf/data/01-06 Grid storage element (SE): dsgridit.cnaf.infn.it:/flatfiles/se00/cdf 1.7T 821G 904G 48% /cdf/data/se-test 6

Dataset Description Files Events Cached Locked skbh01 Bs->Dspi( Ds->Phipi) 108 1780764 none none skbh12 B+ ->D0pi( D0 ->Kpi ) Cabibbo Suppress 390 6976384 none none skbh02 Bs->Dspi( Ds->3pi ) 758 12998672 none none skbh03 Bs->Ds 3pi( Ds->Phipi 698 12154391 none none skbh04 B0->D+pi( D+ ->Kpipi ) 843 14959016 none none skbh05 B0->D+3pi(D+ ->Kpipi ) 1387 25216343 none none skbh06 B+ ->D0pi( D0 ->Kpi ) 404 7207271 none none skbh07 skim-2 Decay D* ->D0pi( D0 ->Kpi ) 281 5350494 none none bhel0d HIGH_PT_ELECTRON 4813 26499559 4759( 99%) 4759( 99%) chdr08 Dataset for BsDsKsK 765 15242505 all all chdl01 Dataset for BD0K3pi 655 12977735 653(100%) 653(100%) chdl02 Dataset for BD0KK 230 4534193 228( 99%) 228( 99%) chdl04 Dataset for BD0pipi 130 2429423 all all chdl07 Dataset for BsDsPhipi 100 1827438 99( 99%) 99( 99%) chdl09 Dataset for BsDsK*K 814 17104942 all all mbot0w Pythia all processes select b semimuonic 33 247144 all all xbmu0d B_MUON_COMPRESSED 4295 25645441 4290(100%) 4290(100%) xbhd0d B_HADRONIC_COMPRESSED, Stream H 42121 270149101 35046( 83%) 35046( 83%) xbhd0h B_HADRONIC_COMPRESSED, Stream H 13859 205848210 6853( 49%) 6816( 49%) xbel0d B_ELECTRON_COMPRESSED 6043 37051608 3638( 60%) none 7

CDF Catalog and disks access SAM is used as catalog for data and right now also as disk space manager but really painful. Maintained by D. Lucchesi and Gabriele Compostella (Trento PhD student) SAM/SRM interface under development (1 person paid by Italy now at fnal to work on that) Need to collaborate also with CNAF to make it CASTOR compliant. 8

GlideCAF: CPU usage 2.0K All running sections Maintenance Use all available resources Give back if others need ~200 active users ~670K jobs 7.0K All waiting sections Maintenance 9

GlideCAF: Users ~200 active users Users in the last 3 months Used for: MC production All users ntuples production Mainly Italians Second skimming performed by LBL people 10

January Tier 1 usage CPU time (hours) Total time (hours) 11

Tier 1 usage since April 24 2005 CPU time (hours) Total time (hours) 12

Porting to GRID: lcgcaf Developed by: F. Delli Paoli, D. Jeans, A. Fella with S. Sarkar and I. Sfiligoi help Maintained by F. Delli Paoli Secure connection via Kerberos Head node User Interface Resource Broker CDF code via AFS Grid site 1 Grid site n CDF Storage Element Job output 13

Monitoring Home made to mimic existing monitor developed by D. Jeans CDF jobs summary: web monitoring. Every 5 minutes queries to LB services and information caching. 14

Lcgcaf: Status Middleware configuration: glite WMS 1.4.1 (CDF patches) to use the DAG jobs CDF configuration: Code: distributed via AFS Run condition data: Frontier server at FNAL and a Squid Proxy/caching service at CNAF Beta test: used to produce Monte Carlo, two million of events produced on CNAF Tier1 and Padova Tier 2 Output on CDF SE then to FNAL Efficiency: ~95% Mainly authentication errors 15

Lcgcaf: Future The lcgcaf portal at CNAF will allow to use the European GRID. Tests performed with Barcelona: OK Lyon is another site we would like to include. Waiting for glite 3 (next week) to start a massive MC production and see if a running experiment efficiency level can be reached. Data Processing for the moment using GlideCAF at T1 16

CDF persons and Tier 1 interaction CDF experience with CNAF is good. So good that now the CDF service work is allowed also at CNAF, Luca Brigliadori (CDF- Bologna) is the first physicist. Currently CDF FTE: D. Lucchesi, F. Delli Paoli (Padova) 1.7 I. Sfiligoi (LNF) now at FNAL ~0.5 S. Sarkar, D. Jeans (CNAF) 1.5 G. Compostella (Trento) 0.2 Luca Brigliadori (Bologna) 0.5 17

CDF persons and Tier 1 interaction cont d Maintained by CNAF: CDF VOMS server (for the all collaboration) CDF farm now part of T1 farm Disks Good interaction also for GRID porting: Support for middleware modification to easy adapt CDF environment (CA support, WMS for DAG submission, speed up submission, ) Support for tests (lcgcaf queue create for us for tests to not interfere with production queue) anything I miss 18