Towards a Computing Grid for the LHC Experiments: ATLAS Data Challenge 1



Similar documents

The AMI Database Project: Atlas Data Challenge Bookkeeping, and the Tag Collector, a new tool for Release Management.

The CMS analysis chain in a distributed environment

ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC

Quality assurance and Data Validation of the ATLAS Data Challenges Simulation Samples

Automated software packaging and installation for the ATLAS experiment

World-wide online monitoring interface of the ATLAS experiment

Bulletin. Introduction. Dates and Venue. History. Important Dates. Registration

BNL Contribution to ATLAS

The next generation of ATLAS PanDA Monitoring

A multi-dimensional view on information retrieval of CMS data

Dual-use tools and systematics-aware analysis workflows in the ATLAS Run-2 analysis model

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

LHC Databases on the Grid: Achievements and Open Issues * A.V. Vaniachine. Argonne National Laboratory 9700 S Cass Ave, Argonne, IL, 60439, USA

The NorduGrid architecture and tools

Software, Computing and Analysis Models at CDF and D0

EDG Project: Database Management Services

Distributed analysis for the ATLAS Experiment in the S.Co.P.E Project

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

Roberto Barbera. Centralized bookkeeping and monitoring in ALICE

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil

Tier-1 Services for Tier-2 Regional Centres

Running a typical ROOT HEP analysis on Hadoop/MapReduce. Stefano Alberto Russo Michele Pinamonti Marina Cobal

A Process for ATLAS Software Development

SPACI & EGEE LCG on IA64

ATLAS job monitoring in the Dashboard Framework

First-year experience with the ATLAS online monitoring framework

Online Monitoring in the CDF II experiment

Guide. Axis Webinar. User guide

An Integrated CyberSecurity Approach for HEP Grids. Workshop Report.

Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007

THE CCLRC DATA PORTAL

Data Grids. Lidan Wang April 5, 2007

Guide. Axis Webinar User Guide

Big Data and Storage Management at the Large Hadron Collider

Context-aware cloud computing for HEP

Status and Evolution of ATLAS Workload Management System PanDA

Database Monitoring Requirements. Salvatore Di Guida (CERN) On behalf of the CMS DB group

Building a Volunteer Cloud

PRODUCT DATA. PULSE WorkFlow Manager Type 7756

FileMaker 11. ODBC and JDBC Guide

STAR-Scheduler: A Batch Job Scheduler for Distributed I/O Intensive Applications V. Mandapaka (a), C. Pruneau (b), J. Lauret (c), S.

SAP Business Objects Business Intelligence platform Document Version: 4.1 Support Package Data Federation Administration Tool Guide

Event display for the International Linear Collider Summer student report

The Persint visualization program for the ATLAS experiment

Shoal: IaaS Cloud Cache Publisher

Site specific monitoring of multiple information systems the HappyFace Project

TestScape. On-line, test data management and root cause analysis system. On-line Visibility. Ease of Use. Modular and Scalable.

PoS(EGICF12-EMITC2)110

The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets

Werkzeuge zur verteilen Analyse im ATLAS-Experiment

HEP Data-Intensive Distributed Cloud Computing System Requirements Specification Document

Evolution of Database Replication Technologies for WLCG

Deploying a distributed data storage system on the UK National Grid Service using federated SRB

iscala CRM: Optimize Your Revenue, Profitability and Customer Satisfaction

Microsoft Research Worldwide Presence

"FRAMEWORKING": A COLLABORATIVE APPROACH TO CONTROL SYSTEMS DEVELOPMENT

Distributed Database Access in the LHC Computing Grid with CORAL

THINK Global: Risk and return

AN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID

How To Use The Correlog With The Cpl Powerpoint Powerpoint Cpl.Org Powerpoint.Org (Powerpoint) Powerpoint (Powerplst) And Powerpoint 2 (Powerstation) (Powerpoints) (Operations

The LHCb Software and Computing NSS/IEEE workshop Ph. Charpentier, CERN

Arbeitspaket 3: Langzeitarchivierung von Forschungsdaten. JHOVE2 module for ROOT files 1

BaBar and ROOT data storage. Peter Elmer BaBar Princeton University ROOT Oct. 2002

GAUDI - A Software Architecture and Framework for building HEP Data Processing Applications

Status and Integration of AP2 Monitoring and Online Steering

Database Services for CERN

An on-line Integrated Bookkeeping: electronic run log book and Meta-Data Repository for ATLAS

The glite File Transfer Service

Managing globally distributed expertise with new competence management solutions a big-science collaboration as a pilot case

Evaluation of the CMT and SCRAM Software Configuration, Build and Release Management Tools

Dashboard applications to monitor experiment activities at sites

Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de

Software installation and condition data distribution via CernVM File System in ATLAS

An Evaluation of the Application Hosting Environment Uk e-science Engineering Task Force

Real Time Analysis of Advanced Photon Source Data

A Tool for Evaluation and Optimization of Web Application Performance

Retrieval on the Grid Results from the European Project GRACE (Grid Search and Categorization Engine)

New Features... 1 Installation... 3 Upgrade Changes... 3 Fixed Limitations... 4 Known Limitations... 5 Informatica Global Customer Support...

Cluster, Grid, Cloud Concepts

The Data Quality Monitoring Software for the CMS experiment at the LHC

A Web Services Data Analysis Grid *

Objectivity Data Migration

Analyzing Network Servers. Disk Space Utilization Analysis. DiskBoss - Data Management Solution

Data Management System for grid and portal services

Tools and strategies to monitor the ATLAS online computing farm

Analisi di un servizio SRM: StoRM

The Grid Monitor. Usage and installation manual. Oxana Smirnova

Status and Prospects of HARP. Malcolm Ellis On behalf of the HARP Collaboration NuFact02 Imperial College, July 2002

Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.

A J2EE based server for Muon Spectrometer Alignment monitoring in the ATLAS detector Journal of Physics: Conference Series

SysPatrol - Server Security Monitor

Data processing goes big

EMBL. International PhD Training. Mikko Taipale, PhD Whitehead Institute/MIT Cambridge, MA USA

Technical. Overview. ~ a ~ irods version 4.x

Software Development Kit

FileMaker 12. ODBC and JDBC Guide

Online data handling with Lustre at the CMS experiment

JD Edwards EnterpriseOne and JD Edwards World Compared. Contrasted.

Transcription:

EUROPEAN ORGANIZATION FOR NUCLEAR RESEARCH CERN-PH-EP/2004-028(rev) 31-January-2006 Towards a Computing Grid for the LHC Experiments: ATLAS Data Challenge 1 The ATLAS DC1 Task Force *) Abstract The ATLAS Collaboration is preparing for data taking and analysis at the CERN Large Hadron Collider, which will start operating in 2007. As part of this preparation, a series of computing Data Challenges was started in 2002 with the goal of the validation of the Computing Model, of the complete software suite, of the data model, and to ensure the correctness of the technical choices to be made for the final offline computing environment. A major feature of the first Data Challenge (DC1) was the preparation and the deployment of the software required for the production of large event samples as a worldwide distributed activity. It should be noted that it was not an option to run the complete production at CERN even if we had wanted to; the resources were not available at CERN to carry out the production on a reasonable time-scale. The great challenge of organising and then carrying out large-scale data production at a significant number of sites around the world had therefore to be faced. The benefits have been manifold: valuable experience has been gained in organizing and arranging access to large-scale, distributed computing resources, and the exercise has created worldwide momentum for ATLAS computing as a whole. This paper describes in detail the main activities carried out in DC1, and what has been learned from them, as a step towards a computing Grid for the LHC experiments. (Submitted to Nucl. Instr. Meth.) *) See next pages for the list of authors 1

R. Sturrock University of Melbourne, AUSTRALIA The ATLAS DC1 Task Force R. Bischof, B. Epp, V. M. Ghete, D. Kuhn Institute for Experimental Physics, University of Innsbruck, AUSTRIA A.G. Mello Universidade Federal do Rio de Janeiro, COPPE/EE/IF, Rio de Janeiro, BRAZIL B. Caron Centre for Subatomic Research, University of Alberta and TRIUMF, Vancouver, CANADA M.C. Vetterli Department of Physics, Simon Fraser University, Burnaby, CANADA G. Karapetian Laboratoire de Physique Nucleaire, Université de Montréal, CANADA K. Martens Department of Physics, University of Toronto, CANADA A. Agarwal, P. Poffenberger, R.A. McPherson a, R.J. Sobie a Department of Physics and Astronomy, University of Victoria, CANADA S. Armstrong, N. Benekos b, V. Boisvert, M. Boonekamp c, S. Brandt, P. Casado, M. Elsing, F. Gianotti, L. Goossens, M. Grote, J.B. Hansen, K. Mair, A. Nairz d, C. Padilla, A. Poppleton, G. Poulard, E. Richter-Was e, S. Rosati, T. Schoerner-Sadenius f, T. Wengler g CERN G.F. Xu Institute of High Energy Physics, Chinese Academy of Sciences, CHINA J.L. Ping Nanjing University, CHINA J. Chudoba, J. Kosina, M. Lokajicek, J. Svec Institute of Physics, Academy of Sciences of the Czech Republic, Praha, CZECH REPUBLIC P. Tas Charles University in Prague, Faculty of Mathematics and Physics, IPNP, Praha, CZECH REPUBLIC J. R. Hansen, E. Lytken, J. L. Nielsen, A. Wäänänen Niels Bohr Institutet for Astronomi, Fysik og Geofysik, Copenhagen, DENMARK S. Tapprogge h Helsinki Institute of Physics, Helsinki, FINLAND D. Calvet Université Blaise Pascal de Clermont-Ferrand, FRANCE a Now at the Institute of Particle Physics of Canada b Now at the MPI, Munich c Now at CEA-Saclay d Now at Leopold-Franzens University, Innsbruck e Now at Institute of Nuclear Physics, Crakow f Now at University of Hamburg g Now at the University of Manchester h Now at University of Mainz 2

S. Albrand, J. Collot, J. Fulachier, F. Ledroit-Guillon, F. Ohlsson-Malek, S. Viret, M. Wielers i LPSC, CNRS-IN2P3, Université Joseph Fourier, Grenoble, FRANCE K. Bernardet, S. Corréard, A. Rozanov, J-B. de Vivie de Regie CPPM, CNRS-IN2P3, Université de la Méditérannée, Marseille, FRANCE C. Arnault, C. Bourdarios, J. Hrivnac, M.Lechowski, G. Parrour, A. Perus, D. Rousseau, A. Schaffer, G. Unal LAL-Orsay, CNRS-IN2P3, Université Paris XI, Orsay, FRANCE F. Derue LPNHEP, CNRS-IN2P3, Université Paris 6/7, Jussieu, Paris, FRANCE L. Chevalier, S. Hassani, J-F. Laporte, R. Nikolaidou, D. Pomarède, M. Virchaux CEA/DAPNIA, Saclay, FRANCE N. Nesvadba j Rheinische Friedrich-Wilhelms Universität, Bonn, GERMANY S. Baranov Universität Freiburg, GERMANY A. Putzer Kirchhoff -Institut für Physik, Universität Heidelberg, GERMANY A. Khomich Universität Mannheim, GERMANY G. Duckeck, P. Schieferdecker LMU München, GERMANY A. Kiryunin, J. Schieck MPI für Physik, München, GERMANY Th. Lagouri Nuclear Physics Laboratory, Aristotle University of Thessaloniki, GREECE E. Duchovni, L. Levinson, D. Schrager, Weizmann Institute of Science, ISRAEL G. Negri CNAF, Bologna, ITALY H. Bilokon, L. Spogli k LNF, Frascati, ITALY D. Barberis, F. Parodi Università di Genova e INFN, ITALY G. Cataldi, E. Gorini, M. Primavera, S. Spagnolo Università di Lecce e INFN, ITALY D. Cavalli, M. Heldmann l, T. Lari, L. Perini, D. Rebatto, S. Resconi, F. Tartarelli, L. Vaccarossa Università di Milano e INFN, ITALY i Now at Rutherford Appleton Laboratory Deceased j Now at Max Planck Institute for Extraterrestrial Physics, Munich k Now at University Roma Tre l Now at University of Freiburg 3

M. Biglietti, G. Carlino, F. Conventi, A. Doria, L. Merola, Università di Napoli Federico II e INFN, ITALY G. Polesello, V. Vercesi Sezione INFN di Pavia, ITALY A. De Salvo, A. Di Mattia, L. Luminari, A. Nisati, M. Reale, M. Testa Università di Roma La Sapienza e INFN, ITALY A. Farilla, M. Verducci m Università di Roma Tre e INFN, ITALY M. Cobal, L.Santi Università di Udine e INFN, ITALY Y. Hasegawa Shinshu University, JAPAN M. Ishino, T. Mashimo, H. Matsumoto, H. Sakamoto, J. Tanaka, I. Ueda International Centre for Elementary Particle Physics (ICEPP), the University of Tokyo, JAPAN S.Bentvelsen, A.Fornaini, G.Gorfine, D.Groep, J.Templon NIKHEF, NETHERLANDS J. Koster n Parallab / UNIFOB, University of Bergen, NORWAY A. Konstantinov o, T. Myklebust p, F. Ould-Saada University of Oslo, Department of Physics, NORWAY T. Bold q, A. Kaczmarska, P. Malecki, T. Szymocha, M. Turala The Henryk Niewodniczanski Institute of Nuclear Physics, Krakow,POLAND Y. Kulchitsky r, G. Khoriauli, N. Gromova, V. Tsulaia s Joint Institute for Nuclear Research, Dubna, RUSSIA A. Minaenko t, R. Rudenko, E. Slabospitskay t, A. Solodkov Institute of High Energy Physics, Protvino, RUSSIA I. Gavrilenko P.N. Lebedev Institute of Physics (FIAN), Moscow, RUSSIA N. Nikitine, S. Sivoklokov, K. Toms Skobeltsyn Institute of Nuclear Physics, Moscow State University, RUSSIA A. Zalite, I. Zalite St.Petersburg Nuclear Physics Institute, RUSSIA B.Kersevan University of Ljubljana and Iozef Stefan Institut, Ljubljana, SLOVENIA m Now at CERN n Now at UNINETT Sigma, Trondheim o Also at IMSAR, Vilnius University p Now with Network Appliance Inc., Ann Arbor q Also at Faculty of Physics and Nuclear techniques, UST-AGH Krakow r Now at Institute of Physics of the National Academy of Sciences, Minsk s Now at University of Pittsburgh t Supported by a grant under contract CERN-INTAS-0052-4297 4

M. Bosman Institut de Física d'altes Energies (IFAE), Barcelona, SPAIN S. Gonzalez, J. Salt, J. Sanchez Instituto de Fisica Corpuscular (IFIC, Centro Mixto CSIC-UVEG), Valencia, SPAIN N. Andersson, L. Nixon NSC, Linköping University, SWEDEN P. Eerola, B. Kónya, O. Smirnova Particle Physics, Institute of Physics, Lund University, SWEDEN Å. Sandgren HPC2N, Umeå University, SWEDEN T. Ekelöf, M. Ellert, N. Gollub Department of Radiation Sciences, Uppsala University, SWEDEN S. Hellman, A. Lipniacka Department of Physics, Stockholm University, SWEDEN A. Corso-Radu u, V. Perez-Reale Laboratory for High Energy Physics, Bern, SWITZERLAND S.C. Lee, S.C. Lin, Z.L. Ren, P.K. Teng Institute of Physics, Academia Sinica, TAIWAN P. Faulkner, S.W. O Neale, A. Watson University of Birmingham, UK F. Brochu, C. Lester Cambridge University, UK J, Kennedy, S. Thompson University of Glasgow, UK E. Bouhova-Thacker, R. Henderson, R. Jones, V.Kartvelishvili, M. Smizanska Lancaster University, UK A. Washbrook Liverpool University, UK J. Drohan, N. Konstantinidis University College London, UK E. Moyse v Queen Mary and Westfield College, University of London, London, UK S. Salih Manchester University, UK J. Loken University of Oxford, UK J. Baines, D. Candlin, R. Candlin, R. Clifft, W. Li w, N. McCubbin RAL, UK u Now at CERN Deceased v Now at University of Massachusetts w Now at IHEP, Beijing 5

S. George, A. Lowe Royal Holloway College, University of London, Egham, UK C. Buttar x, I. Dawson, A. Moraes y, D. Tovey University of Sheffield, UK J. Gieraltowski, T. LeCompte, D. Malon, E. May, A. Vaniachine Argonne National Laboratory, USA D.L. Adams, K. Assamagan, R. Baker z, W. Deng, V. Fine, Y. Fisyak, B. Gibbard, H. Ma, P. Nevski, F. Paige, S.Rajagopalan, J. Smyth, A. Undrus, T. Wenaus, D. Yu Brookhaven National Laboratory, USA P. Calafiura, S. Canon aa, D. Costanzo bb, I. Hinchliffe, W. Lavrijsen., C. Leggett, M. Marino, D.R. Quarrie, I. Sakrejda, G. Stavropoulos, C. Tull Lawrence Berkeley National Laboratory, USA P. Loch University of Arizona, Tucson, USA J. Shank, S. Youssef Boston University, USA D. Engh cc, E. Frank, A. Gupta, R. Gardner, F. Merritt, Y. Smirnov University of Chicago, USA J. Huth Harvard University, USA L. Grundhoefer, F. Luehring Indiana University, USA S. Goldfarb University of Michigan, USA H. Severini, P. Skubic Oklahoma University, USA Y. Gao, T. Ryan Southern Methodist University, USA K. De, M. Sosebee, P. McGuigan, N. Ozturk University of Texas, Arlington, USA S. Gonzalez University of Wisconsin, Madison, USA x Now at University of Glasgow y Now at University of Glasgow z Now at Eureka College, California aa Now at SuperComputing Centre, Pittsburgh bb Now at University of Sheffield cc Now at University of Vanderbilt 6

1 Introduction In 2007 the Large Hadron Collider (LHC) is due to come into service at the European Particle Physics Laboratory (CERN) in Geneva. At the LHC two proton beams, each of energy 7 TeV, are steered to collide head-on at the centre of large complex detectors. These collisions are expected to reveal fundamental new processes in particle physics. The ATLAS 1 (A Toroidal LHC ApparatuS) detector is one of several that are being constructed to exploit the LHC. The ATLAS collaboration consists of about 1800 physicists and engineers from more than 150 universities and laboratories in 34 countries. The ATLAS detector weighs ~7000 tons and comprises several particle-detection technologies that surround the colliding beams almost completely. The ATLAS detector will measure the directions and energies of the particles produced in LHC collisions. At design intensity ( luminosity ) these collisions will occur at a rate of ~1GHz. This very high rate is required in order to accumulate statistically meaningful samples of the extremely rare processes that are expected to be of the highest physics interest. Since each collision (or event ) results in ~100-200 MB of off-line data, very fast real-time selection and on-line data reduction and compression by a factor of several million are required to produce manageable off-line data volumes. Even so, ATLAS will produce annually several petabytes of off-line data to be analysed by institutes around the world. To ensure that the requisite computing resources are available, each experiment will need a worldwide distributed databank and computer system, on a scale that is one to two orders of magnitude larger than previous experiments in particle physics. To prepare for this task, the LHC Computing Review 2 in 2001 recommended that the LHC experiments should carry out Data Challenges (DC) of increasing size and complexity. A Data Challenge comprises, in essence, the simulation, done as realistically as possible, of data (events) from the detector, followed by the processing of that data using the software and computing infrastructure that will, with further development, be used for the real data when the LHC starts operating. The goals of the ATLAS Data Challenges are the validation of the ATLAS Computing Model, of the complete software suite, of the data model, and to ensure the correctness of the technical computing choices to be made. From the point of view of LHC computing, the computational Grid 3 is a very timely development. The Grid concept is that data processing is carried out at the most suitable site(s) on an international multi-site network, with the so-called Grid middleware organizing access and assigning computing capacity in a way that is entirely seamless from the perspective of the user. Grid technologies offer several obvious advantages for a multinational and geographically distributed project: they allow for a uniform computing infrastructure of the project; they simplify the management and coordination of the resources while potentially decentralizing such tasks as software development and analysis; and last, but not least, they provide an affordable way to increase the effective computing power. 1 http://cern.ch/atlas/ 2 http://cern.ch/lhc-computing-review-public/public/report_final.pdf 3 Grid computing has grown rapidly since the late 1990 s, particularly after the publication of the influential The Grid: Blueprint for a New Computing Infrastructure (1999), edited by Foster and Kesselman The term Computational Grid (or Grid for short) has been coined by analogy with the electrical power grid. The world body is the Global Grid Forum, http://www.gridforum.org 7

The ATLAS collaboration intends to perform its Data Challenges using as much as possible Grid tools provided by the LHC Computing Grid (LCG) project 4, the NorduGrid 5 and the USGrid 6. DC1 saw the first usage of these technologies in ATLAS in limited-scale tests. In this report the ATLAS Data Challenge 1 (DC1) is described, with emphasis on the computing aspects. As background and introduction, Section 2 describes the scope and scale of DC1, and summarises the resources used and the amount of data processed. The main part of the paper is sections 3 and 4. Section 3 covers the various elements that comprise the computing infrastructure required for any large-scale exercise of this type: Software distribution, Data Book-keeping, Data Access and replication, and Workflow control. Section 4 reports on our use of some of the emerging Grid technologies for part of the DC1 exercise. Section 5 draws together the major lessons learned from the DC1, and the overall conclusions are given in Section 6. 2 DC1 Scope and Scale The design and construction of ATLAS requires a large amount of simulated data in order to optimise the design of the detectors, to estimate the physics performance that will be achieved, and to test the software and computing infrastructure. Samples of simulated data consist of a large number (~10 7 ) of simulated events. Each event is the result of a collision between two protons, and the full simulation requires the following steps: Event Generation : Particles produced in the proton-proton collision are generated using a program based on physics theories and phenomenology. Several such programs are available, and are referred to as event generators, or simply generators ; Detector Simulation : The produced particles are transported through the detector elements according to the known physics governing the passage of particles through matter. The resulting interactions with the sensitive elements of the detector are converted into information similar to, and in the same format as, the digital output from the real detector (the digitisation step); Because of the very high intensity at which the LHC will operate, many proton-proton collisions will usually occur during the sensitive time of the particle detectors. Hence the digitised output from the ATLAS detector will contain the piled-up signals from the particles produced in several collisions. This leads to the next step: Pile-up : the digitised signals from several simulated collisions are combined to produce events with realistic signal pile-up; These events can now be used to test the software suite that will be used on real LHC data: Reconstruction : the events are reconstructed to recover particle trajectories and energies from the detector data; 4 http://cern.ch/lcg/ 5 http://www.nordugrid.org 6 http://www.usatlas.bnl.gov/computing/grid/ 8

The simulated data fed to the Reconstruction step is, as far as possible, indistinguishable from real data, with one important exception: since the original information on the particles trajectories and energies is known from the Event Generation step, this information ( truth ) is also recorded so that the output from the Reconstruction step can be compared with it. For all Data Challenges it is essential to have realistic physics content in order to engage the physicist community with the exercise, leading to a more thorough validation of the software. For DC1, in 2002-2003, a major goal was to provide simulated data needed for the preparation of the ATLAS High Level Trigger (HLT) TDR 7. Operationally, the DC1 production was split into three phases: Phase 1: Event Generation and Detector Simulation, Phase 2: Pile-up, and Phase 3: Reconstruction, which are described in more detail below. Australia 954 48 9 668 Austria Canada CERN 900 China Czech Rep. 1036 178 France Germany 127 136 105 Greece Israel Italy 346 256 1141 Japan NorduGrid Poland 96 Russia 15 Spain 1215 218 300 133 Taiwan UK USA Fig. 1 Number of normalised processors 8 per country available to DC1 (Countries are displayed clockwise, starting with Australia (48)). The numbers of processors per site varied between 9 and 900, resulting in a total of ~5000 CPUs at peak time corresponding to a processing power of ~2.2 MSI2k. 7 ATLAS TDR 016; CERN/LHCC/2003-022; ISBN 92-9083-205-3; http://cern.ch/atlas-proj-hltdaqdcs-tdr/ 8 A normalised processor corresponds to 1 Pentium III/500 MHz. 9

The following 56 institutes in 21 countries 9 participated in DC1: 1. Australia (Melbourne) 2. Austria (Innsbruck) 3. Canada (Alberta, Montreal, Simon Fraser, Toronto, Victoria) 4. CERN 5. China (Beijing, Nanjing) 6. Czech Republic (Prague) 7. France (Grenoble, Marseille; using CCIN2P3 in Lyon) 8. Germany (Heidelberg, München; using GridKA in Karlsruhe) 9. Greece (Thessaloniki) 10. Israel (Weizmann) 11. Italy (CNAF Bologna, Frascati, Milano, Napoli, Roma) 12. Japan (Tokyo) 13. NorduGrid (NBI, Odense, Bergen, Oslo, Linkoping, Lund, Stockholm, Umea, Uppsala) 14. Poland (Cracow) 15. Russia (Dubna, ITEP Moscow, MSU Moscow, Protvino) 16. Spain (Valencia) 17. Taiwan (Taipei) 18. UK (Birmingham, Cambridge, Glasgow, Lancaster, Liverpool, RAL, Sheffield) 19. USA (ANL, Arlington, BNL, Boston, Dallas, Indiana, LBNL, New Mexico, Oklahoma) Fig. 2 Map of the sites (indicated by red dots) taking part in the ATLAS DC1 activities. The countries shaded in yellow have at least one institute in an LHC experiment. 9 The list of the NorduGrid sites used by the three Nordic countries (Denmark, Norway, Sweden) can be found on http://www.nordugrid.org/hardware.html 10

2.1 Phase 1: Event Generation and Detector Simulation Phase 1 was carried out during summer 2002. The Event Generation step, which is relatively light in terms of computing resources, was carried out at CERN, and the generated events were then distributed widely for the more compute-intensive Detector Simulation step. At peak time ~3200 processors located in 40 institutes in 19 countries were used in Phase 1. This corresponds to ~1400kSI2k 10 or ~50% of the CPU power estimated to be required for one Regional Facility ( Tier 1 ) at the time of LHC start-up (2007). The hardware investment, in cash terms, made by those institutes in the twelve months prior to the DC1 exercise corresponds to ~50 % of the annual hardware investment needed from 2006 onwards for the non-cern part of the ATLAS Offline Computing. 2.1.1 Event Generation The event generation used the event generator Pythia 6.203 11 running in the ATLAS Athena/Gaudi 12 framework. The ATLAS physics and HLT communities required many differing samples of events, with each sample characterised by certain requirements on the type, energies, and directions of subsets of the final-state particles. Because of the physics governing the production of final-state particles, it is not possible to impose such requirements with 100% efficiency through the input parameters to Pythia: it can only be done approximately. In practice ~50 million events were generated by Pythia, of which only ~11 million were selected by a fast filter for full Detector Simulation. In addition, ~40 million single-particle events (muons, photons, electrons, pions) were generated for full Detector Simulation in order to scan the detector response in a systematic way. These events can be generated very rapidly as there is only one particle per event, and it was convenient to use the older Fortran-based Atlsim 13 framework for this task. The total CPU time necessary to generate all the events was about 200kSI2k-days and the resulting data volume was 8 TBytes. 2.1.2 Detector Simulation The ATLAS detector simulation code used for DC1 is Fortran-based. It uses GEANT 3.21 14 to track the particles through the detector and runs in the Atlsim framework. 15 As mentioned above, the events generated by Pythia were filtered at the start of the detector-simulation job to select those (~11 million) that met certain requirements on final-state particles. These selected events, and all the ~40 million single-particle events, were then processed through the full detector simulation, and the results were written out in the form of ZEBRA 16 banks. 10 SI2k stands for SPEC INT 2000, the classification of CPU integer performance defined by the Standard Performance Evaluation Corporation. See http://www.spec.org/cpu2000/ 11 http://www.thep.lu.se/~torbjorn/pythia.html 12 http://cern.ch/atlas-proj-computing-tdr/html/computing-tdr-21.htm 13 http://cern.ch/atlas/groups/software/documents/atlsim/atlsim.html 14 CERN Program Library W5013: http://cern.ch/wwwasdoc/geant_html3/geantall.html 15 ATLAS has now completed its testing and verification of the C++ GEANT4 package (http://cern.ch/wwwasd/geant4/geant4.html ), and the detector simulation starting with DC2 uses GEANT4 running in the Athena/Gaudi framework. 16 CERN Program Library Q100/Q101 (http://cern.ch/wwwasdoc/zebra_html3/zebramain.html ) 11

As already noted, the detector simulation requires substantial computing resources, and this step was carried out at 40 institutes in 19 countries. The compute power required was 1.4 M SI2k-days, and generated an output data volume of 24 TBytes. 2.2 Phase 2: Pile-up production The cross-section for inelastic non-diffractive pp interactions at the LHC is expected to be around 67 mb. At the design intensity (luminosity of 10 34 cm 2 s 1 ) of the LHC, this implies an average of 23 minimum-bias events per bunch crossing, varying according to a Poisson distribution. Any collision recorded in the ATLAS detector therefore contains a superposition of particles coming from several events. In general the particles from a single "interesting physics" event will have triggered the readout, and additional particles will come from other uninteresting pp collisions. The number of particles seen by an ATLAS sub-detector depends not only on the LHC intensity but also on the signal-collection time of that sub-detector. The collisions within a given bunch crossing occur essentially instantaneously (time-spread < 1 ns), and all ATLAS sub-detectors are sensitive to all these collisions. Some of the sub-detectors have a sensitive time much greater than the 25 ns interval between bunch crossings, and hence are sensitive to collisions in bunch crossings occurring before and after the bunch crossing containing the interesting collision that triggered the readout. Furthermore, neutrons fly around the ATLAS cavern (in which the detector is installed) for a few seconds until they are thermalised, thus producing a kind of permanent neutron-photon bath resulting in a steady rate of Compton electrons and spallation protons, which are observed in the outermost parts of the detector. This component, i.e. additional hits created by long-lived particles, is called "cavern background". It is, therefore, quite a complex task to add the pile-up, ensuring that the signals from the uninteresting collisions are properly distributed in time and overlay correctly, sub-detector by sub-detector, the signals from the interesting collision. Pile-up was added to a sub-set of data samples: 3.9M events were produced with low-luminosity (2.10 33 cm 2 s 1 ) pile-up and 2.65M events with pile-up corresponding to design luminosity (10 34 cm 2 s 1 ). New countries, China and Greece, and new institutes from Canada, Italy, NorduGrid, UK and USA joined the effort in the course of the Phase 2 so that 56 institutes in 21 countries participated in the pile-up production, giving a maximum capability of more than 2 MSI2k. This phase of DC1 took about 3.2 MSI2k-days and produced a total data volume of about 34 TBytes. 2.3 Phase 3: Reconstruction The ATLAS reconstruction software has been developed from the earlier Fortran version, largely re-written in C++, and runs in the Athena/Gaudi framework. In essence, the Athena framework embodies a separation between data-specific members and methods on the one hand and algorithm-related members and methods on the other, e.g. for reconstruction. Data objects are handled through the Transient Event Store for event information and a Transient Detector Store for condition information. Data (specified by object type and string key) are read by Algorithms, or persistified by Converters. Algorithms are driven through a flexible event loop. Common utilities are provided through services. ASCII files, called Job Options, 12

allow the specification of algorithms and services parameters and sequencing. The Athena executable itself is very small, as all the significant software is dynamically loaded at run time, with typically one library per package. DC1 was carried out before the new C++ data-persistency system (POOL 17 ), being developed in the LCG context, was available, and hence the output of reconstruction was stored in HBOOK ntuples. This was done using a special algorithm, named CBNT for ComBined NTuple, capable of writing the content into an ntuple through the Gaudi ntuple converter. The algorithm is fully configurable, so that only the needed information is written out, which is especially important for the large truth information. The main drawback was that the downstream analysis could only be done in PAW (or ROOT after conversion), as the C++ design of the original objects was not preserved. In a separate step, algorithms running in the HLT Selection Software environment reconstructed objects and extracted the key features from event data that are used to derive the trigger decision. To facilitate the access to the large distributed datasets, since not all production sites were accessible via Grid tools, the data for the reconstruction step were replicated to just 8 sites, and the reconstruction of the data was (mostly) carried out at those sites. About 6.4M events were processed during the reconstruction phase. This part of DC1 took about 4.4 M SI2kdays and produced an output data (ntuple) volume of about 200 GBytes. 2.4 Software and Site Validation In any computing exercise of this type it is of course essential to test very thoroughly that the software is functioning correctly before embarking on large-scale use of computing resources. As significant parts of DC1 were carried out at geographically distributed sites, it was also essential to validate each site, i.e. to ensure that the installed software was functioning as expected. 2.4.1 Monitoring Generated Events As described in sections 2.1 above, the event-generation step was carried out at CERN. The generated events were monitored by histogramming various characteristic properties of the events, and gross errors (e.g. a mistyped input parameter) were easily detected in such histograms. A characteristic feature of the generated events is that they should contain one or more jets of final-state particles. ( Jets are tightly collimated groups of final-state particles that arise from the fragmentation of the more fundamental quarks or gluons.) Identifying the jets in an event is one of the key tasks of the reconstruction and analysis software. In order to check the jet properties of the generated events without having to go through the full chain of detector simulation, reconstruction, and analysis, the package Atlfast was used. As its name implies, Atlfast is an extremely fast package that approximates well the full detector simulation and particle reconstruction, using carefully tuned parameterisations. Jet finding was performed by running Atlfast, and then using Atlfast utilities to perform the jet finding at the particle level. The output ntuple was then used in a secondary job, running in the Physics Analysis Workstation (PAW 18 ) environment, to produce histograms of the number of 17 POOL: http://lcgapp.cern.ch/project/persist/ 18 http://cern.ch/paw/ 13

reconstructed jets, their transverse momentum (p T ) spectra and pseudo-rapidity distributions. These are normalised in various ways as appropriate: to the number of events; to the number of jets; and to the total cross section. Finally, the two histogram samples (event characteristics and jet properties) were merged and a postscript summary of all of the histograms produced was made and checked for consistency with the physics expectations for the given sample. 2.4.2 Quality Assurance and Validation The aim of the ATLAS DC quality assurance and validation procedure 19 was threefold: - to ensure the compatibility and reproducibility of samples produced at different sites (site validation); - to monitor the changes and improvements to the ATLAS detector geometry; - to check the physics content of the generated samples (physics validation). As the new C++ reconstruction software to be used in DC1 (section 2.3) was undergoing continual development, it was decided to base the quality assurance and validation on the older, but stable, Fortran-based reconstruction framework, Atrecon 20. The validation test suite consists of a modular analysis structure based on PAW, which runs off a general-purpose ntuple from Atrecon and which contains information on Monte Carlo event generation and the reconstruction for all ATLAS sub-detectors. The analysis procedure consists of comparing two output ntuples in two steps. First, an openended list of sub-detector-specific macros is run from a master process to produce the two sets of validation histograms. Second, a histogram-by-histogram comparison is performed between two sets of validation histograms, providing a bin-by-bin significance plot and a χ 2 test. At the end a summary χ 2 bar chart for all compared histograms is made. The site validation was done by comparing the outputs from identical input samples run at different sites and by comparisons of larger, statistically independent, samples of the same physics process. The validation provided an important check of the simulation infrastructure at the contributing DC sites. For example, it made it possible to spot slight but significant differences of the run-time libraries. During the initial phase this was a quite complex and intensive, but absolutely necessary, activity. The physics validation of the data was carried out in parallel with the other checks. A comparison of the number of jets, b-jets, c-jets, electrons and photons in each event, with a similar sample produced in a previous large-scale production, was performed. In addition, new fully simulated samples were inspected and were used for detailed detector calibration and similar purposes. New samples were also used extensively to study b-tagging. The b- physics group validated the DC1 simulation-reconstruction software chain for several detector layouts 21. 19 http://cern.ch/atlas/groups/software/dc/validation/www 20 http://cern.ch/atlas/groups/software/documents/reconstruction.html 21 DC1-b-physics validation: ATL-COM-PHYS-2003-003 14

3 Computing Elements of DC1 In section 2 we have outlined briefly the scope of a large-scale simulation study for the ATLAS experiment, i.e. what DC1 was trying to achieve at each stage. We turn now, and for the rest of this paper, to how DC1 was carried out, in terms of the software infrastructure used. This infrastructure is applicable to any large-scale computing operation of this type, carried out in a worldwide context. We identify the following elements: 1. Software distribution: the ability to distribute and install the software executables and libraries is an essential pre-requisite for the use of worldwide computing resources, and the distribution process used for DC1 is described in section 3.1. 2. Data bookkeeping: DC1 comprises in essence a series of transformations on large quantities of data and the details ( metadata ) of each transformation must be recorded. The database used for this purpose for DC1 is described in section 3.2. 3. Data access and replication: Distributed processing of data requires knowledge of where data resides, so that it can be accessed or copied ( replicated ) to appropriate storage. The infrastructure used for this is described in section 3.3. 4. Workflow control: This covers the preparation of job parameters, the submission of jobs, monitoring their progress, and the action taken if a job fails. This is described in section 3.4. It is worth emphasising that these elements are present, logically, in almost any computing task that involves the processing of data external to the computer programme. As the scale of such data processing has grown, software has been evolved to automate the overall management. From this perspective the Grid is just the latest step in the development of software to automate these elements as much as possible. The use of three of the emerging Grids, NorduGrid, US Grid, and EDG, for part of DC1 is described in section 4. 3.1 Software Distribution The ATLAS software is split into more than 500 packages residing in a single CVS repository at CERN. The Configuration Management Tool, CMT 22, manages package dependencies, libraries and building the executables. New releases are built at CERN approximately every three weeks, following an agreed schedule for the introduction of new features into the software. The compilation process is done on Linux machines. Users with a good network connection and access to AFS could use executables and data files directly from CERN. This approach is of course not suitable for remote sites with a limited network bandwidth or without access to AFS. Therefore, the relevant ATLAS software releases have been packaged into kits in RPM (RedHat Package Manager) format. For DC1 the kits, along with the installation scripts, were available for download via secure web connection or, otherwise, from the various Grid sites, EDG, USGrid or NorduGrid. 22 http://www.cmtsite.org/ 15

The general criteria followed during the package-architecture development phase have been to build a self-consistent distribution procedure, not dependent on the Linux version. The installation has been designed to keep the same directory structure as in the CERN repository. To be consistent with the reference software produced at CERN, all the executables and libraries shipped with the kit are copied from the official ATLAS AFS software repository. The packages are organized in a set of base tools that are required for all installations, and several additional components. The minimal installation provides the following items: -the set-up and management scripts; -the official ATLAS compilers; -the required libraries not part of the ATLAS software (external packages); -the ATLAS executable, libraries and data needed at runtime. Other packages are provided to enable code development. The RPM suites have proven to be robust and efficient. Most of the countries and sites have installed the software using the official set of RPMs, but some sites for the DC1 production have also adopted other types of installations. In particular a procedure based on full mirroring of the distributions directly from the CERN AFS repository, and the production of an alternate set of RPMs, have been used on the NorduGrid testbed. The main drawback found in the use of RPMs was the lack of flexibility: bug fixes in the new reconstruction software required entire new releases to be built and distributed. 3.2 Data Bookkeeping The associated bookkeeping and meta-data services are an essential component of DC1. The DC1 data comprises a large number of datasets and partitions. Each "dataset" is a collection of events, and has a logical dataset name, unique within ATLAS, assigned according to a nomenclature established by the production team. Datasets are further divided into partitions because of file-size limitations. A partition is a file that contains a part of a dataset, and is assigned a unique Logical File Names (LFN) that contains, by convention, at least the key part of the logical dataset name, and a partition number. 3.2.1 AMI The Atlas Metadata Interface (AMI) is a framework consisting of relational databases that contain their own description, a software layer to exploit this auto-description, and a set of interfaces for insertion and searching. AMI was used in the context of DC1 to manage the bookkeeping database that stores descriptive information (meta-attributes) about the data (binary files). The attributes of the binary data file that are stored by the bookkeeping database are "logical". These are application specific properties that could not be guessed by any outside system, such as a grid application. An example of a logical attribute is the type of event contained in the data file. A logical attribute never changes when data files are replicated. 16

The aim of the bookkeeping catalogue is twofold: - to make it possible to understand the contents of a file of binary physics data without actually having to open it, - to search for data given a set of logical attributes. The set of attributes defined for DC1 datasets was established in close collaboration with the production management team. It does include a few parameters that are not strictly logical, for example the data file sizes. These were included to ease the production of statistics. AMI provides several interfaces for use by physicists, and an API for developers. AMI Architecture AMI is written in Java. In consequence, it is independent of platform, operating system and database technology. The only prerequisite is that Java is installed on the client system. The AMI software is built in three levels above the Java Data Base Connection (JDBC) layer, as shown in Figure 3. The lower level packages wrap the remote connections to database, the transmission of SQL commands, and the recuperation of query results. While mysql was used for DC1, any database that understands SQL, and for which a JDBC driver is available, may in principle be used. Each AMI compliant database namespace must contain a certain number of tables that describe it. The middle layer of AMI provides generic classes for making use of these internal descriptions. For example, a table called db_model exists in each namespace, which describes the relations between tables. This generic mechanism hides details of database implementation from clients. A client who queries an AMI database is not expected to have knowledge of the name of the database, or the name of any database tables, or the relation between the database tables. The architecture allows clients to express queries in terms of the application semantics. Thus a user of the DC1 AMI production bookkeeping database should be able to work with a schema of datasets, partitions or files, whereas a client of another AMI based application would work with a different semantic. The lower and middle layers of software are common to the two sets of users. The generic web-search interface is part of the middle layer of AMI. The top layers of the software are application specific. Packages in this top layer have knowledge of non-generic features of the database tables. In the case of DC1 bookkeeping, the top layer contains in particular classes to manage dataset provenance, dataset definition, dataset nomenclature protocol and production statistics. Some specific commands and web interfaces have been developed that make use of this layer. The JAVA API of AMI provides access to both the application specific and the generic functions. The architecture is designed to allow geographic distribution of databases; all connections pass through a central router, which redirects requests to the correct site. This central router should be mirrored. For DC1 however, all the databases are physically at the LPSC Grenoble, and are situated on the same server. 17

APPLICATION SPECIFIC SOFTWARE: Java API for AMI databases Atlas Production Packages Other Packages Middle software layer for AMI compliant databases GENERIC SOFTWARE: USED BY ALL AMI APPLICATIONS Lower level of AMI software handles connections and SQL syntax Java Data Base Connection Layer and DB specific drivers APPLICATION DATABASES Atlas Production Databases Other AMI compliant Databases Fig. 3 The 3-tier architecture of AMI The Command Line Interface A large proportion of the updating of AMI was done by using the production tools AtCom and GRAT described below. AMI also provides a command line interface so that physicists running jobs without using the production tools were able to insert to and update the bookkeeping database. For DC1 the client software was available either as a stand-alone JAR file, or as part of the Atlas software release. The client software makes use of a configuration file containing the server and port details of the database, and other parameters common to all the commands. The first client session must configure the software using a database password distributed by a separate mechanism. A large number of commands are available, including commands to query the database, and to obtain information about the attributes. For jobs that did not use AtCom or Grat, the information for updating AMI was contained in batch job log files. Scripts were provided to parse the log files and to generate a set of AMI commands for insertion or update of AMI records. The Web Interfaces AMI contains a generic read-only web interface for searching. It is generated from the autodescriptions of the core databases, which means that the database schema can be changed without touching the web interface code. The interface has a "quick" search, where certain fields for searching are pre-selected, and an "advanced" search, which gives access to all the 18

database fields. In both cases, the SQL constructed is visible to the user, and can be directly edited if the user desires. Users can navigate from the search results to a graph showing the provenance of a dataset, and also to the MAGDA database that contains information on the physical location of the files. Other special Atlas Production interfaces have been provided: Users can request a new dataset by specifying the desired characteristics. The request is then sent to the production manager. The Production manager can examine the requested dataset, and either approve the dataset, with the possibility of editing some fields, or refuse it. A third interface is available to give an overview of the state of production. Some simple statistics are available, and it is possible to obtain a pie chart of jobs done per production site. 3.3 Data Access and Replication To complement the information stored by AMI, a well-defined strategy for how to replicate and access the large worldwide-distributed datasets is needed to ensure that: - the provenance of each partition is uniquely defined and documented (including all processing and selection steps i.e. the metadata information); - identical results are obtained independent of the actual location of each replica. 3.3.1 Manager for Grid-based Data (MAGDA) MAGDA 23, developed at Brookhaven National Laboratory (BNL), provides automated file registration and replication tools. Thus it is a manager well suited to Grid-based data, but can be used in a non-grid environment. It is built upon a database that describes where the data reside; thus it complements AMI, which describes what the data are. The ATLAS experiment data are distributed across globally dispersed storage facilities. In MAGDA the concept of a site is used to abstract the storage facility, as shown in Figure 4. The term location is used to denote data locations within a site typically directory trees. Locations carry attributes characterizing the data stored there, such as whether the data are a master or replica copy. The concept of a host is used to represent a set of computers that have access to a defined set of sites. Thus a MAGDA service, having knowledge of the host it is running on, is automatically aware of all the data locally accessible to it. MAGDA makes use of MySQL, Perl, Java, and C++ to provide file cataloguing, retrieval and replication services. For data movement, gridftp, bbftp and scp can be chosen depending on available protocols at the given site. Third party transfers are also implemented, if allowed. The principal catalogues maintained by MAGDA, as MySQL tables, are: a file catalogue with logical and physical file information and metadata; site, location and host catalogues; a catalogue of logical file collections; and a task catalogue. File metadata describe the file 23 http://arxiv.org/abs/physics/0306105 19

attributes and content. The file catalogue supports the notion of master and replica instances. Automatic file replication operations are organized into reusable tasks. Fig. 4 Logical relations among site, location and host in MAGDA MAGDA provides the following major services: 1. Setting up and managing distributed sites with associated locations, and locations within those sites, and the hosts on which data-gathering servers and user applications run. 2. Gathering the file information from the various sorts of data sites. A file spider cron job can be set up to automatically scan several storage sites and reports to the catalogue database. 3. Interfacing to users via web interfaces for presenting and querying catalogue information and for modifying the system. 4. Replicating and serving files to production and end-user applications. All input and output files in DC1 were registered in the MAGDA catalogue. Automated tools like GRAT (see section 4.2) searched for and transferred input and output files from grid sites using the MAGDA client tools. Physicists used similar tools to find data files for analysis. MAGDA spiders were used to catalogue files that were generated on batch machines using tools other than GRAT. Over 1.5 million files at hundreds of sites were catalogued and made accessible for DC1 through MAGDA. 3.4 Workflow Control A number of tools were developed and used to facilitate the production and the monitoring of the DC1 data-processing. The principal ones are: AtCom 24 (short for ATLAS Commander), developed in Europe, an automated job definition, submission and monitoring tool, working directly with AMI; GRAT 25 (short for the GRid Application Toolkit), developed in the US, an automated Grid-based workflow management tool. 24 http://atlas-project-atcom.web.cern.ch 25 http://heppc1.uta.edu/atlas/software/grat/ 20