Data analysis for the CMS experiment in the petabyte range at the Vienna Tier-2 centre

Size: px
Start display at page:

Download "Data analysis for the CMS experiment in the petabyte range at the Vienna Tier-2 centre"

Transcription

1 FAKT/ÖPG 2009 Natascha Hörmann, Dietrich Liko, Gerhard Walzel Data analysis for the CMS experiment in the petabyte range at the Vienna Tier-2 centre

2 Overview Motivation for Grid computing tektor Performance Studien: HLT Data analysis and data flow at the LHC experiment Vienna Grid Computing Centre Grid Computing Centre Stress test - Step 09 2

3 Motivation for Grid Computing LHC will start at 2009 and run for the next years Proton bunch crossing rate: 40 MHz Different rate of accepted events after Trigger (15/s 5000/s) Different recording rate ( MB/s) ~15 PetaByte/year data to store (~ 15 Mill. GigaByte/year) LHCb ~ 200 MB/s ATLAS ~ 300 MB/s CMS ~ 200 MB/s ALICE ~ 1250 MB/s 3

4 Motivation for Grid Computing in Austria Austrian-Grid Project Federated Tier-2 centre for LHC Austrian Grid User Vienna: CMS experiment Innsbruck: ATLAS experiment Applications in the context of AustrianGrid Computing research, new middleware,... Other applications: Biomed, Healthscience, Weather-forecast,... National Grid Initiative (NGI) lead by the university of Linz 4

5 Grid Structure Tier-0: CERN central centre Saving the experiment data Central data distributor to Tier-1 5

6 11 Tier-1 centre Grid Structure Copies of experiment data Reconstruction of the events Distributor to Tier-2s 6

7 Grid Structure ~160 Tier-2 centres at universities and research centres (and Tier-3) Re-Reconstruction of the events Skimming of the events MC Simulation of the events Data analysis 7

8 Huge amount of analysis data transfer Cumulative transfer volume of Cosmic and MC data over a year Tier-1->Tier-2: ~ 4.0 PetaByte/year 8

9 Monte Carlo Analysis data flow Tier-0 Cern Actual MC sample: SUSY analysis data: Signal (Low Mass SUSY), Background (ttbar, W +Jets, Z+Jets, QCD) Experiment Tier-1 Tier-1 Tier-1 Tier-1s MC Production and reconstruction of data Duration: 2 months Datasets: 100 TB Tier-2 Vienna Tier-2 Tier-2 Re-Skimming (Vienna) Duration: 1 week SUSY Skim: 1.7 TB Tier-2s Interactive analysis Grid Job submission Duration: 1-5 hours Skimming (PAT) (Physics Analysis Tools) Duration: 1 month SUSY Skim: 10 TB 9

10 Vienna Tier-2 centre Federated Tier-2 Vienna-Innsbruck (Austrian Grid) Resources at T2-Vienna Actual: 420 CPUs, 125 TeraByte Available: 1000 CPUs, 500 TeraByte Storage Element Supermicro RAID DataPoolManager (DPM) Computers Sun (Dual Quad (2x4) Core CPU s) Intel Xenon 2,5 GHz Linked via the university of vienna to the European Network GEANT2 10

11 Data transfers between Tiers Commissioning of the Tier-1 and Tier-2 links Reach defined transfer-rate over a period Tier-1: transfer rate > 20 MB/s for 1 day Tier-2: transfer rate > 5 MB/s for 1 day Tested periodically Tier-1,Tier-2 -> Tier-2 Vienna, links successfully commissioned 11

12 Transfers to Tier-2 Vienna Test data transfer rate over a year Av. transfer rate: 4.2 MB/s Cumulative transfer volume: 128 TeraByte/year 12

13 Computer Centre Stress Test Step 09 Scale Testing for the Experimental Program in 2009 Testing the Grid Infrastructure in June 2009 Tier-0 tests Testing to write to tape for several days Tier-0 and Tier-1 tests Testing the read/write on tape/disc Reconstructing and skimming datasets Testing Tier-1, Tier-2 transfers Tier-2 tests Testing Job submission 13

14 Job Stress Test Step 09 Results of the analysis jobs at the Tier-2s (June 2009) analysis jobs with a runtime of 4h during 2 weeks sent to T2-Vienna Successful: 96.4 % 14

15 Results - Step 09 Results of the analysis jobs at the CMS Tier-2s (June 2009): Successful finished jobs: o %: 24 Tier-2s o %: 7 Tier-2s o < 80 %: 19 Tier-2s 15

16 Vienna Tier-2 for everyone? What do you need? Grid Certificate User Interface installed on your local PC User Software installed or submitted with the job Starting cooperation with the technical university vienna/ theoretical physics institute: Thermodynamic simulation of quark gluon plasma Existing cooperation with the medical university vienna/ MedAustron: Simulations for radiation therapy using hadrons Interested in using the Vienna Grid Infrastructure? Our Grid experts at HEPHY will help you 16

17 Summary Physics analysis at LHC as well as other fields like biomedicine, weather-forecast has extremely high requirements on computer resources Testing the computing infrastructure before the LHC startup is extremely important Successful analysis stress test of the CMS Grid Infrastructure has taken place but there is still a lot to improve e.g. transfer rate and reliability Everyone can use the grid infrastructure 17

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Volker Büge 1, Marcel Kunze 2, OIiver Oberst 1,2, Günter Quast 1, Armin Scheurer 1 1) Institut für Experimentelle

More information

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan Status of Grid Activities in Pakistan FAWAD SAEED National Centre For Physics, Pakistan 1 Introduction of NCP-LCG2 q NCP-LCG2 is the only Tier-2 centre in Pakistan for Worldwide LHC computing Grid (WLCG).

More information

How To Teach Physics At The Lhc

How To Teach Physics At The Lhc LHC discoveries and Particle Physics Concepts for Education Farid Ould- Saada, University of Oslo On behalf of IPPOG EPS- HEP, Vienna, 25.07.2015 A successful program LHC data are successfully deployed

More information

Beyond High Performance Computing: What Matters to CERN

Beyond High Performance Computing: What Matters to CERN Beyond High Performance Computing: What Matters to CERN Pierre VANDE VYVRE for the ALICE Collaboration ALICE Data Acquisition Project Leader CERN, Geneva, Switzerland 2 CERN CERN is the world's largest

More information

CMS: Challenges in Advanced Computing Techniques (Big Data, Data reduction, Data Analytics)

CMS: Challenges in Advanced Computing Techniques (Big Data, Data reduction, Data Analytics) CMS: Challenges in Advanced Computing Techniques (Big Data, Data reduction, Data Analytics) With input from: Daniele Bonacorsi, Ian Fisk, Valentin Kuznetsov, David Lange Oliver Gutsche CERN openlab technical

More information

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015 (Possible) HEP Use Case for NDN Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015 Outline LHC Experiments LHC Computing Models CMS Data Federation & AAA Evolving Computing Models & NDN Summary Phil DeMar:

More information

BNL Contribution to ATLAS

BNL Contribution to ATLAS BNL Contribution to ATLAS Software & Performance S. Rajagopalan April 17, 2007 DOE Review Outline Contributions to Core Software & Support Data Model Analysis Tools Event Data Management Distributed Software

More information

SUPPORT FOR CMS EXPERIMENT AT TIER1 CENTER IN GERMANY

SUPPORT FOR CMS EXPERIMENT AT TIER1 CENTER IN GERMANY The 5th InternaEonal Conference Distributed CompuEng and Grid technologies in Science and EducaEon SUPPORT FOR CMS EXPERIMENT AT TIER1 CENTER IN GERMANY N. Ratnikova, J. Berger, C. Böser, O. Oberst, G.

More information

Scalable stochastic tracing of distributed data management events

Scalable stochastic tracing of distributed data management events Scalable stochastic tracing of distributed data management events Mario Lassnig mario.lassnig@cern.ch ATLAS Data Processing CERN Physics Department Distributed and Parallel Systems University of Innsbruck

More information

Calorimetry in particle physics experiments

Calorimetry in particle physics experiments Calorimetry in particle physics experiments Unit n. 8 Calibration techniques Roberta Arcidiacono Lecture overview Introduction Hardware Calibration Test Beam Calibration In-situ Calibration (EM calorimeters)

More information

A Physics Approach to Big Data. Adam Kocoloski, PhD CTO Cloudant

A Physics Approach to Big Data. Adam Kocoloski, PhD CTO Cloudant A Physics Approach to Big Data Adam Kocoloski, PhD CTO Cloudant 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 Solenoidal Tracker at RHIC (STAR) The life of LHC data Detected by experiment Online

More information

Grid Computing in Aachen

Grid Computing in Aachen GEFÖRDERT VOM Grid Computing in Aachen III. Physikalisches Institut B Berichtswoche des Graduiertenkollegs Bad Honnef, 05.09.2008 Concept of Grid Computing Computing Grid; like the power grid, but for

More information

HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS EKSPERYMENTY FIZYKI WYSOKICH ENERGII W SIECIACH KOMPUTEROWYCH GRID. 1.

HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS EKSPERYMENTY FIZYKI WYSOKICH ENERGII W SIECIACH KOMPUTEROWYCH GRID. 1. Computer Science Vol. 9 2008 Andrzej Olszewski HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS The demand for computing resources used for detector simulations and data analysis in High Energy

More information

Big Data and Storage Management at the Large Hadron Collider

Big Data and Storage Management at the Large Hadron Collider Big Data and Storage Management at the Large Hadron Collider Dirk Duellmann CERN IT, Data & Storage Services Accelerating Science and Innovation CERN was founded 1954: 12 European States Science for Peace!

More information

PHYSICS WITH LHC EARLY DATA

PHYSICS WITH LHC EARLY DATA PHYSICS WITH LHC EARLY DATA ONE OF THE LAST PROPHETIC TALKS ON THIS SUBJECT HOPEFULLY We may have some two month of the Machine operation in 2008 LONG HISTORY... I will extensively use: Fabiola GIANOTTI

More information

HIP Computing Resources for LHC-startup

HIP Computing Resources for LHC-startup HIP Computing Resources for LHC-startup Tomas Lindén Finnish CMS meeting in Kumpula 03.10. 2007 Kumpula, Helsinki October 3, 2007 1 Tomas Lindén Contents 1. Finnish Tier-1/2 computing in 2007 and 2008

More information

NT1: An example for future EISCAT_3D data centre and archiving?

NT1: An example for future EISCAT_3D data centre and archiving? March 10, 2015 1 NT1: An example for future EISCAT_3D data centre and archiving? John White NeIC xx March 10, 2015 2 Introduction High Energy Physics and Computing Worldwide LHC Computing Grid Nordic Tier

More information

Measurement of Neutralino Mass Differences with CMS in Dilepton Final States at the Benchmark Point LM9

Measurement of Neutralino Mass Differences with CMS in Dilepton Final States at the Benchmark Point LM9 Measurement of Neutralino Mass Differences with CMS in Dilepton Final States at the Benchmark Point LM9, Katja Klein, Lutz Feld, Niklas Mohr 1. Physikalisches Institut B RWTH Aachen Introduction Fast discovery

More information

Big Data Needs High Energy Physics especially the LHC. Richard P Mount SLAC National Accelerator Laboratory June 27, 2013

Big Data Needs High Energy Physics especially the LHC. Richard P Mount SLAC National Accelerator Laboratory June 27, 2013 Big Data Needs High Energy Physics especially the LHC Richard P Mount SLAC National Accelerator Laboratory June 27, 2013 Why so much data? Our universe seems to be governed by nondeterministic physics

More information

Online data handling with Lustre at the CMS experiment

Online data handling with Lustre at the CMS experiment Online data handling with Lustre at the CMS experiment Lavinia Darlea, on behalf of CMS DAQ Group MIT/DAQ CMS September 17, 2015 1 / 29 CERN 2 / 29 CERN CERN was founded 1954: 12 European States Science

More information

Performance Monitoring of the Software Frameworks for LHC Experiments

Performance Monitoring of the Software Frameworks for LHC Experiments Proceedings of the First EELA-2 Conference R. mayo et al. (Eds.) CIEMAT 2009 2009 The authors. All rights reserved Performance Monitoring of the Software Frameworks for LHC Experiments William A. Romero

More information

The Data Quality Monitoring Software for the CMS experiment at the LHC

The Data Quality Monitoring Software for the CMS experiment at the LHC The Data Quality Monitoring Software for the CMS experiment at the LHC On behalf of the CMS Collaboration Marco Rovere, CERN CHEP 2015 Evolution of Software and Computing for Experiments Okinawa, Japan,

More information

Science+ Large Hadron Cancer & Frank Wurthwein Virus Hunting

Science+ Large Hadron Cancer & Frank Wurthwein Virus Hunting Shared Computing Driving Discovery: From the Large Hadron Collider to Virus Hunting Frank Würthwein Professor of Physics University of California San Diego February 14th, 2015 The Science of the LHC The

More information

ADVANCEMENTS IN BIG DATA PROCESSING IN THE ATLAS AND CMS EXPERIMENTS 1. A.V. Vaniachine on behalf of the ATLAS and CMS Collaborations

ADVANCEMENTS IN BIG DATA PROCESSING IN THE ATLAS AND CMS EXPERIMENTS 1. A.V. Vaniachine on behalf of the ATLAS and CMS Collaborations ADVANCEMENTS IN BIG DATA PROCESSING IN THE ATLAS AND CMS EXPERIMENTS 1 A.V. Vaniachine on behalf of the ATLAS and CMS Collaborations Argonne National Laboratory, 9700 S Cass Ave, Argonne, IL, 60439, USA

More information

Secure Wide Area Network Access to CMS Analysis Data Using the Lustre Filesystem

Secure Wide Area Network Access to CMS Analysis Data Using the Lustre Filesystem Secure Wide Area Network Access to CMS Analysis Data Using the Lustre Filesystem D Bourilkov 1, P Avery 1, M Cheng 1, Y Fu 1, B Kim 1, J Palencia 2, R Budden 2, K Benninger 2, J L Rodriquez 3, J Dilascio

More information

GRID computing at LHC Science without Borders

GRID computing at LHC Science without Borders GRID computing at LHC Science without Borders Kajari Mazumdar Department of High Energy Physics Tata Institute of Fundamental Research, Mumbai. Disclaimer: I am a physicist whose research field induces

More information

Measurement of the Mass of the Top Quark in the l+ Jets Channel Using the Matrix Element Method

Measurement of the Mass of the Top Quark in the l+ Jets Channel Using the Matrix Element Method Measurement of the Mass of the Top Quark in the l+ Jets Channel Using the Matrix Element Method Carlos Garcia University of Rochester For the DØ Collaboration APS Meeting 2007 Outline Introduction Top

More information

Tier-1 Services for Tier-2 Regional Centres

Tier-1 Services for Tier-2 Regional Centres Tier-1 Services for Tier-2 Regional Centres The LHC Computing MoU is currently being elaborated by a dedicated Task Force. This will cover at least the services that Tier-0 (T0) and Tier-1 centres (T1)

More information

A multi-dimensional view on information retrieval of CMS data

A multi-dimensional view on information retrieval of CMS data A multi-dimensional view on information retrieval of CMS data A. Dolgert, L. Gibbons, V. Kuznetsov, C. D. Jones, D. Riley Cornell University, Ithaca, NY 14853, USA E-mail: vkuznet@gmail.com Abstract. The

More information

The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links. Filippo Costa on behalf of the ALICE DAQ group

The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links. Filippo Costa on behalf of the ALICE DAQ group The new frontier of the DATA acquisition using 1 and 10 Gb/s Ethernet links Filippo Costa on behalf of the ALICE DAQ group DATE software 2 DATE (ALICE Data Acquisition and Test Environment) ALICE is a

More information

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

CMS Tier-3 cluster at NISER. Dr. Tania Moulik CMS Tier-3 cluster at NISER Dr. Tania Moulik What and why? Grid computing is a term referring to the combination of computer resources from multiple administrative domains to reach common goal. Grids tend

More information

The CMS Tier0 goes Cloud and Grid for LHC Run 2. Dirk Hufnagel (FNAL) for CMS Computing

The CMS Tier0 goes Cloud and Grid for LHC Run 2. Dirk Hufnagel (FNAL) for CMS Computing The CMS Tier0 goes Cloud and Grid for LHC Run 2 Dirk Hufnagel (FNAL) for CMS Computing CHEP, 13.04.2015 Overview Changes for the Tier0 between Run 1 and Run 2 CERN Agile Infrastructure (in GlideInWMS)

More information

Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF)

Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF) Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF) Gerardo Ganis CERN E-mail: Gerardo.Ganis@cern.ch CERN Institute of Informatics, University of Warsaw E-mail: Jan.Iwaszkiewicz@cern.ch

More information

Single-Top Production at the Tevatron and the LHC: Results and Prospects

Single-Top Production at the Tevatron and the LHC: Results and Prospects Single-Top Production at the Tevatron and the LHC: Results and Prospects Wolfgang Wagner Bergische Universität Wuppertal DESY Zeuthen, June 16, 2011 Content: 1) Introduction / History 2) Experimental Status

More information

Data sharing and Big Data in the physical sciences. 2 October 2015

Data sharing and Big Data in the physical sciences. 2 October 2015 Data sharing and Big Data in the physical sciences 2 October 2015 Content Digital curation: Data and metadata Why consider the physical sciences? Astronomy: Video Physics: LHC for example. Video The Research

More information

Implications of CMS searches for the Constrained MSSM A Bayesian approach

Implications of CMS searches for the Constrained MSSM A Bayesian approach Implications of CMS searches for the Constrained MSSM A Bayesian approach Małgorzata Kazana, Yue-Lin Sming Tsai On behalf of the BayesFITS group National Centre for Nuclear Research Warsaw, Poland BayesFITS,

More information

File server infrastructure @NIKHEF

File server infrastructure @NIKHEF File server infrastructure @NIKHEF CT system support June 2003 1 CT NIKHEF Outline Protocols Naming scheme (Unix, Windows) Backup and archiving Server systems Disk quota policy AFS June 2003 2 CT NIKHEF

More information

The CMS analysis chain in a distributed environment

The CMS analysis chain in a distributed environment The CMS analysis chain in a distributed environment on behalf of the CMS collaboration DESY, Zeuthen,, Germany 22 nd 27 th May, 2005 1 The CMS experiment 2 The CMS Computing Model (1) The CMS collaboration

More information

Jets energy calibration in ATLAS

Jets energy calibration in ATLAS Jets energy calibration in ATLAS V.Giangiobbe Università di Pisa INFN sezione di Pisa Supported by the ARTEMIS Research Training Network Workshop sui Monte Carlo, la Fisica e le Simulazioni a LHC V.Giangiobbe

More information

Computing Model for SuperBelle

Computing Model for SuperBelle Computing Model for SuperBelle Outline Scale and Motivation Definitions of Computing Model Interplay between Analysis Model and Computing Model Options for the Computing Model Strategy to choose the Model

More information

Real Time Tracking with ATLAS Silicon Detectors and its Applications to Beauty Hadron Physics

Real Time Tracking with ATLAS Silicon Detectors and its Applications to Beauty Hadron Physics Real Time Tracking with ATLAS Silicon Detectors and its Applications to Beauty Hadron Physics Carlo Schiavi Dottorato in Fisica - XVII Ciclo Outline The ATLAS Experiment The SiTrack Algorithm Application

More information

Accelerating Experimental Elementary Particle Physics with the Gordon Supercomputer. Frank Würthwein Rick Wagner August 5th, 2013

Accelerating Experimental Elementary Particle Physics with the Gordon Supercomputer. Frank Würthwein Rick Wagner August 5th, 2013 Accelerating Experimental Elementary Particle Physics with the Gordon Supercomputer Frank Würthwein Rick Wagner August 5th, 2013 The Universe is a strange place! 67% of energy is dark energy We got no

More information

Tier0 plans and security and backup policy proposals

Tier0 plans and security and backup policy proposals Tier0 plans and security and backup policy proposals, CERN IT-PSS CERN - IT Outline Service operational aspects Hardware set-up in 2007 Replication set-up Test plan Backup and security policies CERN Oracle

More information

Open access to data and analysis tools from the CMS experiment at the LHC

Open access to data and analysis tools from the CMS experiment at the LHC Open access to data and analysis tools from the CMS experiment at the LHC Thomas McCauley (for the CMS Collaboration and QuarkNet) University of Notre Dame, USA thomas.mccauley@cern.ch! 5 Feb 2015 Outline

More information

Computing at the HL-LHC

Computing at the HL-LHC Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory Group ALICE: Pierre Vande Vyvre, Thorsten Kollegger, Predrag Buncic; ATLAS: David Rousseau, Benedetto Gorini,

More information

Data analysis in Par,cle Physics

Data analysis in Par,cle Physics Data analysis in Par,cle Physics From data taking to discovery Tuesday, 13 August 2013 Lukasz Kreczko - Bristol IT MegaMeet 1 $ whoami Lukasz (Luke) Kreczko Par,cle Physicist Graduated in Physics from

More information

FCC 1309180800 JGU WBS_v0034.xlsm

FCC 1309180800 JGU WBS_v0034.xlsm 1 Accelerators 1.1 Hadron injectors 1.1.1 Overall design parameters 1.1.1.1 Performance and gap of existing injector chain 1.1.1.2 Performance and gap of existing injector chain 1.1.1.3 Baseline parameters

More information

Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de

Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de Tier-2 cloud Holger Marten Holger. Marten at iwr. fzk. de www.gridka.de 1 GridKa associated Tier-2 sites spread over 3 EGEE regions. (4 LHC Experiments, 5 (soon: 6) countries, >20 T2 sites) 2 region DECH

More information

Linux and the Higgs Particle

Linux and the Higgs Particle Linux and the Higgs Particle Dr. Bernd Panzer-Steindel Computing Fabric Area Manager, CERN/IT Linux World, Frankfurt 27.October 2004 Outline What is CERN The Physics The Physics Tools The Accelerator The

More information

CERN s Scientific Programme and the need for computing resources

CERN s Scientific Programme and the need for computing resources This document produced by Members of the Helix Nebula consortium is licensed under a Creative Commons Attribution 3.0 Unported License. Permissions beyond the scope of this license may be available at

More information

Big Data Processing Experience in the ATLAS Experiment

Big Data Processing Experience in the ATLAS Experiment Big Data Processing Experience in the ATLAS Experiment A. on behalf of the ATLAS Collabora5on Interna5onal Symposium on Grids and Clouds (ISGC) 2014 March 23-28, 2014 Academia Sinica, Taipei, Taiwan Introduction

More information

Top rediscovery at ATLAS and CMS

Top rediscovery at ATLAS and CMS Top rediscovery at ATLAS and CMS on behalf of ATLAS and CMS collaborations CNRS/IN2P3 & UJF/ENSPG, LPSC, Grenoble, France E-mail: julien.donini@lpsc.in2p3.fr We describe the plans and strategies of the

More information

Search for Dark Matter at the LHC

Search for Dark Matter at the LHC Search for Dark Matter at the LHC Steven Lowette Vrije Universiteit Brussel - IIHE 19 November 2014 3rd CosPa Meeting Université de Liège Content Introduction Generic DM searches at the LHC Explicit DM

More information

RO-11-NIPNE, evolution, user support, site and software development. IFIN-HH, DFCTI, LHCb Romanian Team

RO-11-NIPNE, evolution, user support, site and software development. IFIN-HH, DFCTI, LHCb Romanian Team IFIN-HH, DFCTI, LHCb Romanian Team Short overview: The old RO-11-NIPNE site New requirements from the LHCb team User support ( solution offered). Data reprocessing 2012 facts Future plans The old RO-11-NIPNE

More information

Summer Student Project Report

Summer Student Project Report Summer Student Project Report Dimitris Kalimeris National and Kapodistrian University of Athens June September 2014 Abstract This report will outline two projects that were done as part of a three months

More information

LHC Databases on the Grid: Achievements and Open Issues * A.V. Vaniachine. Argonne National Laboratory 9700 S Cass Ave, Argonne, IL, 60439, USA

LHC Databases on the Grid: Achievements and Open Issues * A.V. Vaniachine. Argonne National Laboratory 9700 S Cass Ave, Argonne, IL, 60439, USA ANL-HEP-CP-10-18 To appear in the Proceedings of the IV International Conference on Distributed computing and Gridtechnologies in science and education (Grid2010), JINR, Dubna, Russia, 28 June - 3 July,

More information

LHC GRID computing in Poland

LHC GRID computing in Poland POLAND LHC GRID computing in Poland Michał Turała IFJ PAN/ ACK Cyfronet AGH, Kraków Polish Particle ICFA Physics DDW07, Symposium, Mexicio City, Warszawa, 25.10.2007 21.04.2008 1 Outline Computing needs

More information

Space Users: Status, Requirements and Open Issues

Space Users: Status, Requirements and Open Issues Space Users: Status, Requirements and Open Issues Despite asking a number of key people, I only received 1.5 responses of user requirements... Are space users too busy or just simply happy? :-) I will

More information

EGEE is a project funded by the European Union under contract IST-2003-508833

EGEE is a project funded by the European Union under contract IST-2003-508833 www.eu-egee.org NA4 Applications F.Harris(Oxford/CERN) NA4/HEP coordinator EGEE is a project funded by the European Union under contract IST-2003-508833 Talk Outline The basic goals of NA4 The organisation

More information

IT-INFN-CNAF Status Update

IT-INFN-CNAF Status Update IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, 10-11 December 2009 Stefano Zani 10/11/2009 Stefano Zani INFN CNAF (TIER1 Staff) 1 INFN CNAF CNAF is the main computing facility of the INFN Core business:

More information

Software, Computing and Analysis Models at CDF and D0

Software, Computing and Analysis Models at CDF and D0 Software, Computing and Analysis Models at CDF and D0 Donatella Lucchesi CDF experiment INFN-Padova Outline Introduction CDF and D0 Computing Model GRID Migration Summary III Workshop Italiano sulla fisica

More information

The TOTEM experiment at the LHC: results and perspective

The TOTEM experiment at the LHC: results and perspective The TOTEM experiment at the LHC: results and perspective Edoardo Bossini Università degli studi di Siena and INFN-Pisa (on behalf of the TOTEM collaboration) Trieste, 24 Settembre 2013 OUTLINE: Detector

More information

US NSF s Scientific Software Innovation Institutes

US NSF s Scientific Software Innovation Institutes US NSF s Scientific Software Innovation Institutes S 2 I 2 awards invest in long-term projects which will realize sustained software infrastructure that is integral to doing transformative science. (Can

More information

Status of ALICE activities within FKPPL LIA

Status of ALICE activities within FKPPL LIA Status of ALICE activities within FKPPL LIA 1 Development of the online monitoring software for the ALICE Muon Trigger And Suppression Study for Υ S A N G - U N A H N K O N K U K U N I V E R S I T Y S

More information

Distributed analysis for the ATLAS Experiment in the S.Co.P.E Project

Distributed analysis for the ATLAS Experiment in the S.Co.P.E Project FINAL WORKSHOP OF GRID PROJECTS, PON RICERCA 2000-2006, AVVISO 1575 1 Distributed analysis for the ATLAS Experiment in the S.Co.P.E Project A. Aloisio 1,2, M. Alviggi 1,2, M. Biglietti 1,2, V. Canale 1,2,

More information

ALICE GRID & Kolkata Tier-2

ALICE GRID & Kolkata Tier-2 ALICE GRID & Kolkata Tier-2 Site Name :- IN-DAE-VECC-01 & IN-DAE-VECC-02 VO :- ALICE City:- KOLKATA Country :- INDIA Vikas Singhal VECC, Kolkata Events at LHC Luminosity : 10 34 cm -2 s -1 40 MHz every

More information

CLEO III Data Storage

CLEO III Data Storage CLEO III Data Storage M. Lohner 1, C. D. Jones 1, Dan Riley 1 Cornell University, USA Abstract The CLEO III experiment will collect on the order of 200 TB of data over the lifetime of the experiment. The

More information

First-year experience with the ATLAS online monitoring framework

First-year experience with the ATLAS online monitoring framework First-year experience with the ATLAS online monitoring framework A Corso-Radu 1 for the ATLAS TDAQ Collaboration 2 University of California, Department of Physics and Astronomy, 4129 Frederick Reines Hall,

More information

Introduction to Grid computing

Introduction to Grid computing Introduction to Grid computing The INFNGrid Project Team Introduction This tutorial has been implemented considering as starting point the DataGrid (EDG) tutorial Many thanks to the EDG tutorials team!

More information

Performance monitoring of the software frameworks for LHC experiments

Performance monitoring of the software frameworks for LHC experiments Performance monitoring of the software frameworks for LHC experiments William A. Romero R. wil-rome@uniandes.edu.co J.M. Dana Jose.Dana@cern.ch First EELA-2 Conference Bogotá, COL OUTLINE Introduction

More information

ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC

ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC Wensheng Deng 1, Alexei Klimentov 1, Pavel Nevski 1, Jonas Strandberg 2, Junji Tojo 3, Alexandre Vaniachine 4, Rodney

More information

Theory versus Experiment. Prof. Jorgen D Hondt Vrije Universiteit Brussel jodhondt@vub.ac.be

Theory versus Experiment. Prof. Jorgen D Hondt Vrije Universiteit Brussel jodhondt@vub.ac.be Theory versus Experiment Prof. Jorgen D Hondt Vrije Universiteit Brussel jodhondt@vub.ac.be Theory versus Experiment Pag. 2 Dangerous cocktail!!! Pag. 3 The basics in these lectures Part1 : Theory meets

More information

Integration of Virtualized Worker Nodes in Batch Systems

Integration of Virtualized Worker Nodes in Batch Systems Integration of Virtualized Worker Nodes Dr. A. Scheurer, Dr. V. Büge, O. Oberst, P. Krauß Linuxtag 2010, Berlin, Session: Cloud Computing, Talk ID: #16197 KIT University of the State of Baden-Wuerttemberg

More information

CMS Computing Model: Notes for a discussion with Super-B

CMS Computing Model: Notes for a discussion with Super-B CMS Computing Model: Notes for a discussion with Super-B Claudio Grandi [ CMS Tier-1 sites coordinator - INFN-Bologna ] Daniele Bonacorsi [ CMS Facilities Ops coordinator - University of Bologna ] 1 Outline

More information

ATLAS Data Management Accounting with Hadoop Pig and HBase

ATLAS Data Management Accounting with Hadoop Pig and HBase ATLAS Data Management Accounting with Hadoop Pig and HBase Mario Lassnig, Vincent Garonne, Gancho Dimitrov, Luca Canali, on behalf of the ATLAS Collaboration European Organization for Nuclear Research

More information

Service Challenge Tests of the LCG Grid

Service Challenge Tests of the LCG Grid Service Challenge Tests of the LCG Grid Andrzej Olszewski Institute of Nuclear Physics PAN Kraków, Poland Cracow 05 Grid Workshop 22 nd Nov 2005 The materials used in this presentation come from many sources

More information

ATLAS NOTE ATLAS-CONF-2010-063. July 21, 2010. Search for top pair candidate events in ATLAS at s = 7 TeV. The ATLAS Collaboration.

ATLAS NOTE ATLAS-CONF-2010-063. July 21, 2010. Search for top pair candidate events in ATLAS at s = 7 TeV. The ATLAS Collaboration. ATLAS NOTE ATLAS-CONF-2010-063 July 21, 2010 Search for top pair candidate events in ATLAS at s = 7 TeV The ATLAS Collaboration Abstract A search is performed for events consistent with top quark pair

More information

Werkzeuge zur verteilen Analyse im ATLAS-Experiment

Werkzeuge zur verteilen Analyse im ATLAS-Experiment Werkzeuge zur verteilen Analyse im ATLAS-Experiment Johannes Elmsheuser Ludwig-Maximilians-Universität München, Germany 9 März 27/DPG Tagung Heidelberg ATLAS Grid Infrastructure Heterogeneous grid environment

More information

ATLAS Cloud Computing and Computational Science Center at Fresno State

ATLAS Cloud Computing and Computational Science Center at Fresno State ATLAS Cloud Computing and Computational Science Center at Fresno State Cui Lin and (CS/Physics Departments, Fresno State) 2/24/2012 at CSU Chancellor s Office LHC ATLAS Tier 3 at CSUF Tier 1 France ~PByte/sec

More information

Werkzeuge zur verteilten Analyse im ATLAS-Experiment

Werkzeuge zur verteilten Analyse im ATLAS-Experiment Werkzeuge zur verteilten Analyse im ATLAS-Experiment Johannes Elmsheuser Ludwig-Maximilians-Universität München, Germany 04 März 2008/DPG Tagung Freiburg ATLAS Data replication and distribution Johannes

More information

Radiotherapy in Hungary: present status and future needs. Tibor Major, PhD National Institute of Oncology Radiotherapy Department Budapest, Hungary

Radiotherapy in Hungary: present status and future needs. Tibor Major, PhD National Institute of Oncology Radiotherapy Department Budapest, Hungary Radiotherapy in Hungary: present status and future needs Tibor Major, PhD National Institute of Oncology Radiotherapy Department Budapest, Hungary Academia Europaea Section Workshops, Bergen, 10 September,

More information

From Jet Scaling to Jet Vetos

From Jet Scaling to Jet Vetos From Jet Scaling to Jet Vetos Heidelberg DESY, 2/202 LHC Higgs analyses Two problems for LHC Higgs analyses [talks Rauch, Englert] observe H b b decays [fat Higgs jets, Marcel s talk] 2 understand jet

More information

Advancements in Big Data Processing

Advancements in Big Data Processing Advancements in Big Data Processing Alexandre Vaniachine V Interna0onal Conference "Distributed Compu0ng and Grid- technologies in Science and Educa0on" (Grid2012) July 16 21, 2012 JINR, Dubna, Russia

More information

CPU Benchmarks Over 600,000 CPUs Benchmarked

CPU Benchmarks Over 600,000 CPUs Benchmarked Shopping cart Search Home Software Hardware Benchmarks Services Store Support Forums About Us Home» CPU Benchmarks» Multiple CPU Systems CPU Benchmarks Video Card Benchmarks Hard Drive Benchmarks RAM PC

More information

Distributed Database Services - a Fundamental Component of the WLCG Service for the LHC Experiments - Experience and Outlook

Distributed Database Services - a Fundamental Component of the WLCG Service for the LHC Experiments - Experience and Outlook Distributed Database Services - a Fundamental Component of the WLCG Service for the LHC Experiments - Experience and Outlook Maria Girone CERN, 1211 Geneva 23, Switzerland Maria.Girone@cern.ch Abstract.

More information

An Affordable Commodity Network Attached Storage Solution for Biological Research Environments.

An Affordable Commodity Network Attached Storage Solution for Biological Research Environments. An Affordable Commodity Network Attached Storage Solution for Biological Research Environments. Ari E. Berman, Ph.D. Senior Systems Engineer Buck Institute for Research on Aging aberman@buckinstitute.org

More information

Data Deduplication in a Hybrid Architecture for Improving Write Performance

Data Deduplication in a Hybrid Architecture for Improving Write Performance Data Deduplication in a Hybrid Architecture for Improving Write Performance Data-intensive Salable Computing Laboratory Department of Computer Science Texas Tech University Lubbock, Texas June 10th, 2013

More information

Running a typical ROOT HEP analysis on Hadoop/MapReduce. Stefano Alberto Russo Michele Pinamonti Marina Cobal

Running a typical ROOT HEP analysis on Hadoop/MapReduce. Stefano Alberto Russo Michele Pinamonti Marina Cobal Running a typical ROOT HEP analysis on Hadoop/MapReduce Stefano Alberto Russo Michele Pinamonti Marina Cobal CHEP 2013 Amsterdam 14-18/10/2013 Topics The Hadoop/MapReduce model Hadoop and High Energy Physics

More information

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server)

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server) Scalability Results Select the right hardware configuration for your organization to optimize performance Table of Contents Introduction... 1 Scalability... 2 Definition... 2 CPU and Memory Usage... 2

More information

Report from SARA/NIKHEF T1 and associated T2s

Report from SARA/NIKHEF T1 and associated T2s Report from SARA/NIKHEF T1 and associated T2s Ron Trompert SARA About SARA and NIKHEF NIKHEF SARA High Energy Physics Institute High performance computing centre Manages the Surfnet 6 network for the Dutch

More information

Tier 1 Services - CNAF to T1

Tier 1 Services - CNAF to T1 CDF Report on Tier 1 Usage Donatella Lucchesi for the CDF Italian Computing Group INFN Padova Outline The CDF Computing Model Tier1 resources usage as today CDF portal for European GRID: lcgcaf People

More information

Legal Notices and Important Information

Legal Notices and Important Information Legal Notices and Important Information Intel processor numbers are not a measure of performance. Processor numbers differentiate features within each processor family, not across different processor families.

More information

Top-Quark Studies at CMS

Top-Quark Studies at CMS Top-Quark Studies at CMS Tim Christiansen (CERN) on behalf of the CMS Collaboration ICHEP 2010, Paris 35th International Conference on High-Energy Physics tt 2 km 22 28 July 2010 Single-top 4 km New Physics

More information

Theoretical Particle Physics FYTN04: Oral Exam Questions, version ht15

Theoretical Particle Physics FYTN04: Oral Exam Questions, version ht15 Theoretical Particle Physics FYTN04: Oral Exam Questions, version ht15 Examples of The questions are roughly ordered by chapter but are often connected across the different chapters. Ordering is as in

More information

variables to investigate Monte Carlo methods of t t production

variables to investigate Monte Carlo methods of t t production Using the M 2 and variables to investigate Monte Carlo methods of t t production Caitlin Jones September 8, 25 Abstract In this project the behaviour of Monte Carlo simulations for the event t t! ` `+b

More information

Highlights of Recent CMS Results. Dmytro Kovalskyi (UCSB)

Highlights of Recent CMS Results. Dmytro Kovalskyi (UCSB) Highlights of Recent CMS Results Dmytro Kovalskyi (UCSB) Introduction Number of CMS publication is over 0 already It is very hard to review all the recent results in one talk This talk concentrates on

More information

CERN analysis preservation (CAP) - Use Cases. Sünje Dallmeier Tiessen, Patricia Herterich, Peter Igo-Kemenes, Tibor Šimko, Tim Smith

CERN analysis preservation (CAP) - Use Cases. Sünje Dallmeier Tiessen, Patricia Herterich, Peter Igo-Kemenes, Tibor Šimko, Tim Smith CERN analysis preservation (CAP) - Use Cases Sünje Dallmeier Tiessen, Patricia Herterich, Peter Igo-Kemenes, Tibor Šimko, Tim Smith Created in April 2015, published in November 2015 Abstract In this document

More information

MANAGED STORAGE SYSTEMS AT CERN

MANAGED STORAGE SYSTEMS AT CERN MANAGED STORAGE SYSTEMS AT CERN Ingo Augustin and Fabrizio Gagliardi CERN, Geneva, Switzerland 1. INTRODUCTION Abstract The amount of data produced by the future LHC experiments requires fundamental changes

More information

Oracle Database Reliability, Performance and scalability on Intel Xeon platforms Mitch Shults, Intel Corporation October 2011

Oracle Database Reliability, Performance and scalability on Intel Xeon platforms Mitch Shults, Intel Corporation October 2011 Oracle Database Reliability, Performance and scalability on Intel platforms Mitch Shults, Intel Corporation October 2011 1 Intel Processor E7-8800/4800/2800 Product Families Up to 10 s and 20 Threads 30MB

More information

HEP GROUP UNIANDES - COLOMBIA

HEP GROUP UNIANDES - COLOMBIA HEP GROUP UNIANDES - COLOMBIA Carlos Avila On behalf of the group January 8th 2014 C. Avila, UNIANDES 1 LATIN AMERICAN COUNTRIES IN LHC EXPERIMENTS Uniandes is a CMS collaborating Institution since March

More information