HAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions
|
|
- Michael Logan
- 8 years ago
- Views:
Transcription
1 DESY Tier 2 and NAF Peter Wegner, Birgit Lewendel for DESY-IT/DV Tier 2: Status and News NAF: Status, Plans and Questions
2 Basics T2: 1.5 average Tier 2 are requested by CMS-groups for Germany Desy commitment: 1 av. Tier 2 for CMS av. Tier 2 for Atlas av. Tier 2 for LHCb Aachen commitment: 0.5 average Tier 2 Desy s Tier 2 is distributed between Hamburg and Zeuthen Birgit Lewendel 2
3 The Tier 2 is part of larger Grid@DESY Managed via Virtual Organisations (VOs) see: grid.desy.de H1/Zeus Grids Lattice Data Grid LHC-Tier 2 others Ice Cube ILC/Calice Birgit Lewendel 3
4 Grid Infrastructure and T2 ssh UI JDL output VO exclusive RB WMS LFC BDII LFC certs $HOME/.globus/ VOMS RB CE SE CE Batch GRIS GIIS GRIS SE world /etc/grid-security/grid-mapfile WN WN WN WN R-GMA HMS DESY Site2 Site Birgit Lewendel 4
5 Counts, Developments, Limits ~580 CPU cores (~770kSpecInt), 900GB, ~250TB disk All cores on SL4, 32bit (grid-ce3, grid-batch3) Several dcache SE, 1 dedicated for CMS (default SE), 1 for Atlas ~30 boxes, core services - steering/management (RB(4), WMS, VOMS, LFC,...) Grid@DESY/ZN 180 CPU cores (~280kSpecInt) 100 cores on SL3, 32bit, 80 cores on SL4, 64bit ~45TB disk dcache SE Birgit Lewendel 5
6 Counts, Developments, Limits (2) networking - wide area 1 Gb/s connecticity to XWIN officially 10 Gb/s in place (XWIN) 10 Gb/s VPN to Zeuthen pool accounts as deployed by yaim and EGEE (special treatment for CMS software deployment needed?) 2GB/core RAM, 5GB/job disk VO user support by CMS Birgit Lewendel 6
7 DESY T2 Usage 2007 Wallclock time HH+ZN lhcb dteam 6% ops icecube atlas 16% calice 7% dech hone 2 ilc 2% dcms desy dgtest geant4 cms 28% zeus 21% geant 1% hermes cms zeus hone atlas calice dteam ilc geant dgtest ops geant4 lhcb dech icecube dcms desy hermes Birgit Lewendel 7
8 DESY T2 Usage 2007 CPU time HH+ZN calice 7% ilc 3% dteam dgtest geant 1% atlas 15% hone 29% geant4 icecube dech dcms desy lhcb cms 2 hermes ops zeus 25% cms zeus hone atlas calice dteam ilc geant dgtest ops geant4 lhcb dech icecube dcms desy hermes Birgit Lewendel 8
9 DESY T2 Usage 2007 Wallclock/CPU time (h) HH+ZN Wallclock CPU cms zeus hone atlas calice dteam ilc geant dgtest ops geant4 lhcb dech icecube dcms desy hermes ATLAS/CMS: Relative high I/O fraction Birgit Lewendel 9
10 DESY T2 Usage 2007 Total Wallclock/CPU time (days) HERA vs LHC WallClock CPU atlas+cms hone+zeus Birgit Lewendel 10
11 DESY T2 Usage 2007 Wallclock time ZN ilc 2% icecube lhcb ops zeus 8% atlas calice cms dech desy dteam ghep hone 29% atlas 53% geant4 ghep hone geant4 1% dteam desy dech calice 7% cms icecube ilc lhcb ops zeus Birgit Lewendel 11
12 DESY T2 Usage 2007 CPU time ZN ilc 3% lhcb ops zeus 8% atlas icecube calice cms dech desy CMS VO Supported (!) atlas 45% dteam geant4 hone 36% ghep hone icecube ilc geant4 1% ghep dteam desy dech cms calice 7% lhcb ops zeus Birgit Lewendel 12
13 Overview National Analysis Facility (NAF) architecture (simplified) Certificate based login gsissh Batch Access Local Grid?? Interactive Batch (Sun Grid Engine SGE) proof AFS Cluster (parallel) FS dcache SRM Tier 2 grid-access Birgit Lewendel 13
14 Overview NAF (2) Distributed at DESY Zeuthen and Hamburg for users from german institutes from CMS, Atlas, LHCb and ILC. Hardware Computing: interactive server WN -> 2008 proof same blades to keep number flexible 1. order: blades with several hundred cores, 2GB/core, 2*150GB disk/blade, Infiniband Hardware Storage: probably SUN thumpers (17.5TB/box at raid 6) for dcache pools, for afs home and group space, high-performance I/O Back-up: more user input needed Birgit Lewendel 14
15 NAF Worker Node Architecture CPUs today Quad-Core Intel CPUs 5435 ( Clovertown ), if appropriate Dual-Core (5160 Woodcrest ) Per core - 2 GB main memory - 20 GB disk - 20 MB/s bandwidth for I/O applications (PROOF MB/s?) - n(?) MB/s bandwidth for disk I/O LO Management incl. digital KVM, IPMI at the minimum, service processor, interface for automatic upgrades, diagnostics RedHat certified redundant power supply (not clear whether necessary) good performance/watt ratio Service Level: 3-years On-side service NBD Format: Blade Birgit Lewendel 15
16 NAF: Org., Plans, Schedule Def. of work packages, each WP with 3-5 people and a convener OS I/O file system batch montoringinteractive software operation procurement SL4+ shared cluster file system for high-performance access SGE grid tools +qsub common afs cell afs home, group dirs login with proxy WGS env. Root, UI no desktop env. Soon (October?) meeting with users from all groups to review the plans. Pilot system should be ready to use end of November (Helmholtz Alliance kick-off meeting Dec 3 rd -5 th ) Finances: Helmholtz Alliance and DGrid Sonderinvestition Birgit Lewendel 16
17 Summary Grid Infrastructure and T2 in good shape and ready for ramp up heavily in use by several VOs shows good performance in CMS tests (thanks to UniHH T2 monitoring shift team) National Analysis Facility in preparation many open questions, items user feedback meeting soon Birgit Lewendel 17
18 Back-up Birgit Lewendel 18
19 Revised hardware resources plan (C=CTDR, N=New) for a average CMS Tier 2 (Ass.: 25 Tier 2 s) CPU [ksi2k] 300 N 600 N 1000 N 1810 N 400 C 900 C 1400 C 2300 C Disk [TB] 50 N 170 N 340 N 530 N 100 C 200 C 400 C 700 C Tape [TB] (?) Birgit Lewendel 19
20 Revised hardware resources plan ramp up (total for DESY) CPU [ksi2k] Disk [TB] Tape [TB] (?) Birgit Lewendel 20
Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de
Tier-2 cloud Holger Marten Holger. Marten at iwr. fzk. de www.gridka.de 1 GridKa associated Tier-2 sites spread over 3 EGEE regions. (4 LHC Experiments, 5 (soon: 6) countries, >20 T2 sites) 2 region DECH
More informationInternal ROC DECH Report
Internal ROC DECH Report Sven Hermann et.al. Karlsruhe Institute of Technology www.eu-egee.org EGEE-III INFSO-RI-222667 EGEE and glite are registered trademarks DECH Region U Dortmund DESY DESY Zeuthen
More informationThe Technical Realization of the National Analysis Facility
Waltraut Niepraschk Andreas Haupt Wolfgang Friebel Kai Leffhalm Götz Waschk Peter Wegner for the NAF team The Technical Realization of the National Analysis Facility Technical Seminar Zeuthen 2008-04-15
More informationDCMS Tier 2/3 prototype infrastructure
DCMS Tier 2/3 prototype infrastructure 1 Anja Vest, Uni Karlsruhe DCMS Meeting in Aachen, Overview LCG Queues/mapping set up Hardware capacities Supported software Summary DCMS overview CMS DCMS: Tier
More informationReport from SARA/NIKHEF T1 and associated T2s
Report from SARA/NIKHEF T1 and associated T2s Ron Trompert SARA About SARA and NIKHEF NIKHEF SARA High Energy Physics Institute High performance computing centre Manages the Surfnet 6 network for the Dutch
More informationThe dcache Storage Element
16. Juni 2008 Hamburg The dcache Storage Element and it's role in the LHC era for the dcache team Topics for today Storage elements (SEs) in the grid Introduction to the dcache SE Usage of dcache in LCG
More informationStatus of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan
Status of Grid Activities in Pakistan FAWAD SAEED National Centre For Physics, Pakistan 1 Introduction of NCP-LCG2 q NCP-LCG2 is the only Tier-2 centre in Pakistan for Worldwide LHC computing Grid (WLCG).
More informationThe Grid-it: the Italian Grid Production infrastructure
n 1 Maria Cristina Vistoli INFN CNAF, Bologna Italy The Grid-it: the Italian Grid Production infrastructure INFN-Grid goals!promote computational grid technologies research & development: Middleware and
More informationIntegrating a heterogeneous and shared Linux cluster into grids
Integrating a heterogeneous and shared Linux cluster into grids 1,2 1 1,2 1 V. Büge, U. Felzmann, C. Jung, U. Kerzel, 1 1 1 M. Kreps, G. Quast, A. Vest 1 2 DPG Frühjahrstagung March 28 31, 2006 Dortmund
More informationTier0 plans and security and backup policy proposals
Tier0 plans and security and backup policy proposals, CERN IT-PSS CERN - IT Outline Service operational aspects Hardware set-up in 2007 Replication set-up Test plan Backup and security policies CERN Oracle
More informationData storage services at CC-IN2P3
Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief Agenda Hardware: Storage on disk. Storage on tape. Software:
More informationComputing in High- Energy-Physics: How Virtualization meets the Grid
Computing in High- Energy-Physics: How Virtualization meets the Grid Yves Kemp Institut für Experimentelle Kernphysik Universität Karlsruhe Yves Kemp Barcelona, 10/23/2006 Outline: Problems encountered
More informationVirtualization Infrastructure at Karlsruhe
Virtualization Infrastructure at Karlsruhe HEPiX Fall 2007 Volker Buege 1),2), Ariel Garcia 1), Marcus Hardt 1), Fabian Kulla 1),Marcel Kunze 1), Oliver Oberst 1),2), Günter Quast 2), Christophe Saout
More informationRO-11-NIPNE, evolution, user support, site and software development. IFIN-HH, DFCTI, LHCb Romanian Team
IFIN-HH, DFCTI, LHCb Romanian Team Short overview: The old RO-11-NIPNE site New requirements from the LHCb team User support ( solution offered). Data reprocessing 2012 facts Future plans The old RO-11-NIPNE
More informationManagement of a Grid Infrastructure in glite with Virtualization
Management of a Grid Infrastructure in glite with Virtualization Miguel Cardenas Montes Javier Perez-Griffo Callejon Raul Ramos Pollan Manuel Rubio del Solar Ibergrid 2007 Santiago Compostela / 16 of May
More informationGridKa site report. Manfred Alef, Andreas Heiss, Jos van Wezel. www.kit.edu. Steinbuch Centre for Computing
GridKa site report Manfred Alef, Andreas Heiss, Jos van Wezel Steinbuch Centre for Computing KIT The cooperation of and Universität Karlsruhe (TH) www.kit.edu KIT? SCC? { = University ComputingCentre +
More informationCERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006
CERN local High Availability solutions and experiences Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006 1 Introduction Different h/w used for GRID services Various techniques & First
More informationIT-infrastructure in Zeuthen. Andreas Haupt DESY - DV DESY Computing-Seminar Hamburg, 6th June 2011
IT-infrastructure in Zeuthen. Andreas Haupt DESY - DV DESY Computing-Seminar Hamburg, 6th June 2011 Outline > DESY in Zeuthen an overview > The computing centre in Zeuthen > Network & communication infrastructure
More informationDynamic Extension of a Virtualized Cluster by using Cloud Resources CHEP 2012
Dynamic Extension of a Virtualized Cluster by using Cloud Resources CHEP 2012 Thomas Hauth,, Günter Quast IEKP KIT University of the State of Baden-Wuerttemberg and National Research Center of the Helmholtz
More informationSingle Sign-In User Centered Computing for High Energy Physics
Single Sign-In User Centered Computing for High Energy Physics Max Fischer, Oliver Oberst, Günter Quast, Marian Zvada ISGC 2013: March 19-22, 2013 Taipei, Taiwan INSTITUT FÜR EXPERIMENTELLE KERNPHYSIK
More informationIntegration of Virtualized Worker Nodes in Standard-Batch-Systems CHEP 2009 Prague Oliver Oberst
Integration of Virtualized Worker Nodes in Standard-Batch-Systems CHEP 2009 Prague Oliver Oberst Outline General Description of Virtualization / Virtualization Solutions Shared HPC Infrastructure Virtualization
More informationAnalisi di un servizio SRM: StoRM
27 November 2007 General Parallel File System (GPFS) The StoRM service Deployment configuration Authorization and ACLs Conclusions. Definition of terms Definition of terms 1/2 Distributed File System The
More informationALICE GRID & Kolkata Tier-2
ALICE GRID & Kolkata Tier-2 Site Name :- IN-DAE-VECC-01 & IN-DAE-VECC-02 VO :- ALICE City:- KOLKATA Country :- INDIA Vikas Singhal VECC, Kolkata Events at LHC Luminosity : 10 34 cm -2 s -1 40 MHz every
More informationLHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch. CERN, July 2007
WLCG Service Schedule LHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch WLCG Storage Workshop CERN, July 2007 Agenda The machine The experiments The service LHC Schedule Mar. Apr.
More informationDcache Support and Strategy
Helmholtz Alliance 2nd Grid Workshop HGF Mass Storage Support Group Christoph Anton Mitterer christoph.anton.mitterer@physik.uni-muenchen.de for the group Group members Filled positions Christopher Jung
More informationSun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
More informationCY-01-KIMON operational issues
CY-01-KIMON operational issues Speaker: Kyriacos Neocleous High Performance Computing systems Lab, University Of Cyprus SA1 workshop: Stockholm, Sweden, June 2007 www.eu-egee.org Presentation outline (1)
More informationSPACI & EGEE LCG on IA64
SPACI & EGEE LCG on IA64 Dr. Sandro Fiore, University of Lecce and SPACI December 13 th 2005 www.eu-egee.org Outline EGEE Production Grid SPACI Activity Status of the LCG on IA64 SPACI & EGEE Farm Configuration
More informationEstablishing Applicability of SSDs to LHC Tier-2 Hardware Configuration
Establishing Applicability of SSDs to LHC Tier-2 Hardware Configuration A CHEP 2010 presentation by: Sam Skipsey and The GridPP Storage Group With particular acknowledgments to: Wahid Bhimji (go see his
More informationMichael Thomas, Dorian Kcira California Institute of Technology. CMS Offline & Computing Week
Michael Thomas, Dorian Kcira California Institute of Technology CMS Offline & Computing Week San Diego, April 20-24 th 2009 Map-Reduce plus the HDFS filesystem implemented in java Map-Reduce is a highly
More informationATLAS GridKa T1/T2 Status
ATLAS GridKa T1/T2 Status GridKa TAB, FZK, 19 Oct 2007 München GridKa T1/T2 status Production and data management operations Computing team & cloud organization T1/T2 meeting summary Site monitoring/gangarobot
More informationA quantitative comparison between xen and kvm
Home Search Collections Journals About Contact us My IOPscience A quantitative comparison between xen and kvm This content has been downloaded from IOPscience. Please scroll down to see the full text.
More informationA Web-based Portal to Access and Manage WNoDeS Virtualized Cloud Resources
A Web-based Portal to Access and Manage WNoDeS Virtualized Cloud Resources Davide Salomoni 1, Daniele Andreotti 1, Luca Cestari 2, Guido Potena 2, Peter Solagna 3 1 INFN-CNAF, Bologna, Italy 2 University
More informationResume. Wenjing. Date of birth: June 11th, 1982 Nationality: Chinese Phone number: 8610-88236012-608 Cell phone: 13366466802 wuwj@ihep.ac.
Resume Personal information First name: Wenjing surname: Wu Gender: Female Date of birth: June 11th, 1982 Nationality: Chinese Phone number: 8610-88236012-608 Cell phone: 13366466802 Email: wuwj@ihep.ac.cn
More informationVirtual Server and Storage Provisioning Service. Service Description
RAID Virtual Server and Storage Provisioning Service Service Description November 28, 2008 Computer Services Page 1 TABLE OF CONTENTS INTRODUCTION... 4 VIRTUAL SERVER AND STORAGE PROVISIONING SERVICE OVERVIEW...
More informationGrid Computing in Aachen
GEFÖRDERT VOM Grid Computing in Aachen III. Physikalisches Institut B Berichtswoche des Graduiertenkollegs Bad Honnef, 05.09.2008 Concept of Grid Computing Computing Grid; like the power grid, but for
More informationMaurice Askinazi Ofer Rind Tony Wong. HEPIX @ Cornell Nov. 2, 2010 Storage at BNL
Maurice Askinazi Ofer Rind Tony Wong HEPIX @ Cornell Nov. 2, 2010 Storage at BNL Traditional Storage Dedicated compute nodes and NFS SAN storage Simple and effective, but SAN storage became very expensive
More informationHardware and Software Requirements. Release 7.5.x PowerSchool Student Information System
Release 7.5.x PowerSchool Student Information System Released October 2012 Document Owner: Documentation Services This edition applies to Release 7.5.x of the PowerSchool software and to all subsequent
More informationEvolution of the Italian Tier1 (INFN-T1) Umea, May 2009 Felice.Rosso@cnaf.infn.it
Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009 Felice.Rosso@cnaf.infn.it 1 In 2001 the project of the Italian Tier1 in Bologna at CNAF was born. First computers were based on Intel Pentium III
More informationSun Grid Engine, a new scheduler for EGEE
Sun Grid Engine, a new scheduler for EGEE G. Borges, M. David, J. Gomes, J. Lopez, P. Rey, A. Simon, C. Fernandez, D. Kant, K. M. Sephton IBERGRID Conference Santiago de Compostela, Spain 14, 15, 16 May
More informationBetriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil
Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Volker Büge 1, Marcel Kunze 2, OIiver Oberst 1,2, Günter Quast 1, Armin Scheurer 1 1) Institut für Experimentelle
More informationTHE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC
THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC The Right Data, in the Right Place, at the Right Time José Martins Storage Practice Sun Microsystems 1 Agenda Sun s strategy and commitment to the HPC or technical
More informationComputing at the HL-LHC
Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory Group ALICE: Pierre Vande Vyvre, Thorsten Kollegger, Predrag Buncic; ATLAS: David Rousseau, Benedetto Gorini,
More informationPerformance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage
Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage Technical white paper Table of contents Executive summary... 2 Introduction... 2 Test methodology... 3
More information3 Red Hat Enterprise Linux 6 Consolidation
Whitepaper Consolidation EXECUTIVE SUMMARY At this time of massive and disruptive technological changes where applications must be nimbly deployed on physical, virtual, and cloud infrastructure, Red Hat
More informationScientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015 Index - Storage use cases - Bluearc - Lustre - EOS - dcache disk only - dcache+enstore Data distribution by solution
More informationMicrosoft Hyper-V chose a Primary Server Virtualization Platform
Roger Shupert, Integration Specialist } Lake Michigan College has been using Microsoft Hyper-V as it s primary server virtualization platform since 2008, in this presentation we will discuss the following;
More informationRecommendations for Static Firewall Configuration in D-Grid
D-Grid Integrationsprojekt (DGI-2) Fachgebiet 3-3 Firewalls Recommendations for Static Firewall Configuration in D-Grid Version 1.5, 21. Mai 2008 D-Grid Integrationsprojekt (DGI-2) Autoren: Gian Luca Volpato
More informationHigh Performance Computing Infrastructure at DESY
High Performance Computing Infrastructure at DESY Sven Sternberger & Frank Schlünzen High Performance Computing Infrastructures at DESY DV-Seminar / 04 Feb 2013 Compute Infrastructures at DESY - Outline
More informationBuilding a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
More informationDSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group
DSS High performance storage pools for LHC Łukasz Janyst on behalf of the CERN IT-DSS group CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Introduction The goal of EOS is to provide a
More informationMass Storage System for Disk and Tape resources at the Tier1.
Mass Storage System for Disk and Tape resources at the Tier1. Ricci Pier Paolo et al., on behalf of INFN TIER1 Storage pierpaolo.ricci@cnaf.infn.it ACAT 2008 November 3-7, 2008 Erice Summary Tier1 Disk
More informationHP reference configuration for entry-level SAS Grid Manager solutions
HP reference configuration for entry-level SAS Grid Manager solutions Up to 864 simultaneous SAS jobs and more than 3 GB/s I/O throughput Technical white paper Table of contents Executive summary... 2
More informationRemoving Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
More informationVirtualization, Grid, Cloud: Integration Paths for Scientific Computing
Virtualization, Grid, Cloud: Integration Paths for Scientific Computing Or, where and how will my efficient large-scale computing applications be executed? D. Salomoni, INFN Tier-1 Computing Manager Davide.Salomoni@cnaf.infn.it
More informationMichael Kagan. michael@mellanox.com
Virtualization in Data Center The Network Perspective Michael Kagan CTO, Mellanox Technologies michael@mellanox.com Outline Data Center Transition Servers S as a Service Network as a Service IO as a Service
More informationOracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers
Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers Masood Ahmed EMEA Infrastructure Solutions Oracle/SAP Relationship Overview First SAP R/3 release
More informationStatus and Evolution of ATLAS Workload Management System PanDA
Status and Evolution of ATLAS Workload Management System PanDA Univ. of Texas at Arlington GRID 2012, Dubna Outline Overview PanDA design PanDA performance Recent Improvements Future Plans Why PanDA The
More informationIntroduction to Cloud Computing
Introduction to Cloud Computing Cloud Computing II (Qloud) 15 319, spring 2010 3 rd Lecture, Jan 19 th Majd F. Sakr Lecture Motivation Introduction to a Data center Understand the Cloud hardware in CMUQ
More informationJUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
More informationKIT Site Report. Andreas Petzold. www.kit.edu STEINBUCH CENTRE FOR COMPUTING - SCC
KIT Site Report Andreas Petzold STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association www.kit.edu GridKa Tier 1 - Batch
More informationcity cloud - the leading kvm based iaas in europe
city cloud - the leading kvm based iaas in europe How City Network created one of the leading IaaS in Europe using KVM and op5 Author: Özgür Bal - City Network Andreas Månsson - op5 AB 2014 Content Executive
More informationScheduling and Load Balancing in the Parallel ROOT Facility (PROOF)
Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF) Gerardo Ganis CERN E-mail: Gerardo.Ganis@cern.ch CERN Institute of Informatics, University of Warsaw E-mail: Jan.Iwaszkiewicz@cern.ch
More informationwww.see-grid-sci.eu Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009
SEE-GRID-SCI Virtualization and Grid Computing with XEN www.see-grid-sci.eu Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009 Milan Potocnik University
More informationIntegration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept
Integration of Virtualized Workernodes in Batch Queueing Systems, Dr. Armin Scheurer, Oliver Oberst, Prof. Günter Quast INSTITUT FÜR EXPERIMENTELLE KERNPHYSIK FAKULTÄT FÜR PHYSIK KIT University of the
More informationPreview of a Novel Architecture for Large Scale Storage
Preview of a Novel Architecture for Large Scale Storage Andreas Petzold, Christoph-Erdmann Pfeiler, Jos van Wezel Steinbuch Centre for Computing STEINBUCH CENTRE FOR COMPUTING - SCC KIT University of the
More informationSage 300 ERP 2014 Compatibility guide
Sage 300 ERP 2014 Compatibility guide 03 17 2014 Table of contents 1.0 Overview 3 1.1 Overview 3 1.1.1 Unlisted platforms are not supported 3 1.1.2 Product updates and program fixes 3 1.1.3 Compatibility
More informationAn Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
More informationGridKa: Roles and Status
GridKa: Roles and Status GmbH Institute for Scientific Computing P.O. Box 3640 D-76021 Karlsruhe, Germany Holger Marten http://www.gridka.de History 10/2000: First ideas about a German Regional Centre
More informationForschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Global Grid User Support - GGUS - within the LCG & EGEE environment
Global Grid User Support - GGUS - within the LCG & EGEE environment Abstract: For very large projects like the LHC Computing Grid Project (LCG) involving some 8,000 scientists from universities and laboratories
More informationIT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez
IT of SPIM Data Storage and Compression EMBO Course - August 27th Jeff Oegema, Peter Steinbach, Oscar Gonzalez 1 Talk Outline Introduction and the IT Team SPIM Data Flow Capture, Compression, and the Data
More informationDragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers
Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers This section includes system requirements for DMENE Network configurations that utilize virtual
More informationBatch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb
Batch and Cloud overview Andrew McNab University of Manchester GridPP and LHCb Overview Assumptions Batch systems The Grid Pilot Frameworks DIRAC Virtual Machines Vac Vcycle Tier-2 Evolution Containers
More informationUsing VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems
Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems Applied Technology Abstract By migrating VMware virtual machines from one physical environment to another, VMware VMotion can
More informationN-central 8.0 On-Premise Software and N-compass 3.1 Advanced Reporting Software
System Requirements N-central 8.0 On-Premise Software and N-compass 3.1 Advanced Reporting Software The following describes the server requirements for N-central 8.0 and its optional advanced reporting
More informationForschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction
dcache Introduction Forschungszentrum Karlsruhe GmbH Institute for Scientific Computing P.O. Box 3640 D-76021 Karlsruhe, Germany Dr. http://www.gridka.de What is dcache? Developed at DESY and FNAL Disk
More informationAdvanced Techniques with Newton. Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011
Advanced Techniques with Newton Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011 Workshop Goals Gain independence Executing your work Finding Information Fixing Problems Optimizing Effectiveness
More informationCNR-INFM DEMOCRITOS and SISSA elab Trieste
elab and the FVG grid Stefano Cozzini CNR-INFM DEMOCRITOS and SISSA elab Trieste Agenda/Aims Present elab ant its computational infrastructure GRID-FVG structure basic requirements technical choices open
More informationIOS110. Virtualization 5/27/2014 1
IOS110 Virtualization 5/27/2014 1 Agenda What is Virtualization? Types of Virtualization. Advantages and Disadvantages. Virtualization software Hyper V What is Virtualization? Virtualization Refers to
More informationSAP Business One Hardware Requirements Guide
Hardware Requirements Guide Document Version: 1.04 2014-08-25 Release 8.8 and higher Typographic Conventions Type Style Example Description Words or characters quoted from the screen. These include field
More informationVirtualisation Cloud Computing at the RAL Tier 1. Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013
Virtualisation Cloud Computing at the RAL Tier 1 Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013 Virtualisation @ RAL Context at RAL Hyper-V Services Platform Scientific Computing Department
More informationAgenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.
Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance
More informationManaged Storage @ GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team
Managed Storage @ GRID or why NFSv4.1 is not enough Tigran Mkrtchyan for dcache Team What the hell do physicists do? Physicist are hackers they just want to know how things works. In moder physics given
More informationSystem Requirements Table of contents
Table of contents 1 Introduction... 2 2 Knoa Agent... 2 2.1 System Requirements...2 2.2 Environment Requirements...4 3 Knoa Server Architecture...4 3.1 Knoa Server Components... 4 3.2 Server Hardware Setup...5
More informationVBLOCK SOLUTION FOR SAP: SAP APPLICATION AND DATABASE PERFORMANCE IN PHYSICAL AND VIRTUAL ENVIRONMENTS
Vblock Solution for SAP: SAP Application and Database Performance in Physical and Virtual Environments Table of Contents www.vce.com V VBLOCK SOLUTION FOR SAP: SAP APPLICATION AND DATABASE PERFORMANCE
More informationStatus and Integration of AP2 Monitoring and Online Steering
Status and Integration of AP2 Monitoring and Online Steering Daniel Lorenz - University of Siegen Stefan Borovac, Markus Mechtel - University of Wuppertal Ralph Müller-Pfefferkorn Technische Universität
More informationInteract Intranet Version 7. Technical Requirements. August 2014. 2014 Interact
Interact Intranet Version 7 Technical Requirements August 2014 2014 Interact Definitions... 3 Licenses... 3 On-Premise... 3 Cloud... 3 Pulic Cloud... 3 Private Cloud... 3 Perpetual... 3 Self-Hosted...
More informationSage ERP Accpac. Compatibility Guide Version 6.0. Revised: November 18, 2010. Version 6.0 Compatibility Guide
Sage ERP Accpac Compatibility Guide Version 6.0 Revised: November 18, 2010 Version 6.0 Compatibility Guide i Contents Overview... 1 Version 6.0 Compatibility... 2 All Environments... 2 Virtual Environments...
More information2009 Oracle Corporation 1
The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material,
More informationDatabase Services for Physics @ CERN
Database Services for Physics @ CERN Deployment and Monitoring Radovan Chytracek CERN IT Department Outline Database services for physics Status today How we do the services tomorrow? Performance tuning
More informationSMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
More informationMeasurement of BeStMan Scalability
Measurement of BeStMan Scalability Haifeng Pi, Igor Sfiligoi, Frank Wuerthwein, Abhishek Rana University of California San Diego Tanya Levshina Fermi National Accelerator Laboratory Alexander Sim, Junmin
More informationDistributed Computing for CEPC. YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep. 12-13, 2014 1 Outline Introduction Experience of BES-DIRAC Distributed
More informationIntegrated Grid Solutions. and Greenplum
EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving
More informationBuilding Clusters for Gromacs and other HPC applications
Building Clusters for Gromacs and other HPC applications Erik Lindahl lindahl@cbr.su.se CBR Outline: Clusters Clusters vs. small networks of machines Why do YOU need a cluster? Computer hardware Network
More informationPRIMERGY server-based High Performance Computing solutions
PRIMERGY server-based High Performance Computing solutions PreSales - May 2010 - HPC Revenue OS & Processor Type Increasing standardization with shift in HPC to x86 with 70% in 2008.. HPC revenue by operating
More informationCapacity Management for Oracle Database Machine Exadata v2
Capacity Management for Oracle Database Machine Exadata v2 Dr. Boris Zibitsker, BEZ Systems NOCOUG 21 Boris Zibitsker Predictive Analytics for IT 1 About Author Dr. Boris Zibitsker, Chairman, CTO, BEZ
More informationHIP Computing Resources for LHC-startup
HIP Computing Resources for LHC-startup Tomas Lindén Finnish CMS meeting in Kumpula 03.10. 2007 Kumpula, Helsinki October 3, 2007 1 Tomas Lindén Contents 1. Finnish Tier-1/2 computing in 2007 and 2008
More informationPES. Batch virtualization and Cloud computing. Part 1: Batch virtualization. Batch virtualization and Cloud computing
Batch virtualization and Cloud computing Batch virtualization and Cloud computing Part 1: Batch virtualization Tony Cass, Sebastien Goasguen, Belmiro Moreira, Ewan Roche, Ulrich Schwickerath, Romain Wartel
More informationECLIPSE Performance Benchmarks and Profiling. January 2009
ECLIPSE Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox, Schlumberger HPC Advisory Council Cluster
More information