Integrating a heterogeneous and shared Linux cluster into grids



Similar documents
Computing in High- Energy-Physics: How Virtualization meets the Grid

DCMS Tier 2/3 prototype infrastructure

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil

Dynamic Extension of a Virtualized Cluster by using Cloud Resources CHEP 2012

The dcache Storage Element

Virtualization of a Cluster Batch System

Virtualization Infrastructure at Karlsruhe

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Global Grid User Support - GGUS - within the LCG & EGEE environment

Global Grid User Support - GGUS - in the LCG & EGEE environment

Mass Storage at GridKa

Experiences with the GLUE information schema in the LCG/EGEE production Grid

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Basics of Virtualisation

HAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions

NT1: An example for future EISCAT_3D data centre and archiving?

Scientific Storage at FNAL. Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015

Software, Computing and Analysis Models at CDF and D0

The Grid-it: the Italian Grid Production infrastructure

Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009

Report from SARA/NIKHEF T1 and associated T2s

Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept

Data storage services at CC-IN2P3

Virtual Server and Storage Provisioning Service. Service Description

The ENEA gateway approach providing EGEE/gLite access to unsupported platforms and operating systems

Recommendations for Static Firewall Configuration in D-Grid

High Availability Databases based on Oracle 10g RAC on Linux

This presentation provides an overview of the architecture of the IBM Workload Deployer product.

Symantec NetBackup Appliances

GridKa: Roles and Status

IBM Virtual Server Services. A smarter way to support and grow your business

Virtual Environments for Prototyping Tier-3 Clusters

GridKa site report. Manfred Alef, Andreas Heiss, Jos van Wezel. Steinbuch Centre for Computing

Single Sign-In User Centered Computing for High Energy Physics

dcache, Software for Big Data

CERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN

Global Grid User Support - GGUS - start up schedule

Status and Integration of AP2 Monitoring and Online Steering

Analisi di un servizio SRM: StoRM

The CMS analysis chain in a distributed environment

Integration of Virtualized Worker Nodes in Standard-Batch-Systems CHEP 2009 Prague Oliver Oberst

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

The GRID and the Linux Farm at the RCF

HADOOP, a newly emerged Java-based software framework, Hadoop Distributed File System for the Grid

Addressing research data challenges at the. University of Colorado Boulder

GRIDSEED: A Virtual Training Grid Infrastructure

The GridKa Installation for HEP Computing

Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware

GRIDSEED: A Virtual Training Grid Infrastructure

Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

The ENEA-EGEE site: Access to non-standard platforms

CMS Software Deployment on OSG

FermiGrid Highly Available Grid Services

Cluster, Grid, Cloud Concepts

Installing, Running and Maintaining Large Linux Clusters at CERN

Virtualisation Cloud Computing at the RAL Tier 1. Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013

OIS. Update on Windows 7 at CERN & Remote Desktop Gateway. Operating Systems & Information Services CERN IT-OIS

OVERVIEW. CEP Cluster Server is Ideal For: First-time users who want to make applications highly available

How To Understand The Architecture Of An Ulteo Virtual Desktop Server Farm

Instruments in Grid: the New Instrument Element

Towards a Comprehensive Accounting Solution in the Multi-Middleware Environment of the D-Grid Initiative

Roadmap for Applying Hadoop Distributed File System in Scientific Grid Computing

An approach to grid scheduling by using Condor-G Matchmaking mechanism

Grid Computing in Aachen

CLOUD COMPUTING AS THE BASIS OF ELECTRONIC UNIVERSITY CAMPUS. Gmar, D., Kryukov, V., Shakhgeldyan, K. Vladivostok State University of Economics

Solution for private cloud computing

SPRACE Site Report. Guilherme Amadio SPRACE UNESP

CMS Dashboard of Grid Activity

Network & HEP Computing in China. Gongxing SUN CJK Workshop & CFI

Stateless Compute Cluster

Installation der Software des LHC Computing Grid auf einem Institutscluster. Hartmut Stadie, Christopher Jung, Günter Quast, Klaus Rabbertz, Jens Rehn

Integration of Virtualized Worker Nodes in Batch Systems

University of Huddersfield Repository

Data Management. Network transfers

A Physics Approach to Big Data. Adam Kocoloski, PhD CTO Cloudant

File Transfer Software and Service SC3

Tier0 plans and security and backup policy proposals

U-LITE Network Infrastructure

Building a Linux Cluster

RO-11-NIPNE, evolution, user support, site and software development. IFIN-HH, DFCTI, LHCb Romanian Team

IPv6 Traffic Analysis and Storage

How To Speed Up A Flash Flash Storage System With The Hyperq Memory Router

Das HappyFace Meta-Monitoring Framework

VDI: What Does it Mean, Deploying challenges & Will It Save You Money?


Plateforme de Calcul pour les Sciences du Vivant. SRB & glite. V. Breton.

Hadoop Architecture. Part 1

Tier 1 Services - CNAF to T1

An Oracle White Paper July Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide

IT-INFN-CNAF Status Update

Measurement of BeStMan Scalability

RED HAT ENTERPRISE VIRTUALIZATION

Chapter 12 Distributed Storage

Evolution of the Italian Tier1 (INFN-T1) Umea, May 2009

MAKING YOUR VIRTUAL INFRASTUCTURE NON-STOP Making availability efficient with Veritas products

Computing at the HL-LHC

System Requirements Table of contents

CNR-INFM DEMOCRITOS and SISSA elab Trieste

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Improvement Options for LHC Mass Storage and Data Management

Transcription:

Integrating a heterogeneous and shared Linux cluster into grids 1,2 1 1,2 1 V. Büge, U. Felzmann, C. Jung, U. Kerzel, 1 1 1 M. Kreps, G. Quast, A. Vest 1 2 DPG Frühjahrstagung March 28 31, 2006 Dortmund

Outline Why do university groups need grid computing? Computing environments at universities Representative example: IEKP cluster at Karlsruhe Peculiarities of a typical university cluster Site specific grid services The IEKP LCG site Conclusion and Outlook Many thanks to: the BMBF for financial support, the Universitäts-Rechenzentrum Karlsruhe for technical support, the IEKP system administrators and the dcms LCG administrators DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 2

Why do university groups need grid computing? Current and future HEP experiments at Tevatron or LHC: huge data production rates and event sizes Worldwide distributed datasets already now: Simulated data in the LHC experiments, O(100 TB) Real data in CDF, D0, H1, ZEUS, Babar, etc., O(1 PB) Processing power widely available in associated institutes Collaborating groups cope with these challenges using grid tools Opportunistic/shared use of the resources between local users and grid users Benefits of integrating an institute's cluster into grids: Minimisation of idle times Interception of peak loads Shared data storage Shared deployment effort of common services DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 3

Computing environment at universities A typical university computer cluster has to cope with diverse challenges: Heterogeneous structure in: hardware, software, funding and ownership Support of multiple groups with different applications and sometimes conflicting interests Infrastructural facilities have grown in the course of time characteristic history and resulting inhomogeneities Embedded in structures imposed by institute, faculty and university Integration into existing grids not easy at all! Idea of sharing resources still not present in all minds DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 4

Example: IEKP cluster at Karlsruhe Representative example: IEKP (Institut für Exp. Kernphysik) at the University of Karlsruhe integrated in: SAMGrid (Sequential Access via Metadata Grid) for CDF and LCG (LHC Computing Grid) for CMS Local users in working groups CDF CMS AMS SAMGrid LCG heterogeneous environment in: software & hardware local and grid users access policies grid middleware DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 5

Example: IEKP at Karlsruhe Cluster component specifications: One or more portal machines for each experiment (3 for CDF, 1 for CMS, 1 for AMS) 5 file servers, ~ 15 20 TB disk space 27 Computing Nodes, 36 CPU's Linux cluster independent of desktop cluster IEKP (Tier-3) GridKa/FZK (Tier-1) DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 6

Architecture of the IEKP Linux cluster DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 7

Peculiarities of a typical university cluster Network architecture: Inner network: computing nodes, file servers, cluster control machine Outer network: publicly accessible portals development of analysis software connection to inner network access to file servers usage of Worker Nodes via the local batch system Network protocols and services: User accounts exported by the cluster control machine to all nodes via Network Information Service (NIS) File/root systems exported via Network File System (NFS) Other supported protocols: GSIFTP, SRM DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 8

Peculiarities of a typical university cluster Local batch system: PBS/Torque (Open Source) Scheduler: MAUI fair share principle (group and user fair share) Batch queues Firewall Grid components behind the firewall of the institute Allow external access to grid services: some ports of the firewall have to be opened Internal campus net: protected by the university's computing department (switched off for IEKP cluster) Desktop cluster User wokstation and access point to the portal machines IEKP Linux cluster: connected to the desktop network by a 1 GBit connection DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 9

Operating systems Software on portal machines and Worker Nodes (WNs): experiment dependent Operating system on all machines: Linux, but the flavour is not identical on all components (due to experiment specific extensions/modifications) CDF portals: Linux distribution based on Fermi RedHat 7.3 CMS portals: Scientific Linux CERN 3 WNs: Scientific Linux CERN 3 (the only OS under which all AMS, CDF, CMS software and grid software runs, at least for the moment) No major problems occurred running different Linux distributions on the same cluster! WNs: 32-bit operating system; upgrade to a 64-bit operating system foreseen DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 10

Site specific grid services (LCG) Layers of abstraction, local peculiarities irrelevant LCG: Computing Element (CE) Gateway to local computing resources, i.e. Worker Nodes via local batch system Globus Gatekeeper Storage Element (SE) Gateway to local storage (disk, tape) Globus GridFTP server, SRM interface Monitoring Box (MON) User Interface (UI) User's access point to the grid and portal for (grid) file access Client programs using grid services Could also be a desktop machine SAM: Only one dedicated machine needed per cluster (portal machine) DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 11

The IEKP LCG site IEKP site: Tier-2/3 prototype centre within LCG Dedicated configuration IEKP offers full grid functionality: Grid based physics analyses Software installation for grid and local users Data storage No Virtual Organisation Management Service (VOMS) yet but: need for a prioritisation of certain user groups! Present workaround: treatment of grid users using the IEKP capacities depending on their affiliation organised by Mapping to different accounts and user groups Configuration of the corresponding batch queues, e.g. different queue priorities managed by fair share targets: Local CMS users (cms) local accounts German CMS users (dcms) generic mapping accounts CMS Collaboration (cmsgrid) generic mapping accounts DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 12

Conclusion & Outlook Computer clusters at universities: often shared and situated in a heterogeneous environment but: beneficial integration into computing grids is possible IEKP cluster successfully integrated in SAMGrid and LCG without compromising local user groups (IEKP-KA/2006-3) Whishlist: Precondition: Motivate university groups to share their free computing resources within grid environments Mechanisms for priorisation, accounting and charge-back ( billing ) Operating System independence of experiment specific and grid software, in particular the LCG middleware Lightweight and non invasive LCG installation procedure on existing computing environments needed Virtualisation LCG hosts on a single powerful server (ongoing at IEKP): improves utilisation of resources and security aspects Worker Nodes: may cope with significant differences of the Linux flavours required by certain user groups DPG 2006, March 28-31 Anja Vest, University of Karlsruhe 13