Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Global Grid User Support - GGUS - within the LCG & EGEE environment

Size: px
Start display at page:

Download "Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Global Grid User Support - GGUS - within the LCG & EGEE environment"

Transcription

1 Global Grid User Support - GGUS - within the LCG & EGEE environment

2 Abstract: For very large projects like the LHC Computing Grid Project (LCG) involving some 8,000 scientists from universities and laboratories all around the world, it is an indispensable requirement to have a well organized user support. The Institute for Scientific Computing at the Forschungszentrum Karlsruhe started implementing a Global Grid User Support (GGUS) after official assignment of the Grid Deployment Board in March For this purpose a web portal together with a helpdesk application has been developed. As a single entry point for all Grid related issues and problems GGUS follows the objectives of providing news, documentation and status information about Grid resources. The user will find forms to submit and track service requests. GGUS collaborates with different support teams in the Grid environment like the Grid Operations Center and the Experiment Specific Support. They can access the helpdesk system via web interface. GGUS stores all the incoming trouble tickets and outgoing solutions in a central database and plans to build up a knowledge base where all the information can be offered in a structured manner. As a prototype GGUS started operation at the Forschungszentrum Karlsruhe in October 2003 and supported local user groups of the German Tier 1 Computing Center, called GridKa. 4 month later the GGUS system was opened for the LCG community and since March 2004 ASCC Taiwan is the second GGUS partner. Due to the time difference between the two support centers GGUS is operational 15 hours per day now. A major target for GGUS is to be available 24 hours a day, 7 days a week. Therefore we are trying to find a qualified third partner in the US time zone. In the talk the GGUS system will be explained and integration into the Grid environment will be explained.

3 About the Author Name: Wolfgang Thöne Born 1965, living in Linkenheim near Karlsruhe Working at the Forschungszentrum Karlsruhe, Institute for Science Computing. He is member of the Global Grid User Support team, that is established since October 2002 within the LCG project. He leads the concept phase of the project and is currently working on the concepts for user support within the EGEE project.

4 Outline Support? GGUS concept and objectives Support teams Current status of GGUS web portal and helpdesk system Outlook and next steps GGUS staff

5 Forschungszentrum One of the biggest research institutions in Europe, 3500 employees 5 Research areas: - Structure of Matter - Earth and Environment - Health -Energy - Key Technologies (Nanotechnology, Grid Computing) Institute for Scientific Computing: - providing IT services - establishing GridKa, GGUS

6 Forschungszentrum Karlsruhe International Grid Projects Through links with sister projects, there is the potential for a truely global scientific applications grid

7 Goal EGEE manifesto: Enabling Grids for E-science in Europe Create a wide European Grid production quality infrastructure on top of present and future EU RN infrastructure ERA Applications Build on EU and EU member states major investments in Grid Technology International connections (US and AP) Several pioneering prototype results Larg Grid development teams in EU Requires major EU funding effort Approach EGEE Leverage current and planned national and regional Grid programmes Geant Geant Research network Network Work closely with relevant industrial Grid developers, NRENs and US-AP projects Grid infrastructure

8 Operations Center Regional Support Regional Support GGUS Regional Support Resource Center (Processors, disks) Regional Support Center (Support for Applications Local Resources) Grid server Nodes Resource Center Resource Center Resource Center EGEE Operations Structure

9 Why Support? New support groups Network layer Resource centers CIC / GOC / etc. more to come New Support challenges: Jobs are spread over several locations New types of Job and Data access Often more than one group is involved Virtual Organizations

10 Why Support? Need to have a common platform for: Collect service requests Exchange service requests Document solutions Need to document: Quality feedback To the VO s Middleware development To the Grid Operations FAQ s Knowledge Base

11 Concept of Global Grid User Support - GGUS Target: 24 7 support via time difference and 3 support teams Currently: GGUS FZK GGUS ASCC Planned: GGUS USA

12 Objectives Single point of contact and information for Grid related questions (by now for LCG and EGEE) FAQ s Downloads News and problems Support entry point for 8 HEP experiments (and other Grid user communities, EGEE) Operate a central trouble ticket database and system Design, implementation and operation of a knowledge base Be the last resort for all support requests

13 Support Teams within LCG & EGEE Deployment Support (DS) Middleware Problems Operations Center (CIC / GOC / ROC) Operations Problems Resource Centers (RC) Hardware Problems Global Grid User Support (GGUS) Single Point of Contact Coordination of UserSupport Experiment Specific User Support (ESUS) VO spec. (Software) Problems Other Communities (VOs), e.g. EGEE LHC experiments (Alice Atlas CMS LHCb) non-lhc experiments (BaBar, CDF, Compass, D0)

14 Support Workflow (LCG) ESUS Application problem GGUS DS Middleware problem User CIC / ROC +Resource Center Operation problem

15 User Problem Process model Support request Interaction Support request GGUS Support Feedback & Solution Support request Interaction ESUS Red lines: Interface the GGUS support system and/or supply an interface to GGUS system DS CIC/ROC Interaction Resource Center

16 Process model Update Web Portal Info Interaction GGUS Support Interaction Create Solve Ticket ESUS Red lines: Interface the GGUS support system and/or supply an interface to GGUS system DS CIC/ROC Interaction Resource Center

17 Architecture User Resources and Status Central User DB VO DB Web-Interface for user interaction Middleware & Interfaces GGUS Helpdesk Application (Remedy) Reporting & Analysis Workflow engine Middleware & Interfaces Web-Interface & JAVA Client for support interaction Universal Database for for GRID GRID related related problems ESUS, DS, GOC/ROC

18 Application model SuSE Linux 9.x Apache 1.3.x with mod_ssl Tomcat 4.1.x (for Remedy 6.0) MySQL 4.0.x PHP 4.x Remedy 5.1 Server & Middleware Next Release will base on Remedy 6.0 Remedy 5.1 C/C++ API Next Release will base on Remedy 6.0 Oracle OCI 9.x with Oracle in a cluster environment

19 Application model access via Remedy Client access via web-browser Apache webserver PHP Oracle/Remedy Interface Remedy API Interface Remedy Mid-Tier Application Remedy Client Remedy Application Oracle database read

20 Current status GGUS team 4 people at FZK 5 people at ASCC Teamed up with ASCC since April 2004 Support time from 02:00 16:30 CEST

21 Current status Helpdesk application Release version 1.0 is deployed and operational Workflows and escalations are running Application is accessible via web interface and via windows client User authentication via certificates and userid/password are in place DS, GOC/ROC and ESUS may use a central application via the Web

22 Current status Web Portal is online since October 2003 User interface to submit, modify and track tickets is deployed and operational Rudimental news information system is in place First version of FAQ system and a documentation section is online Monitoring and status information is available for LCG-2 Sites and local operations (GridKa)

23 Screenshots GGUS-website Home GGUS-website Submit a Service Request Remedy web view for support staff

24

25

26

27 Outlook New Helpdesk application (Remedy release 6.0) Enhancements based on user and support feedback Provide a simple interface for other helpdesk systems (short term) design a web-services interface for other helpdesk applications (mid term) Develop further interfaces to other helpdesk systems New FAQ system with high user integration First plans on how to establish the knowledge base Provide support for EGEE community Search for an third partner in the US to achieve the goal of 24h support

28 Access to GGUS GGUS web portal: Phone FZK: Support time: - ASCC: Mon. to Fri. 0:00 to 08:00 UTC (local time: 8 am to 4 pm) - FZK: Mon. to Fri. 6:00 to 14:00 UTC (local time: 8 am to 4 pm)

29 Thank you for your attention!

Global Grid User Support - GGUS - in the LCG & EGEE environment

Global Grid User Support - GGUS - in the LCG & EGEE environment Global Grid User Support - GGUS - in the LCG & EGEE environment Torsten Antoni (torsten.antoni@iwr.fzk.de) Why Support? New support groups Network layer Resource centers CIC / GOC / etc. more to come New

More information

Global Grid User Support - GGUS - start up schedule

Global Grid User Support - GGUS - start up schedule Global Grid User Support - GGUS - start up schedule GDB Meeting 2004-07 07-13 Concept Target: 24 7 support via time difference and 3 support teams Currently: GGUS FZK GGUS ASCC Planned: GGUS USA Support

More information

Welcome to the User Support for EGEE Task Force Meeting

Welcome to the User Support for EGEE Task Force Meeting Welcome to the User Support for EGEE Task Force Meeting The agenda is as follows: Welcome Note & Presentation of the current GGUS Support system Basic Support Model Coffee brake Processes Lunch Break Interfaces

More information

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Support in EGEE. (SA1 View) Torsten Antoni GGUS, FZK

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Support in EGEE. (SA1 View) Torsten Antoni GGUS, FZK Support in EGEE (SA1 View) Torsten Antoni GGUS, FZK (torsten.antoni@iwr.fzk.de) with input from LCG Operations Workshop, e-irg e Workshop Why is there need for support? New support groups Network layer

More information

Support Model for SC4 Pilot WLCG Service

Support Model for SC4 Pilot WLCG Service Model for SC4 Pilot WLCG Flavia Donno CERN www.eu-egee.org Problems reporting SC : what s implied? Deployment and configuration, middleware, external components, mass storage support, etc. (from site admins,

More information

Regional Dashboard. Cyril L Orphelin - CNRS/IN2P3 Abingdon, England. www.eu-egee.org. EGEE and glite are registered trademarks

Regional Dashboard. Cyril L Orphelin - CNRS/IN2P3 Abingdon, England. www.eu-egee.org. EGEE and glite are registered trademarks Regional Dashboard Cyril L Orphelin - CNRS/IN2P3 Abingdon, England www.eu-egee.org egee EGEE and glite are registered trademarks One common view with 3 different group of roles. If you are registered with

More information

GridKa Database Services

GridKa Database Services Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft GridKa Database Services Overview about our activities and possible tasks in 3D Rainer Kupsch 13.12.2004 in LCG-3D Kickoff-Workshop Institut für

More information

Global Grid User Support Building a worldwide distributed user support infrastructure

Global Grid User Support Building a worldwide distributed user support infrastructure Global Grid User Support Building a worldwide distributed user support infrastructure T. Antoni (1), W. Bühler (1), H. Dres (1), G. Grein (1), T.Kamps (2), R. Kupsch (1), M. Roth (1), R.Stenzel (2) (1)

More information

DSA1.5 U SER SUPPORT SYSTEM

DSA1.5 U SER SUPPORT SYSTEM DSA1.5 U SER SUPPORT SYSTEM H ELP- DESK SYSTEM IN PRODUCTION AND USED VIA WEB INTERFACE Document Filename: Activity: Partner(s): Lead Partner: Document classification: BG-DSA1.5-v1.0-User-support-system.doc

More information

Sustainable Grid User Support

Sustainable Grid User Support Sustainable Grid User Support Dr. Torsten Antoni torsten.antoni@kit.edu www.eu-egee.org EGEE and glite are registered trademarks User education User support is Simple access to a broad range of information

More information

Bob Jones Technical Director bob.jones@cern.ch

Bob Jones Technical Director bob.jones@cern.ch Bob Jones Technical Director bob.jones@cern.ch CERN - August 2003 EGEE is proposed as a project to be funded by the European Union under contract IST-2003-508833 EGEE Goal & Strategy Goal: Create a wide

More information

Integrating a heterogeneous and shared Linux cluster into grids

Integrating a heterogeneous and shared Linux cluster into grids Integrating a heterogeneous and shared Linux cluster into grids 1,2 1 1,2 1 V. Büge, U. Felzmann, C. Jung, U. Kerzel, 1 1 1 M. Kreps, G. Quast, A. Vest 1 2 DPG Frühjahrstagung March 28 31, 2006 Dortmund

More information

GridKa: Roles and Status

GridKa: Roles and Status GridKa: Roles and Status GmbH Institute for Scientific Computing P.O. Box 3640 D-76021 Karlsruhe, Germany Holger Marten http://www.gridka.de History 10/2000: First ideas about a German Regional Centre

More information

Grid @ Forschungszentrum Karlsruhe: GridKa and GGUS

Grid @ Forschungszentrum Karlsruhe: GridKa and GGUS Grid @ Forschungszentrum Karlsruhe: GridKa and GGUS Forschungszentrum Karlsruhe GmbH Institute for Scientific omputing P.O. Box 3640 D-76021 Karlsruhe, Germany Holger Marten (for the GridKa and GGUS teams)

More information

Tier-1 Services for Tier-2 Regional Centres

Tier-1 Services for Tier-2 Regional Centres Tier-1 Services for Tier-2 Regional Centres The LHC Computing MoU is currently being elaborated by a dedicated Task Force. This will cover at least the services that Tier-0 (T0) and Tier-1 centres (T1)

More information

Internal ROC DECH Report

Internal ROC DECH Report Internal ROC DECH Report Sven Hermann et.al. Karlsruhe Institute of Technology www.eu-egee.org EGEE-III INFSO-RI-222667 EGEE and glite are registered trademarks DECH Region U Dortmund DESY DESY Zeuthen

More information

Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware

Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware R. Goranova University of Sofia St. Kliment Ohridski,

More information

CMS Dashboard of Grid Activity

CMS Dashboard of Grid Activity Enabling Grids for E-sciencE CMS Dashboard of Grid Activity Julia Andreeva, Juha Herrala, CERN LCG ARDA Project, EGEE NA4 EGEE User Forum Geneva, Switzerland March 1-3, 2006 http://arda.cern.ch ARDA and

More information

File Transfer Software and Service SC3

File Transfer Software and Service SC3 File Transfer Software and Service SC3 Gavin McCance JRA1 Data Management Cluster Service Challenge Meeting April 26 2005, Taipei www.eu-egee.org Outline Overview of Components Tier-0 / Tier-1 / Tier-2

More information

The dcache Storage Element

The dcache Storage Element 16. Juni 2008 Hamburg The dcache Storage Element and it's role in the LHC era for the dcache team Topics for today Storage elements (SEs) in the grid Introduction to the dcache SE Usage of dcache in LCG

More information

The GENIUS Grid Portal

The GENIUS Grid Portal The GENIUS Grid Portal (*) work in collaboration with A. Falzone and A. Rodolico EGEE NA4 Workshop, Paris, 18.12.2003 CHEP 2000, 10.02.2000 Outline Introduction Grid portal architecture and requirements

More information

Knowledge Management and Semantics in Global Grid User Support

Knowledge Management and Semantics in Global Grid User Support Knowledge Management and Semantics in Global Grid User Support T. Antoni (1), W. Bühler (1), H. Dres (1), G. Grein (1), T.Kamps (2), R. Kupsch (1), M. Roth (1), R.Stenzel (2) (1) Institute for Scientific

More information

Report from SARA/NIKHEF T1 and associated T2s

Report from SARA/NIKHEF T1 and associated T2s Report from SARA/NIKHEF T1 and associated T2s Ron Trompert SARA About SARA and NIKHEF NIKHEF SARA High Energy Physics Institute High performance computing centre Manages the Surfnet 6 network for the Dutch

More information

Document: 8600 FAQ FOR TPM TPM. Version Authors Date Comments

Document: 8600 FAQ FOR TPM TPM. Version Authors Date Comments Document: 8600 FAQ FOR TPM TPM Version Authors Date Comments 0.6 Alistair Mills 24/08/2005 Added 8618-8623, and made minor corrections 0.5 Torsten Antoni 10/08/2005 Minor corrections 0.4 Flavia Donno 07/08/2005

More information

Helmholtz Alliance 2nd Grid Workshop. Christoph Anton Mitterer christoph.anton.mitterer@physik.uni-muenchen.de. for the group

Helmholtz Alliance 2nd Grid Workshop. Christoph Anton Mitterer christoph.anton.mitterer@physik.uni-muenchen.de. for the group Helmholtz Alliance 2nd Grid Workshop HGF Mass Storage Support Group Christoph Anton Mitterer christoph.anton.mitterer@physik.uni-muenchen.de for the group Group members Filled positions Christopher Jung

More information

DSA1.4 R EPORT ON IMPLEMENTATION OF MONITORING AND OPERATIONAL SUPPORT SYSTEM. Activity: SA1. Partner(s): EENet, NICPB. Lead Partner: EENet

DSA1.4 R EPORT ON IMPLEMENTATION OF MONITORING AND OPERATIONAL SUPPORT SYSTEM. Activity: SA1. Partner(s): EENet, NICPB. Lead Partner: EENet R EPORT ON IMPLEMENTATION OF MONITORING AND OPERATIONAL SUPPORT SYSTEM Document Filename: Activity: Partner(s): Lead Partner: Document classification: BG-DSA1.4-v1.0-Monitoring-operational-support-.doc

More information

ARDA Experiment Dashboard

ARDA Experiment Dashboard ARDA Experiment Dashboard Ricardo Rocha (ARDA CERN) on behalf of the Dashboard Team www.eu-egee.org egee INFSO-RI-508833 Outline Background Dashboard Framework VO Monitoring Applications Job Monitoring

More information

The Italian Grid Infrastructure (IGI) CGW08 Cracow. Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director

The Italian Grid Infrastructure (IGI) CGW08 Cracow. Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director The Italian Grid Infrastructure (IGI) CGW08 Cracow Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director IGI -> National Grid Initiative Applications: Archeology Astronomy Astrophysics

More information

Computing in High- Energy-Physics: How Virtualization meets the Grid

Computing in High- Energy-Physics: How Virtualization meets the Grid Computing in High- Energy-Physics: How Virtualization meets the Grid Yves Kemp Institut für Experimentelle Kernphysik Universität Karlsruhe Yves Kemp Barcelona, 10/23/2006 Outline: Problems encountered

More information

The GridKa Installation for HEP Computing

The GridKa Installation for HEP Computing The GridKa Installation for HEP Computing Forschungszentrum Karlsruhe GmbH Central Information and Communication Technologies Department Hermann-von-Helmholtz-Platz 1 D-76344 Eggenstein-Leopoldshafen Holger

More information

Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de

Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de Tier-2 cloud Holger Marten Holger. Marten at iwr. fzk. de www.gridka.de 1 GridKa associated Tier-2 sites spread over 3 EGEE regions. (4 LHC Experiments, 5 (soon: 6) countries, >20 T2 sites) 2 region DECH

More information

Virtualization Infrastructure at Karlsruhe

Virtualization Infrastructure at Karlsruhe Virtualization Infrastructure at Karlsruhe HEPiX Fall 2007 Volker Buege 1),2), Ariel Garcia 1), Marcus Hardt 1), Fabian Kulla 1),Marcel Kunze 1), Oliver Oberst 1),2), Günter Quast 2), Christophe Saout

More information

Database Services for Physics @ CERN

Database Services for Physics @ CERN Database Services for Physics @ CERN Deployment and Monitoring Radovan Chytracek CERN IT Department Outline Database services for physics Status today How we do the services tomorrow? Performance tuning

More information

Tier0 plans and security and backup policy proposals

Tier0 plans and security and backup policy proposals Tier0 plans and security and backup policy proposals, CERN IT-PSS CERN - IT Outline Service operational aspects Hardware set-up in 2007 Replication set-up Test plan Backup and security policies CERN Oracle

More information

Status and Integration of AP2 Monitoring and Online Steering

Status and Integration of AP2 Monitoring and Online Steering Status and Integration of AP2 Monitoring and Online Steering Daniel Lorenz - University of Siegen Stefan Borovac, Markus Mechtel - University of Wuppertal Ralph Müller-Pfefferkorn Technische Universität

More information

The dashboard Grid monitoring framework

The dashboard Grid monitoring framework The dashboard Grid monitoring framework Benjamin Gaidioz on behalf of the ARDA dashboard team (CERN/EGEE) ISGC 2007 conference The dashboard Grid monitoring framework p. 1 introduction/outline goals of

More information

The Grid-it: the Italian Grid Production infrastructure

The Grid-it: the Italian Grid Production infrastructure n 1 Maria Cristina Vistoli INFN CNAF, Bologna Italy The Grid-it: the Italian Grid Production infrastructure INFN-Grid goals!promote computational grid technologies research & development: Middleware and

More information

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan

Status of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan Status of Grid Activities in Pakistan FAWAD SAEED National Centre For Physics, Pakistan 1 Introduction of NCP-LCG2 q NCP-LCG2 is the only Tier-2 centre in Pakistan for Worldwide LHC computing Grid (WLCG).

More information

ITIL and Grid services at GridKa CHEP 2009, 21-27 March, Prague

ITIL and Grid services at GridKa CHEP 2009, 21-27 March, Prague ITIL Grid services at GridKa CHEP 2009, 21-27 March, Prague Tobias König, Dr. Holger Marten Steinbuch Centre for Computing (SCC) KIT The cooperation of Forschungszentrum Karlsruhe GmbH und Universität

More information

PoS(EGICF12-EMITC2)110

PoS(EGICF12-EMITC2)110 User-centric monitoring of the analysis and production activities within the ATLAS and CMS Virtual Organisations using the Experiment Dashboard system Julia Andreeva E-mail: Julia.Andreeva@cern.ch Mattia

More information

CY-01-KIMON operational issues

CY-01-KIMON operational issues CY-01-KIMON operational issues Speaker: Kyriacos Neocleous High Performance Computing systems Lab, University Of Cyprus SA1 workshop: Stockholm, Sweden, June 2007 www.eu-egee.org Presentation outline (1)

More information

EGEE a worldwide Grid infrastructure

EGEE a worldwide Grid infrastructure EGEE a worldwide Grid infrastructure Fabrizio Gagliardi Project Director EGEE CERN, Switzerland IFIC, 6 October 2005 www.eu-egee.org Presentation overview Data intensive science and the rationale for Grid

More information

EELA Operations: A standalone regional dashboard implementation

EELA Operations: A standalone regional dashboard implementation Proceedings of the Second EELA-2 Conference R. Mayo et al. (Eds.) CIEMAT 2009 2009 The authors. All rights reserved EELA Operations: A standalone regional dashboard implementation Cyril L'Orphelin, Hélène

More information

Report from Italian ROC

Report from Italian ROC Report from Italian ROC Paolo Veronesi for ROC It www.eu-egee.org ARM-7, ROC(111) 15th - 17th May 2006 - Krakow, Outline Changes in ROC structure over the last 3 months (people/institutes involved) People

More information

Solution for private cloud computing

Solution for private cloud computing The CC1 system Solution for private cloud computing 1 Outline What is CC1? Features Technical details Use cases By scientist By HEP experiment System requirements and installation How to get it? 2 What

More information

irods at CC-IN2P3: managing petabytes of data

irods at CC-IN2P3: managing petabytes of data Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules irods at CC-IN2P3: managing petabytes of data Jean-Yves Nief Pascal Calvat Yonny Cardenas Quentin Le Boulc h

More information

Virtualization, Grid, Cloud: Integration Paths for Scientific Computing

Virtualization, Grid, Cloud: Integration Paths for Scientific Computing Virtualization, Grid, Cloud: Integration Paths for Scientific Computing Or, where and how will my efficient large-scale computing applications be executed? D. Salomoni, INFN Tier-1 Computing Manager Davide.Salomoni@cnaf.infn.it

More information

Operational Model for E2E links in the NREN/GÉANT2 and NREN/Cross-Border-Fibre supplied optical platform

Operational Model for E2E links in the NREN/GÉANT2 and NREN/Cross-Border-Fibre supplied optical platform Operational Model for E2E links in the NREN/GÉANT2 and NREN/Cross-Border-Fibre supplied optical platform 0 Background In the GÉANT2 project plan it is foreseen that end-to-end (E2E) links (essentially

More information

Data storage services at CC-IN2P3

Data storage services at CC-IN2P3 Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief Agenda Hardware: Storage on disk. Storage on tape. Software:

More information

SPACI & EGEE LCG on IA64

SPACI & EGEE LCG on IA64 SPACI & EGEE LCG on IA64 Dr. Sandro Fiore, University of Lecce and SPACI December 13 th 2005 www.eu-egee.org Outline EGEE Production Grid SPACI Activity Status of the LCG on IA64 SPACI & EGEE Farm Configuration

More information

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil

Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Volker Büge 1, Marcel Kunze 2, OIiver Oberst 1,2, Günter Quast 1, Armin Scheurer 1 1) Institut für Experimentelle

More information

The CMS analysis chain in a distributed environment

The CMS analysis chain in a distributed environment The CMS analysis chain in a distributed environment on behalf of the CMS collaboration DESY, Zeuthen,, Germany 22 nd 27 th May, 2005 1 The CMS experiment 2 The CMS Computing Model (1) The CMS collaboration

More information

DCMS Tier 2/3 prototype infrastructure

DCMS Tier 2/3 prototype infrastructure DCMS Tier 2/3 prototype infrastructure 1 Anja Vest, Uni Karlsruhe DCMS Meeting in Aachen, Overview LCG Queues/mapping set up Hardware capacities Supported software Summary DCMS overview CMS DCMS: Tier

More information

Introduction to Grid computing

Introduction to Grid computing Introduction to Grid computing The INFNGrid Project Team Introduction This tutorial has been implemented considering as starting point the DataGrid (EDG) tutorial Many thanks to the EDG tutorials team!

More information

Recent grid activities at INFN Catania(*) Roberto Barbera

Recent grid activities at INFN Catania(*) Roberto Barbera RecentgridactivitiesatINFNCatania(*) RobertoBarbera workincollaborationwithnicesrl (*) HEPiX/HEPNT2002,Catania,18.04.2002 CHEP2000,10.02.2000 1 RobertoBarbera DipartimentodiFisicadell UniversitàdiCataniaandINFNCatania

More information

Distributed Database Access in the LHC Computing Grid with CORAL

Distributed Database Access in the LHC Computing Grid with CORAL Distributed Database Access in the LHC Computing Grid with CORAL Dirk Duellmann, CERN IT on behalf of the CORAL team (R. Chytracek, D. Duellmann, G. Govi, I. Papadopoulos, Z. Xie) http://pool.cern.ch &

More information

Web Hosting. E-Mail Hosting. Cloud File Hosting. The Genio Group (214) 732-7411 info@thegeniogroup.com www.thegeniogroup.com

Web Hosting. E-Mail Hosting. Cloud File Hosting. The Genio Group (214) 732-7411 info@thegeniogroup.com www.thegeniogroup.com Web Hosting E-Mail Hosting Cloud File Hosting Genio Hosting Servers All of Genio s Hosting Servers run on Apple hardware running Mac OS X Server. Mac OS X Server leverages the computing power of 64-bit

More information

Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC

Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC EGEE and glite are registered trademarks Enabling Grids for E-sciencE Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC Elisa Lanciotti, Arnau Bria, Gonzalo

More information

Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb

Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb Batch and Cloud overview Andrew McNab University of Manchester GridPP and LHCb Overview Assumptions Batch systems The Grid Pilot Frameworks DIRAC Virtual Machines Vac Vcycle Tier-2 Evolution Containers

More information

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. GridKa User Meeting

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. GridKa User Meeting GridKa User Meeting Forschungszentrum Karlsruhe GmbH Central Information and Communication Technologies Department Hermann-von-Helmholtz-Platz 1 D-76344 Eggenstein-Leopoldshafen Dr. Holger Marten http://grid.fzk.de

More information

HAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions

HAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions DESY Tier 2 and NAF Peter Wegner, Birgit Lewendel for DESY-IT/DV Tier 2: Status and News NAF: Status, Plans and Questions Basics T2: 1.5 average Tier 2 are requested by CMS-groups for Germany Desy commitment:

More information

Grid Computing in Aachen

Grid Computing in Aachen GEFÖRDERT VOM Grid Computing in Aachen III. Physikalisches Institut B Berichtswoche des Graduiertenkollegs Bad Honnef, 05.09.2008 Concept of Grid Computing Computing Grid; like the power grid, but for

More information

Instruments in Grid: the New Instrument Element

Instruments in Grid: the New Instrument Element Instruments in Grid: the New Instrument Element C. Vuerli (1,2), G. Taffoni (1,2), I. Coretti (1), F. Pasian (1,2), P. Santin (1), M. Pucillo (1) (1) INAF Astronomical Observatory of Trieste (2) INAF Informative

More information

vtiger Customer Portal 4.2 User Manual

vtiger Customer Portal 4.2 User Manual - 1 - vtiger Customer Portal 4.2 User Manual (Version 2.0) - 2 - Table of Contents 1. Introduction... 3 2. Installing vtiger Customer Portal... 4 2.1. System Requirements... 4 2.2. Installation Prerequisites...

More information

Virtualization of a Cluster Batch System

Virtualization of a Cluster Batch System Virtualization of a Cluster Batch System Christian Baun, Volker Büge, Benjamin Klein, Jens Mielke, Oliver Oberst and Armin Scheurer Die Kooperation von Cluster Batch System Batch system accepts computational

More information

AN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID

AN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID AN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID R. D. Goranova 1, V. T. Dimitrov 2 Faculty of Mathematics and Informatics, University of Sofia S. Kliment Ohridski, 1164, Sofia, Bulgaria

More information

Unterstützung datenintensiver Forschung am KIT Aktivitäten, Dienste und Erfahrungen

Unterstützung datenintensiver Forschung am KIT Aktivitäten, Dienste und Erfahrungen Unterstützung datenintensiver Forschung am KIT Aktivitäten, Dienste und Erfahrungen Achim Streit Steinbuch Centre for Computing (SCC) KIT Universität des Landes Baden-Württemberg und nationales Forschungszentrum

More information

LCG POOL, Distributed Database Deployment and Oracle Services@CERN

LCG POOL, Distributed Database Deployment and Oracle Services@CERN LCG POOL, Distributed Database Deployment and Oracle Services@CERN Dirk Düllmann, D CERN HEPiX Fall 04, BNL Outline: POOL Persistency Framework and its use in LHC Data Challenges LCG 3D Project scope and

More information

Improvements of the grid infrastructure and services within SEE-GRID

Improvements of the grid infrastructure and services within SEE-GRID Improvements of the grid infrastructure and services within SEE-GRID Anastas Misev MARNET/MARGI/UKIM Macedonia Introduction SEE-GRID Project series SEE-GRID establish infrastructure SEE-GRID-2 extend infrastructure,

More information

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12 XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines A.Zydroń 18 April 2009 Page 1 of 12 1. Introduction...3 2. XTM Database...4 3. JVM and Tomcat considerations...5 4. XTM Engine...5

More information

Deploying Multiscale Applications on European e-infrastructures

Deploying Multiscale Applications on European e-infrastructures Deploying Multiscale Applications on European e-infrastructures 04/06/2013 Ilya Saverchenko The MAPPER project receives funding from the EC's Seventh Framework Programme (FP7/2007-2013) under grant agreement

More information

OPEN CALL FOR TENDERS No MARE/2014/14. Appendix 2. EUMOFA IT architecture and design

OPEN CALL FOR TENDERS No MARE/2014/14. Appendix 2. EUMOFA IT architecture and design EUROPEAN COMMISSION DIRECTORATE-GENERAL FOR MARITIME AFFAIRS AND FISHERIES INTERNATIONAL AFFAIRS AND MARKETS TRADE AND MARKETS OPEN CALL FOR TENDERS No MARE/2014/14 Appendix 2 EUMOFA IT architecture and

More information

LHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch. CERN, July 2007

LHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch. CERN, July 2007 WLCG Service Schedule LHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch WLCG Storage Workshop CERN, July 2007 Agenda The machine The experiments The service LHC Schedule Mar. Apr.

More information

Linux Titan Distribution. Presented By: Adham Helal Amgad Madkour Ayman El Sayed Emad Zakaria

Linux Titan Distribution. Presented By: Adham Helal Amgad Madkour Ayman El Sayed Emad Zakaria Linux Titan Distribution Presented By: Adham Helal Amgad Madkour Ayman El Sayed Emad Zakaria What Is a Linux Distribution? What is a Linux Distribution? The distribution contains groups of packages and

More information

Big Data and Storage Management at the Large Hadron Collider

Big Data and Storage Management at the Large Hadron Collider Big Data and Storage Management at the Large Hadron Collider Dirk Duellmann CERN IT, Data & Storage Services Accelerating Science and Innovation CERN was founded 1954: 12 European States Science for Peace!

More information

Fast, Easy to use and On-demand A Content Platform from the 21st Century

Fast, Easy to use and On-demand A Content Platform from the 21st Century OpenIMS Open Information Management Server Fast, Easy to use and On-demand A Content Platform from the 21st Century E-mail Management Server Whitepaper Open Information Management Server www.openims.com

More information

Deliverable DS4.3.2: Report on Development Infrastructure Usage and Adoption

Deliverable DS4.3.2: Report on Development Infrastructure Usage and Adoption 05-10-2010 Deliverable DS4.3.2 Contractual Date: 30-06-2010 Actual Date: 05-10-2010 Grant Agreement No.: 238875 Activity: SA4 Task Item: T3 Nature of Deliverable: R (Report) Dissemination Level: PU (Public)

More information

Configuration Management of Massively Scalable Systems

Configuration Management of Massively Scalable Systems 1 KKIO 2005 Configuration Management of Massively Scalable Systems Configuration Management of Massively Scalable Systems Marcin Jarząb, Krzysztof Zieliński, Jacek Kosiński SUN Center of Excelence Department

More information

HISP: a data-driven portal for hadron therapy

HISP: a data-driven portal for hadron therapy HISP: a data-driven portal for hadron therapy Faustin Laurentiu Roman CERN / IFIC Prototype architecture Tools, implementation & services Conclusions (& demo) 1 One slide situation: ereferral and escience

More information

Report of the LHC Computing Grid Project. Software Management Process RTAG CERN

Report of the LHC Computing Grid Project. Software Management Process RTAG CERN Report of the LHC Computing Grid Project Software Management Process RTAG Marco Cattaneo, Gabriele Cosmo, Simon George, Fons Rademakers (chair), Stephan Wynhoff CERN 6 May 2002 Table of Contents 1 Chair

More information

WHITE PAPER Managing Linux in the Enterprise: The Red Hat Network Approach

WHITE PAPER Managing Linux in the Enterprise: The Red Hat Network Approach Global Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com WHITE PAPER Managing Linux in the Enterprise: The Red Hat Network Approach Sponsored by: Red Hat Tim

More information

Grid Site Monitoring tools developed and used at at SCL

Grid Site Monitoring tools developed and used at at SCL Grid Site Monitoring tools developed and used at at SCL V. Slavnić, B. Acković, D. Vudragović, A. Balaž, A. Belić Scientific Computing Laboratory Institute of Physics Belgrade Pregrevica 118, 11080 Belgrade,

More information

Deliverable D 6.1 Website

Deliverable D 6.1 Website Biogas2PEM-FC Biogas Reforming and Valorisation through PEM Fuel Cells FP7-SME-2012, Grant Agreement No. 314940 Deliverable D 6.1 Website Deliverable details Deliverable version: v1.0 Date of delivery:

More information

HP Service Manager Compatibility Matrix

HP Service Manager Compatibility Matrix HP Service Manager Compatibility Matrix Software Version 9.21 January 12, 2011 Click one of the following links to see more detailed information. Tier Definitions Servers Applications Support Windows Client

More information

GiftCardXpress - Elavon Brief

GiftCardXpress - Elavon Brief GiftCardXpress - Elavon Brief CFXWORKS, INC 2015 http://www.cfxworks.com GiftCardXpress - Elavon Brief GiftCardXpress (Elavon) Version 16.1: This gift card solution, GiftCardXpress (Elavon), is an open

More information

Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases

Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases NASA Ames NASA Advanced Supercomputing (NAS) Division California, May 24th, 2012 Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases Ignacio M. Llorente Project Director OpenNebula Project.

More information

The Italian Regional Helpdesk System

The Italian Regional Helpdesk System The Italian Regional Helpdesk System A. Paolini (INFN-CNAF) R. Brunetti (INFN-TORINO) www.eu-egee.org EGEE and glite are registered trademarks1 Outline Overview of the system Framework Base features Extended

More information

Open-source and Standards - Unleashing the Potential for Innovation of Cloud Computing

Open-source and Standards - Unleashing the Potential for Innovation of Cloud Computing Cloudscape IV Advances on Interoperability & Cloud Computing Standards Brussels, Belgium, February 23th, 2012 Open-source and Standards - Unleashing the Potential for Innovation of Cloud Computing Ignacio

More information

Deploying a distributed data storage system on the UK National Grid Service using federated SRB

Deploying a distributed data storage system on the UK National Grid Service using federated SRB Deploying a distributed data storage system on the UK National Grid Service using federated SRB Manandhar A.S., Kleese K., Berrisford P., Brown G.D. CCLRC e-science Center Abstract As Grid enabled applications

More information

Grids Computing and Collaboration

Grids Computing and Collaboration Grids Computing and Collaboration Arto Teräs CSC, the Finnish IT center for science University of Pune, India, March 12 th 2007 Grids Computing and Collaboration / Arto Teräs 2007-03-12 Slide

More information

CloudCERT (Testbed framework to exercise critical infrastructure protection)

CloudCERT (Testbed framework to exercise critical infrastructure protection) WP2. CONCEPTUAL MODELLING AND ARCHITECTURE CloudCERT (Testbed framework to exercise critical infrastructure protection) With the financial support of the Prevention, Preparedness and Consequence Management

More information

CHEP 2013. Cloud Bursting with glideinwms Means to satisfy ever increasing computing needs for Scientific Workflows

CHEP 2013. Cloud Bursting with glideinwms Means to satisfy ever increasing computing needs for Scientific Workflows CHEP 2013 Cloud Bursting with glideinwms Means to satisfy ever increasing computing needs for Scientific Workflows by I. Sfiligoi 1, P. Mhashilkar 2, A. Tiradani 2, B. Holzman 2, K. Larson 2 and M. Rynge

More information

HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS EKSPERYMENTY FIZYKI WYSOKICH ENERGII W SIECIACH KOMPUTEROWYCH GRID. 1.

HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS EKSPERYMENTY FIZYKI WYSOKICH ENERGII W SIECIACH KOMPUTEROWYCH GRID. 1. Computer Science Vol. 9 2008 Andrzej Olszewski HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS The demand for computing resources used for detector simulations and data analysis in High Energy

More information

Intershop 7 System Requirements Sheet

Intershop 7 System Requirements Sheet Specific System Requirements on Microsoft Windows Server 2008 Windows Server 2008 Enterprise Edition (64bit), Windows Server 2008 Standard Edition (64bit) all SP2 per application server process recommended

More information

Azienda sul Web. Il Mio Sito

Azienda sul Web. Il Mio Sito REV.01 2013 p. 1/11 Roma, 16/01/2014 Dear Telefonica S.A. marketing department SUBJECT: Revenue Sharing Webee Telefonica INTRODUCTION Architecture will be the same as that already established in partnership

More information

BMC Remedy Integration Guide 7.6.04

BMC Remedy Integration Guide 7.6.04 BMC Remedy Integration Guide 7.6.04 2015 Bomgar Corporation. All rights reserved worldwide. BOMGAR and the BOMGAR logo are trademarks of Bomgar Corporation; other trademarks shown are the property of their

More information

CERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006

CERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006 CERN local High Availability solutions and experiences Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006 1 Introduction Different h/w used for GRID services Various techniques & First

More information

Monitoring Remedy with BMC Solutions

Monitoring Remedy with BMC Solutions Monitoring Remedy with BMC Solutions Overview How does BMC Software monitor Remedy with our own solutions? The challenge is many fold with a solution like Remedy and this does not only apply to Remedy,

More information

Real Time Monitor of Grid Job Executions. Janusz Martyniak Imperial College London

Real Time Monitor of Grid Job Executions. Janusz Martyniak Imperial College London Real Time Monitor of Grid Job Executions Janusz Martyniak Imperial College London What is the RTM RTM is a GRID monitoring system which gives on overview of current state of the infrastructure. The RTM's

More information