Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Global Grid User Support - GGUS - within the LCG & EGEE environment
|
|
- Lester Victor Chambers
- 8 years ago
- Views:
Transcription
1 Global Grid User Support - GGUS - within the LCG & EGEE environment
2 Abstract: For very large projects like the LHC Computing Grid Project (LCG) involving some 8,000 scientists from universities and laboratories all around the world, it is an indispensable requirement to have a well organized user support. The Institute for Scientific Computing at the Forschungszentrum Karlsruhe started implementing a Global Grid User Support (GGUS) after official assignment of the Grid Deployment Board in March For this purpose a web portal together with a helpdesk application has been developed. As a single entry point for all Grid related issues and problems GGUS follows the objectives of providing news, documentation and status information about Grid resources. The user will find forms to submit and track service requests. GGUS collaborates with different support teams in the Grid environment like the Grid Operations Center and the Experiment Specific Support. They can access the helpdesk system via web interface. GGUS stores all the incoming trouble tickets and outgoing solutions in a central database and plans to build up a knowledge base where all the information can be offered in a structured manner. As a prototype GGUS started operation at the Forschungszentrum Karlsruhe in October 2003 and supported local user groups of the German Tier 1 Computing Center, called GridKa. 4 month later the GGUS system was opened for the LCG community and since March 2004 ASCC Taiwan is the second GGUS partner. Due to the time difference between the two support centers GGUS is operational 15 hours per day now. A major target for GGUS is to be available 24 hours a day, 7 days a week. Therefore we are trying to find a qualified third partner in the US time zone. In the talk the GGUS system will be explained and integration into the Grid environment will be explained.
3 About the Author Name: Wolfgang Thöne Born 1965, living in Linkenheim near Karlsruhe Working at the Forschungszentrum Karlsruhe, Institute for Science Computing. He is member of the Global Grid User Support team, that is established since October 2002 within the LCG project. He leads the concept phase of the project and is currently working on the concepts for user support within the EGEE project.
4 Outline Support? GGUS concept and objectives Support teams Current status of GGUS web portal and helpdesk system Outlook and next steps GGUS staff
5 Forschungszentrum One of the biggest research institutions in Europe, 3500 employees 5 Research areas: - Structure of Matter - Earth and Environment - Health -Energy - Key Technologies (Nanotechnology, Grid Computing) Institute for Scientific Computing: - providing IT services - establishing GridKa, GGUS
6 Forschungszentrum Karlsruhe International Grid Projects Through links with sister projects, there is the potential for a truely global scientific applications grid
7 Goal EGEE manifesto: Enabling Grids for E-science in Europe Create a wide European Grid production quality infrastructure on top of present and future EU RN infrastructure ERA Applications Build on EU and EU member states major investments in Grid Technology International connections (US and AP) Several pioneering prototype results Larg Grid development teams in EU Requires major EU funding effort Approach EGEE Leverage current and planned national and regional Grid programmes Geant Geant Research network Network Work closely with relevant industrial Grid developers, NRENs and US-AP projects Grid infrastructure
8 Operations Center Regional Support Regional Support GGUS Regional Support Resource Center (Processors, disks) Regional Support Center (Support for Applications Local Resources) Grid server Nodes Resource Center Resource Center Resource Center EGEE Operations Structure
9 Why Support? New support groups Network layer Resource centers CIC / GOC / etc. more to come New Support challenges: Jobs are spread over several locations New types of Job and Data access Often more than one group is involved Virtual Organizations
10 Why Support? Need to have a common platform for: Collect service requests Exchange service requests Document solutions Need to document: Quality feedback To the VO s Middleware development To the Grid Operations FAQ s Knowledge Base
11 Concept of Global Grid User Support - GGUS Target: 24 7 support via time difference and 3 support teams Currently: GGUS FZK GGUS ASCC Planned: GGUS USA
12 Objectives Single point of contact and information for Grid related questions (by now for LCG and EGEE) FAQ s Downloads News and problems Support entry point for 8 HEP experiments (and other Grid user communities, EGEE) Operate a central trouble ticket database and system Design, implementation and operation of a knowledge base Be the last resort for all support requests
13 Support Teams within LCG & EGEE Deployment Support (DS) Middleware Problems Operations Center (CIC / GOC / ROC) Operations Problems Resource Centers (RC) Hardware Problems Global Grid User Support (GGUS) Single Point of Contact Coordination of UserSupport Experiment Specific User Support (ESUS) VO spec. (Software) Problems Other Communities (VOs), e.g. EGEE LHC experiments (Alice Atlas CMS LHCb) non-lhc experiments (BaBar, CDF, Compass, D0)
14 Support Workflow (LCG) ESUS Application problem GGUS DS Middleware problem User CIC / ROC +Resource Center Operation problem
15 User Problem Process model Support request Interaction Support request GGUS Support Feedback & Solution Support request Interaction ESUS Red lines: Interface the GGUS support system and/or supply an interface to GGUS system DS CIC/ROC Interaction Resource Center
16 Process model Update Web Portal Info Interaction GGUS Support Interaction Create Solve Ticket ESUS Red lines: Interface the GGUS support system and/or supply an interface to GGUS system DS CIC/ROC Interaction Resource Center
17 Architecture User Resources and Status Central User DB VO DB Web-Interface for user interaction Middleware & Interfaces GGUS Helpdesk Application (Remedy) Reporting & Analysis Workflow engine Middleware & Interfaces Web-Interface & JAVA Client for support interaction Universal Database for for GRID GRID related related problems ESUS, DS, GOC/ROC
18 Application model SuSE Linux 9.x Apache 1.3.x with mod_ssl Tomcat 4.1.x (for Remedy 6.0) MySQL 4.0.x PHP 4.x Remedy 5.1 Server & Middleware Next Release will base on Remedy 6.0 Remedy 5.1 C/C++ API Next Release will base on Remedy 6.0 Oracle OCI 9.x with Oracle in a cluster environment
19 Application model access via Remedy Client access via web-browser Apache webserver PHP Oracle/Remedy Interface Remedy API Interface Remedy Mid-Tier Application Remedy Client Remedy Application Oracle database read
20 Current status GGUS team 4 people at FZK 5 people at ASCC Teamed up with ASCC since April 2004 Support time from 02:00 16:30 CEST
21 Current status Helpdesk application Release version 1.0 is deployed and operational Workflows and escalations are running Application is accessible via web interface and via windows client User authentication via certificates and userid/password are in place DS, GOC/ROC and ESUS may use a central application via the Web
22 Current status Web Portal is online since October 2003 User interface to submit, modify and track tickets is deployed and operational Rudimental news information system is in place First version of FAQ system and a documentation section is online Monitoring and status information is available for LCG-2 Sites and local operations (GridKa)
23 Screenshots GGUS-website Home GGUS-website Submit a Service Request Remedy web view for support staff
24
25
26
27 Outlook New Helpdesk application (Remedy release 6.0) Enhancements based on user and support feedback Provide a simple interface for other helpdesk systems (short term) design a web-services interface for other helpdesk applications (mid term) Develop further interfaces to other helpdesk systems New FAQ system with high user integration First plans on how to establish the knowledge base Provide support for EGEE community Search for an third partner in the US to achieve the goal of 24h support
28 Access to GGUS GGUS web portal: Phone FZK: Support time: - ASCC: Mon. to Fri. 0:00 to 08:00 UTC (local time: 8 am to 4 pm) - FZK: Mon. to Fri. 6:00 to 14:00 UTC (local time: 8 am to 4 pm)
29 Thank you for your attention!
Global Grid User Support - GGUS - in the LCG & EGEE environment
Global Grid User Support - GGUS - in the LCG & EGEE environment Torsten Antoni (torsten.antoni@iwr.fzk.de) Why Support? New support groups Network layer Resource centers CIC / GOC / etc. more to come New
More informationGlobal Grid User Support - GGUS - start up schedule
Global Grid User Support - GGUS - start up schedule GDB Meeting 2004-07 07-13 Concept Target: 24 7 support via time difference and 3 support teams Currently: GGUS FZK GGUS ASCC Planned: GGUS USA Support
More informationWelcome to the User Support for EGEE Task Force Meeting
Welcome to the User Support for EGEE Task Force Meeting The agenda is as follows: Welcome Note & Presentation of the current GGUS Support system Basic Support Model Coffee brake Processes Lunch Break Interfaces
More informationForschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. Support in EGEE. (SA1 View) Torsten Antoni GGUS, FZK
Support in EGEE (SA1 View) Torsten Antoni GGUS, FZK (torsten.antoni@iwr.fzk.de) with input from LCG Operations Workshop, e-irg e Workshop Why is there need for support? New support groups Network layer
More informationSupport Model for SC4 Pilot WLCG Service
Model for SC4 Pilot WLCG Flavia Donno CERN www.eu-egee.org Problems reporting SC : what s implied? Deployment and configuration, middleware, external components, mass storage support, etc. (from site admins,
More informationGridKa Database Services
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft GridKa Database Services Overview about our activities and possible tasks in 3D Rainer Kupsch 13.12.2004 in LCG-3D Kickoff-Workshop Institut für
More informationRegional Dashboard. Cyril L Orphelin - CNRS/IN2P3 Abingdon, England. www.eu-egee.org. EGEE and glite are registered trademarks
Regional Dashboard Cyril L Orphelin - CNRS/IN2P3 Abingdon, England www.eu-egee.org egee EGEE and glite are registered trademarks One common view with 3 different group of roles. If you are registered with
More informationGlobal Grid User Support Building a worldwide distributed user support infrastructure
Global Grid User Support Building a worldwide distributed user support infrastructure T. Antoni (1), W. Bühler (1), H. Dres (1), G. Grein (1), T.Kamps (2), R. Kupsch (1), M. Roth (1), R.Stenzel (2) (1)
More informationDSA1.5 U SER SUPPORT SYSTEM
DSA1.5 U SER SUPPORT SYSTEM H ELP- DESK SYSTEM IN PRODUCTION AND USED VIA WEB INTERFACE Document Filename: Activity: Partner(s): Lead Partner: Document classification: BG-DSA1.5-v1.0-User-support-system.doc
More informationSustainable Grid User Support
Sustainable Grid User Support Dr. Torsten Antoni torsten.antoni@kit.edu www.eu-egee.org EGEE and glite are registered trademarks User education User support is Simple access to a broad range of information
More informationBob Jones Technical Director bob.jones@cern.ch
Bob Jones Technical Director bob.jones@cern.ch CERN - August 2003 EGEE is proposed as a project to be funded by the European Union under contract IST-2003-508833 EGEE Goal & Strategy Goal: Create a wide
More informationGridKa: Roles and Status
GridKa: Roles and Status GmbH Institute for Scientific Computing P.O. Box 3640 D-76021 Karlsruhe, Germany Holger Marten http://www.gridka.de History 10/2000: First ideas about a German Regional Centre
More informationIntegrating a heterogeneous and shared Linux cluster into grids
Integrating a heterogeneous and shared Linux cluster into grids 1,2 1 1,2 1 V. Büge, U. Felzmann, C. Jung, U. Kerzel, 1 1 1 M. Kreps, G. Quast, A. Vest 1 2 DPG Frühjahrstagung March 28 31, 2006 Dortmund
More informationGrid @ Forschungszentrum Karlsruhe: GridKa and GGUS
Grid @ Forschungszentrum Karlsruhe: GridKa and GGUS Forschungszentrum Karlsruhe GmbH Institute for Scientific omputing P.O. Box 3640 D-76021 Karlsruhe, Germany Holger Marten (for the GridKa and GGUS teams)
More informationTier-1 Services for Tier-2 Regional Centres
Tier-1 Services for Tier-2 Regional Centres The LHC Computing MoU is currently being elaborated by a dedicated Task Force. This will cover at least the services that Tier-0 (T0) and Tier-1 centres (T1)
More informationInternal ROC DECH Report
Internal ROC DECH Report Sven Hermann et.al. Karlsruhe Institute of Technology www.eu-egee.org EGEE-III INFSO-RI-222667 EGEE and glite are registered trademarks DECH Region U Dortmund DESY DESY Zeuthen
More informationAnalyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware
Analyses on functional capabilities of BizTalk Server, Oracle BPEL Process Manger and WebSphere Process Server for applications in Grid middleware R. Goranova University of Sofia St. Kliment Ohridski,
More informationCMS Dashboard of Grid Activity
Enabling Grids for E-sciencE CMS Dashboard of Grid Activity Julia Andreeva, Juha Herrala, CERN LCG ARDA Project, EGEE NA4 EGEE User Forum Geneva, Switzerland March 1-3, 2006 http://arda.cern.ch ARDA and
More informationFile Transfer Software and Service SC3
File Transfer Software and Service SC3 Gavin McCance JRA1 Data Management Cluster Service Challenge Meeting April 26 2005, Taipei www.eu-egee.org Outline Overview of Components Tier-0 / Tier-1 / Tier-2
More informationReport from SARA/NIKHEF T1 and associated T2s
Report from SARA/NIKHEF T1 and associated T2s Ron Trompert SARA About SARA and NIKHEF NIKHEF SARA High Energy Physics Institute High performance computing centre Manages the Surfnet 6 network for the Dutch
More informationThe dcache Storage Element
16. Juni 2008 Hamburg The dcache Storage Element and it's role in the LHC era for the dcache team Topics for today Storage elements (SEs) in the grid Introduction to the dcache SE Usage of dcache in LCG
More informationThe GENIUS Grid Portal
The GENIUS Grid Portal (*) work in collaboration with A. Falzone and A. Rodolico EGEE NA4 Workshop, Paris, 18.12.2003 CHEP 2000, 10.02.2000 Outline Introduction Grid portal architecture and requirements
More informationDcache Support and Strategy
Helmholtz Alliance 2nd Grid Workshop HGF Mass Storage Support Group Christoph Anton Mitterer christoph.anton.mitterer@physik.uni-muenchen.de for the group Group members Filled positions Christopher Jung
More informationKnowledge Management and Semantics in Global Grid User Support
Knowledge Management and Semantics in Global Grid User Support T. Antoni (1), W. Bühler (1), H. Dres (1), G. Grein (1), T.Kamps (2), R. Kupsch (1), M. Roth (1), R.Stenzel (2) (1) Institute for Scientific
More informationDocument: 8600 FAQ FOR TPM TPM. Version Authors Date Comments
Document: 8600 FAQ FOR TPM TPM Version Authors Date Comments 0.6 Alistair Mills 24/08/2005 Added 8618-8623, and made minor corrections 0.5 Torsten Antoni 10/08/2005 Minor corrections 0.4 Flavia Donno 07/08/2005
More informationThe Italian Grid Infrastructure (IGI) CGW08 Cracow. Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director
The Italian Grid Infrastructure (IGI) CGW08 Cracow Mirco Mazzucato Italian Grid Infrastructure Coordinator INFN CNAF Director IGI -> National Grid Initiative Applications: Archeology Astronomy Astrophysics
More informationThe GridKa Installation for HEP Computing
The GridKa Installation for HEP Computing Forschungszentrum Karlsruhe GmbH Central Information and Communication Technologies Department Hermann-von-Helmholtz-Platz 1 D-76344 Eggenstein-Leopoldshafen Holger
More informationForschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de
Tier-2 cloud Holger Marten Holger. Marten at iwr. fzk. de www.gridka.de 1 GridKa associated Tier-2 sites spread over 3 EGEE regions. (4 LHC Experiments, 5 (soon: 6) countries, >20 T2 sites) 2 region DECH
More informationDSA1.4 R EPORT ON IMPLEMENTATION OF MONITORING AND OPERATIONAL SUPPORT SYSTEM. Activity: SA1. Partner(s): EENet, NICPB. Lead Partner: EENet
R EPORT ON IMPLEMENTATION OF MONITORING AND OPERATIONAL SUPPORT SYSTEM Document Filename: Activity: Partner(s): Lead Partner: Document classification: BG-DSA1.4-v1.0-Monitoring-operational-support-.doc
More informationARDA Experiment Dashboard
ARDA Experiment Dashboard Ricardo Rocha (ARDA CERN) on behalf of the Dashboard Team www.eu-egee.org egee INFSO-RI-508833 Outline Background Dashboard Framework VO Monitoring Applications Job Monitoring
More informationTier0 plans and security and backup policy proposals
Tier0 plans and security and backup policy proposals, CERN IT-PSS CERN - IT Outline Service operational aspects Hardware set-up in 2007 Replication set-up Test plan Backup and security policies CERN Oracle
More informationComputing in High- Energy-Physics: How Virtualization meets the Grid
Computing in High- Energy-Physics: How Virtualization meets the Grid Yves Kemp Institut für Experimentelle Kernphysik Universität Karlsruhe Yves Kemp Barcelona, 10/23/2006 Outline: Problems encountered
More informationVirtualization Infrastructure at Karlsruhe
Virtualization Infrastructure at Karlsruhe HEPiX Fall 2007 Volker Buege 1),2), Ariel Garcia 1), Marcus Hardt 1), Fabian Kulla 1),Marcel Kunze 1), Oliver Oberst 1),2), Günter Quast 2), Christophe Saout
More informationDatabase Services for Physics @ CERN
Database Services for Physics @ CERN Deployment and Monitoring Radovan Chytracek CERN IT Department Outline Database services for physics Status today How we do the services tomorrow? Performance tuning
More informationITIL and Grid services at GridKa CHEP 2009, 21-27 March, Prague
ITIL Grid services at GridKa CHEP 2009, 21-27 March, Prague Tobias König, Dr. Holger Marten Steinbuch Centre for Computing (SCC) KIT The cooperation of Forschungszentrum Karlsruhe GmbH und Universität
More informationStatus and Integration of AP2 Monitoring and Online Steering
Status and Integration of AP2 Monitoring and Online Steering Daniel Lorenz - University of Siegen Stefan Borovac, Markus Mechtel - University of Wuppertal Ralph Müller-Pfefferkorn Technische Universität
More informationThe Grid-it: the Italian Grid Production infrastructure
n 1 Maria Cristina Vistoli INFN CNAF, Bologna Italy The Grid-it: the Italian Grid Production infrastructure INFN-Grid goals!promote computational grid technologies research & development: Middleware and
More informationCY-01-KIMON operational issues
CY-01-KIMON operational issues Speaker: Kyriacos Neocleous High Performance Computing systems Lab, University Of Cyprus SA1 workshop: Stockholm, Sweden, June 2007 www.eu-egee.org Presentation outline (1)
More informationEGEE a worldwide Grid infrastructure
EGEE a worldwide Grid infrastructure Fabrizio Gagliardi Project Director EGEE CERN, Switzerland IFIC, 6 October 2005 www.eu-egee.org Presentation overview Data intensive science and the rationale for Grid
More informationThe dashboard Grid monitoring framework
The dashboard Grid monitoring framework Benjamin Gaidioz on behalf of the ARDA dashboard team (CERN/EGEE) ISGC 2007 conference The dashboard Grid monitoring framework p. 1 introduction/outline goals of
More informationStatus of Grid Activities in Pakistan. FAWAD SAEED National Centre For Physics, Pakistan
Status of Grid Activities in Pakistan FAWAD SAEED National Centre For Physics, Pakistan 1 Introduction of NCP-LCG2 q NCP-LCG2 is the only Tier-2 centre in Pakistan for Worldwide LHC computing Grid (WLCG).
More informationReport from Italian ROC
Report from Italian ROC Paolo Veronesi for ROC It www.eu-egee.org ARM-7, ROC(111) 15th - 17th May 2006 - Krakow, Outline Changes in ROC structure over the last 3 months (people/institutes involved) People
More informationPoS(EGICF12-EMITC2)110
User-centric monitoring of the analysis and production activities within the ATLAS and CMS Virtual Organisations using the Experiment Dashboard system Julia Andreeva E-mail: Julia.Andreeva@cern.ch Mattia
More informationEELA Operations: A standalone regional dashboard implementation
Proceedings of the Second EELA-2 Conference R. Mayo et al. (Eds.) CIEMAT 2009 2009 The authors. All rights reserved EELA Operations: A standalone regional dashboard implementation Cyril L'Orphelin, Hélène
More informationSPACI & EGEE LCG on IA64
SPACI & EGEE LCG on IA64 Dr. Sandro Fiore, University of Lecce and SPACI December 13 th 2005 www.eu-egee.org Outline EGEE Production Grid SPACI Activity Status of the LCG on IA64 SPACI & EGEE Farm Configuration
More informationBetriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil
Betriebssystem-Virtualisierung auf einem Rechencluster am SCC mit heterogenem Anwendungsprofil Volker Büge 1, Marcel Kunze 2, OIiver Oberst 1,2, Günter Quast 1, Armin Scheurer 1 1) Institut für Experimentelle
More informationRecent grid activities at INFN Catania(*) Roberto Barbera
RecentgridactivitiesatINFNCatania(*) RobertoBarbera workincollaborationwithnicesrl (*) HEPiX/HEPNT2002,Catania,18.04.2002 CHEP2000,10.02.2000 1 RobertoBarbera DipartimentodiFisicadell UniversitàdiCataniaandINFNCatania
More informationDCMS Tier 2/3 prototype infrastructure
DCMS Tier 2/3 prototype infrastructure 1 Anja Vest, Uni Karlsruhe DCMS Meeting in Aachen, Overview LCG Queues/mapping set up Hardware capacities Supported software Summary DCMS overview CMS DCMS: Tier
More informationIntroduction to Grid computing
Introduction to Grid computing The INFNGrid Project Team Introduction This tutorial has been implemented considering as starting point the DataGrid (EDG) tutorial Many thanks to the EDG tutorials team!
More informationVirtualization, Grid, Cloud: Integration Paths for Scientific Computing
Virtualization, Grid, Cloud: Integration Paths for Scientific Computing Or, where and how will my efficient large-scale computing applications be executed? D. Salomoni, INFN Tier-1 Computing Manager Davide.Salomoni@cnaf.infn.it
More informationirods at CC-IN2P3: managing petabytes of data
Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules irods at CC-IN2P3: managing petabytes of data Jean-Yves Nief Pascal Calvat Yonny Cardenas Quentin Le Boulc h
More informationSolution for private cloud computing
The CC1 system Solution for private cloud computing 1 Outline What is CC1? Features Technical details Use cases By scientist By HEP experiment System requirements and installation How to get it? 2 What
More informationData storage services at CC-IN2P3
Centre de Calcul de l Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief Agenda Hardware: Storage on disk. Storage on tape. Software:
More informationOperational Model for E2E links in the NREN/GÉANT2 and NREN/Cross-Border-Fibre supplied optical platform
Operational Model for E2E links in the NREN/GÉANT2 and NREN/Cross-Border-Fibre supplied optical platform 0 Background In the GÉANT2 project plan it is foreseen that end-to-end (E2E) links (essentially
More informationAlternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC
EGEE and glite are registered trademarks Enabling Grids for E-sciencE Alternative models to distribute VO specific software to WLCG sites: a prototype set up at PIC Elisa Lanciotti, Arnau Bria, Gonzalo
More informationForschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. GridKa User Meeting
GridKa User Meeting Forschungszentrum Karlsruhe GmbH Central Information and Communication Technologies Department Hermann-von-Helmholtz-Platz 1 D-76344 Eggenstein-Leopoldshafen Dr. Holger Marten http://grid.fzk.de
More informationBatch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb
Batch and Cloud overview Andrew McNab University of Manchester GridPP and LHCb Overview Assumptions Batch systems The Grid Pilot Frameworks DIRAC Virtual Machines Vac Vcycle Tier-2 Evolution Containers
More informationThe CMS analysis chain in a distributed environment
The CMS analysis chain in a distributed environment on behalf of the CMS collaboration DESY, Zeuthen,, Germany 22 nd 27 th May, 2005 1 The CMS experiment 2 The CMS Computing Model (1) The CMS collaboration
More informationVirtualization of a Cluster Batch System
Virtualization of a Cluster Batch System Christian Baun, Volker Büge, Benjamin Klein, Jens Mielke, Oliver Oberst and Armin Scheurer Die Kooperation von Cluster Batch System Batch system accepts computational
More informationLCG POOL, Distributed Database Deployment and Oracle Services@CERN
LCG POOL, Distributed Database Deployment and Oracle Services@CERN Dirk Düllmann, D CERN HEPiX Fall 04, BNL Outline: POOL Persistency Framework and its use in LHC Data Challenges LCG 3D Project scope and
More informationSee-GRID Project and Business Model
Improvements of the grid infrastructure and services within SEE-GRID Anastas Misev MARNET/MARGI/UKIM Macedonia Introduction SEE-GRID Project series SEE-GRID establish infrastructure SEE-GRID-2 extend infrastructure,
More informationDistributed Database Access in the LHC Computing Grid with CORAL
Distributed Database Access in the LHC Computing Grid with CORAL Dirk Duellmann, CERN IT on behalf of the CORAL team (R. Chytracek, D. Duellmann, G. Govi, I. Papadopoulos, Z. Xie) http://pool.cern.ch &
More informationXTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12
XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines A.Zydroń 18 April 2009 Page 1 of 12 1. Introduction...3 2. XTM Database...4 3. JVM and Tomcat considerations...5 4. XTM Engine...5
More informationLHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch. CERN, July 2007
WLCG Service Schedule LHC schedule: what does it imply for SRM deployment? Jamie.Shiers@cern.ch WLCG Storage Workshop CERN, July 2007 Agenda The machine The experiments The service LHC Schedule Mar. Apr.
More informationDeploying Multiscale Applications on European e-infrastructures
Deploying Multiscale Applications on European e-infrastructures 04/06/2013 Ilya Saverchenko The MAPPER project receives funding from the EC's Seventh Framework Programme (FP7/2007-2013) under grant agreement
More informationHAMBURG ZEUTHEN. DESY Tier 2 and NAF. Peter Wegner, Birgit Lewendel for DESY-IT/DV. Tier 2: Status and News NAF: Status, Plans and Questions
DESY Tier 2 and NAF Peter Wegner, Birgit Lewendel for DESY-IT/DV Tier 2: Status and News NAF: Status, Plans and Questions Basics T2: 1.5 average Tier 2 are requested by CMS-groups for Germany Desy commitment:
More informationBig Data and Storage Management at the Large Hadron Collider
Big Data and Storage Management at the Large Hadron Collider Dirk Duellmann CERN IT, Data & Storage Services Accelerating Science and Innovation CERN was founded 1954: 12 European States Science for Peace!
More informationDeliverable DS4.3.2: Report on Development Infrastructure Usage and Adoption
05-10-2010 Deliverable DS4.3.2 Contractual Date: 30-06-2010 Actual Date: 05-10-2010 Grant Agreement No.: 238875 Activity: SA4 Task Item: T3 Nature of Deliverable: R (Report) Dissemination Level: PU (Public)
More informationWeb Hosting. E-Mail Hosting. Cloud File Hosting. The Genio Group (214) 732-7411 info@thegeniogroup.com www.thegeniogroup.com
Web Hosting E-Mail Hosting Cloud File Hosting Genio Hosting Servers All of Genio s Hosting Servers run on Apple hardware running Mac OS X Server. Mac OS X Server leverages the computing power of 64-bit
More informationInstruments in Grid: the New Instrument Element
Instruments in Grid: the New Instrument Element C. Vuerli (1,2), G. Taffoni (1,2), I. Coretti (1), F. Pasian (1,2), P. Santin (1), M. Pucillo (1) (1) INAF Astronomical Observatory of Trieste (2) INAF Informative
More informationConfiguration Management of Massively Scalable Systems
1 KKIO 2005 Configuration Management of Massively Scalable Systems Configuration Management of Massively Scalable Systems Marcin Jarząb, Krzysztof Zieliński, Jacek Kosiński SUN Center of Excelence Department
More informationGrid Computing in Aachen
GEFÖRDERT VOM Grid Computing in Aachen III. Physikalisches Institut B Berichtswoche des Graduiertenkollegs Bad Honnef, 05.09.2008 Concept of Grid Computing Computing Grid; like the power grid, but for
More informationDeliverable D 6.1 Website
Biogas2PEM-FC Biogas Reforming and Valorisation through PEM Fuel Cells FP7-SME-2012, Grant Agreement No. 314940 Deliverable D 6.1 Website Deliverable details Deliverable version: v1.0 Date of delivery:
More informationUnterstützung datenintensiver Forschung am KIT Aktivitäten, Dienste und Erfahrungen
Unterstützung datenintensiver Forschung am KIT Aktivitäten, Dienste und Erfahrungen Achim Streit Steinbuch Centre for Computing (SCC) KIT Universität des Landes Baden-Württemberg und nationales Forschungszentrum
More informationAN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID
AN APPROACH TO DEVELOPING BUSINESS PROCESSES WITH WEB SERVICES IN GRID R. D. Goranova 1, V. T. Dimitrov 2 Faculty of Mathematics and Informatics, University of Sofia S. Kliment Ohridski, 1164, Sofia, Bulgaria
More informationvtiger Customer Portal 4.2 User Manual
- 1 - vtiger Customer Portal 4.2 User Manual (Version 2.0) - 2 - Table of Contents 1. Introduction... 3 2. Installing vtiger Customer Portal... 4 2.1. System Requirements... 4 2.2. Installation Prerequisites...
More informationHIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS EKSPERYMENTY FIZYKI WYSOKICH ENERGII W SIECIACH KOMPUTEROWYCH GRID. 1.
Computer Science Vol. 9 2008 Andrzej Olszewski HIGH ENERGY PHYSICS EXPERIMENTS IN GRID COMPUTING NETWORKS The demand for computing resources used for detector simulations and data analysis in High Energy
More informationMonitoring Remedy with BMC Solutions
Monitoring Remedy with BMC Solutions Overview How does BMC Software monitor Remedy with our own solutions? The challenge is many fold with a solution like Remedy and this does not only apply to Remedy,
More informationCERN local High Availability solutions and experiences. Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006
CERN local High Availability solutions and experiences Thorsten Kleinwort CERN IT/FIO WLCG Tier 2 workshop CERN 16.06.2006 1 Introduction Different h/w used for GRID services Various techniques & First
More informationOPEN CALL FOR TENDERS No MARE/2014/14. Appendix 2. EUMOFA IT architecture and design
EUROPEAN COMMISSION DIRECTORATE-GENERAL FOR MARITIME AFFAIRS AND FISHERIES INTERNATIONAL AFFAIRS AND MARKETS TRADE AND MARKETS OPEN CALL FOR TENDERS No MARE/2014/14 Appendix 2 EUMOFA IT architecture and
More informationIT-INFN-CNAF Status Update
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, 10-11 December 2009 Stefano Zani 10/11/2009 Stefano Zani INFN CNAF (TIER1 Staff) 1 INFN CNAF CNAF is the main computing facility of the INFN Core business:
More informationThis presentation provides an overview of the architecture of the IBM Workload Deployer product.
This presentation provides an overview of the architecture of the IBM Workload Deployer product. Page 1 of 17 This presentation starts with an overview of the appliance components and then provides more
More informationFast, Easy to use and On-demand A Content Platform from the 21st Century
OpenIMS Open Information Management Server Fast, Easy to use and On-demand A Content Platform from the 21st Century E-mail Management Server Whitepaper Open Information Management Server www.openims.com
More informationGrid Site Monitoring tools developed and used at at SCL
Grid Site Monitoring tools developed and used at at SCL V. Slavnić, B. Acković, D. Vudragović, A. Balaž, A. Belić Scientific Computing Laboratory Institute of Physics Belgrade Pregrevica 118, 11080 Belgrade,
More informationReport of the LHC Computing Grid Project. Software Management Process RTAG CERN
Report of the LHC Computing Grid Project Software Management Process RTAG Marco Cattaneo, Gabriele Cosmo, Simon George, Fons Rademakers (chair), Stephan Wynhoff CERN 6 May 2002 Table of Contents 1 Chair
More informationWHITE PAPER Managing Linux in the Enterprise: The Red Hat Network Approach
Global Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com WHITE PAPER Managing Linux in the Enterprise: The Red Hat Network Approach Sponsored by: Red Hat Tim
More informationProduct Training Services. Training Options and Procedures for JobScheduler and YADE
Product Services Product Services Options and Procedures for JobScheduler and YADE 2 Contents Product Services JobScheduler Levels Level: JobScheduler Operations Level: JobScheduler Installation Level:
More informationOnline Fuzzy-C-Means clustering
Online Fuzzy-C-Means clustering Authors: Author s Addresses: Contact: Dezső Kancsár, Ágnes B. Simon H-1157 Budapest, Nyírpalota u. 79/C 2/8; College of Nyíregyháza, Rákóczi út 69. berszoft@ax.hu, simona@nyf.hu
More informationSteinbuch Centre for Computing (SCC) The Information Technology Centre of KIT
Steinbuch Centre for Computing (SCC) The Information Technology Centre of KIT SCIENTIFIC COMPUTING, HPC AND GRIDS KIT the cooperation of Forschungszentrum Karlsruhe GmbH and Universität Karlsruhe (TH)
More informationHP Service Manager Compatibility Matrix
HP Service Manager Compatibility Matrix Software Version 9.21 January 12, 2011 Click one of the following links to see more detailed information. Tier Definitions Servers Applications Support Windows Client
More informationLHCONE Operational Framework
LHCONE Operational Framework Part 1 : principles and ideas for the operational model Part 2 : LHCONE VRF operational handbook Part 3 : Next step Xavier Jeannin RENATER 2013/01/28 Part 1 : principles and
More informationHISP: a data-driven portal for hadron therapy
HISP: a data-driven portal for hadron therapy Faustin Laurentiu Roman CERN / IFIC Prototype architecture Tools, implementation & services Conclusions (& demo) 1 One slide situation: ereferral and escience
More informationTHE OPEN SOURCE DEVELOPER REPORT
THE OPEN SOURCE DEVELOPER REPORT 2009 ECLIPSE COMMUNITY SURVEY M A Y 2 0 0 9 COPYRIGHT 2009 THE ECLIPSE FOUNDATION ALL RIGHTS RESERVED THIS WORK I S LIC EN SE D UND ER TH E C RE ATI VE COMMON S ATTRIBUT
More informationHow To Install Linux Titan
Linux Titan Distribution Presented By: Adham Helal Amgad Madkour Ayman El Sayed Emad Zakaria What Is a Linux Distribution? What is a Linux Distribution? The distribution contains groups of packages and
More informationCloud Computing Architecture with OpenNebula HPC Cloud Use Cases
NASA Ames NASA Advanced Supercomputing (NAS) Division California, May 24th, 2012 Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases Ignacio M. Llorente Project Director OpenNebula Project.
More informationExpansion Through Acquisitions
Oracle, Open Source and Open Standards Edgars RuĦăis Oracle Latvijā Oracle Corporation World s largest enterprise software vendor 275,000+ global customers 30,000+ applications customers
More informationUS NSF s Scientific Software Innovation Institutes
US NSF s Scientific Software Innovation Institutes S 2 I 2 awards invest in long-term projects which will realize sustained software infrastructure that is integral to doing transformative science. (Can
More informationA Cloud-based Architecture for Visual Effect Rendering System
A Cloud-based Architecture for Visual Effect System Teng-Kai Fan Researcher Chunghwa Telecom Labs 20 August 2015 Outline Preliminary Motivation & Objective Cloud-based Visual Effect System 2/24 Copyright
More informationJBoss Enterprise MIDDLEWARE
JBoss Enterprise MIDDLEWARE WHAT IS IT? JBoss Enterprise Middleware integrates and hardens the latest enterprise-ready features from JBoss community projects into supported, stable, enterprise-class middleware
More informationPortal for ArcGIS. Satish Sankaran Robert Kircher
Portal for ArcGIS Satish Sankaran Robert Kircher ArcGIS A Complete GIS Data Management Planning & Analysis Field Mobility Operational Awareness Constituent Engagement End to End Integration Collect, Organize,
More information