HTCondor at the RAL Tier-1
|
|
|
- Dinah Young
- 10 years ago
- Views:
Transcription
1 HTCondor at the RAL Tier-1 Andrew Lahiff, Alastair Dewhurst, John Kelly, Ian Collier, James Adams STFC Rutherford Appleton Laboratory HTCondor Week 2014
2 Outline Overview of HTCondor at RAL Monitoring Multi-core jobs Dynamically-provisioned worker nodes 2
3 Introduction RAL is a Tier-1 for all 4 LHC experiments Provide computing & disk resources, and tape for custodial storage of data In terms of Tier-1 computing requirements, RAL provides 2% ALICE 13% ATLAS 8% CMS 32% LHCb Also support ~12 non-lhc experiments, including non-hep Computing resources 784 worker nodes, over 14K cores Generally have 40-60K jobs submitted per day 3
4 Torque/Maui had been used for many years Many issues Migration to HTCondor Severity & number of problems increased as size of farm increased Migration 2012 Aug Started evaluating alternatives to Torque/Maui (LSF, Grid Engine, Torque 4, HTCondor, SLURM) 2013 Jun Began testing HTCondor with ATLAS & CMS 2013 Aug Choice of HTCondor approved by management 2013 Sep HTCondor declared production service Moved 50% of pledged CPU resources to HTCondor 2013 Nov Migrated remaining resources to HTCondor 4
5 Experience Very stable operation Experience so far Staff don t need to spend all their time fire-fighting problems Job start rate much higher than Torque/Maui, even when throttled Farm utilization much better Very good support 5
6 Status Version Currently Trying to stay up to date with the latest stable release Features Partitionable slots Hierarchical accounting groups HA central managers PID namespaces Python API condor_gangliad In progress CPU affinity being phased in cgroups has been tested, probably will be phased-in next 6
7 Computing elements All job submission to RAL is via the Grid No local users Currently have 5 CEs, schedd on each: 2 CREAM CEs 3 ARC CEs CREAM doesn t currently support HTCondor We developed the missing functionality ourselves Will feed this back so that it can be included in an official release ARC better But didn t originally handle partitionable slots, passing CPU/memory requirements to HTCondor, We wrote lots of patches, all included in upcoming release Will make it easier for more European sites to move to HTCondor 7
8 HTCondor in the UK Increasing usage of HTCondor in WLCG sites in the UK : None : RAL Tier-1, RAL Tier-2, Bristol, Oxford (in progress) The future 7 sites currently running Torque/Maui Considering moving to HTCondor or will move if others do 8
9 Monitoring 9
10 Jobs monitoring Useful to store details about completed jobs in a database What we currently do Nightly cron reads HTCondor history files, inserts data into MySQL Problems Currently only use subset of content of job ClassAds Could try to put in everything What happens then if jobs have new attributes? Modify DB table? Experience with similar database for Torque As database grew in size, queries took longer & longer Database tuning important Is there a better alternative? 10
11 Jobs monitoring CASTOR team at RAL have been testing Elasticsearch Why not try using it with HTCondor? Elasticsearch ELK stack Logstash: parses log files Elasticsearch: search & analyze data in real-time Kibana: data visualization HTCondor history files Logstash Elastic search Kibana Hardware setup Test cluster of 13 servers (old diskservers & worker nodes) But 3 servers could handle 16 GB of CASTOR logs per day Adding HTCondor Wrote config file for Logstash to enable history files to be parsed Add Logstash to machines running schedds 11
12 Jobs monitoring Can see full job ClassAds 12
13 Jobs monitoring Custom plots E.g. completed jobs by schedd 13
14 Jobs monitoring Custom dashboards 14
15 Benefits Easy to setup Took less than a day to setup the initial cluster Seems to be able to handle the load from HTCondor For us (so far): < 1 GB, < 100K documents per day Arbitrary queries Queries are faster than using condor_history Horizontal construction Need more capacity? Just add more nodes Jobs monitoring 15
16 Multi-core jobs 16
17 Multi-core jobs Situation so far ATLAS have been running multi-core jobs at RAL since November CMS submitted a few test jobs, will submit more eventually Interest so far only for multi-core jobs, not whole-node jobs Only 8-core jobs Our aims Fully dynamic No manual partitioning of resources Number of running multi-core jobs determined by group quotas 17
18 Defrag daemon Essential to allow multi-core jobs to run Want to drain 8 cores only. Changed: Multi-core jobs DEFRAG_WHOLE_MACHINE_EXPR = Cpus == TotalCpus && Offline=!=True! to DEFRAG_WHOLE_MACHINE_EXPR = (Cpus >= 8) && Offline=!=True! Which machines more desirable to drain? Changed DEFRAG_RANK = -ExpectedMachineGracefulDrainingBadput! to DEFRAG_RANK = ifthenelse(cpus >= 8, -10, (TotalCpus - Cpus)/(8.0 - Cpus))! Why make this change? With default DEFRAG_RANK, only older full 8-core WNs were being selected for draining Now: (Number of slots that can be freed up)/(number of needed cores) For us this does a better job of finding the best worker nodes to drain 18
19 Multi-core jobs Effect of changing DEFRAG_RANK Running multi-core jobs No change in the number of concurrent draining machines Rate in increase in number of running multi-core jobs much higher 19
20 Multi-core jobs Group quotas Added accounting groups for ATLAS and CMS multi-core jobs Force accounting groups to be specified for jobs using SUBMIT_EXPRS Easy to include groups for multi-core jobs AccountingGroup =!! ifthenelse(regexp("patl",owner) && RequestCpus > 1, group_atlas.prodatls_multicore", \! ifthenelse(regexp("patl",owner), group_atlas.prodatls", \!! SUBMIT_EXPRS = $(SUBMIT_EXPRS) AccountingGroup! Negotiator Modified GROUP_SORT_EXPR so that the order is: High priority groups (Site Usability Monitor tests) Multi-core groups Remaining groups Helps to ensure multi-core slots not lost too quickly 20
21 Defrag daemon issues No knowledge of demand for multi-core jobs Multi-core jobs Always drains the same number of nodes, irrespective of demand Can result in large amount of wasted resources Wrote simple cron script which adjusts defrag daemon config based on demand Currently very simple, considers 3 cases: Many idle multi-core jobs, few running multi-core jobs Need aggressive draining Many idle multi-core jobs, many running multi-core jobs Less agressive draining Otherwise Very little draining May need to make changes when other VOs start submitting multi-core jobs in bulk 21
22 Multi-core jobs Recent ATLAS activity Running & idle multi-core jobs Gaps in submission by ATLAS results in loss of multi-core slots. Number of whole machines & draining machines Data from condor_gangliad Significantly reduced CPU wastage due to the cron 22
23 Other issues Defrag daemon designed for whole-node, not multi-core Won t drain nodes already running multi-core jobs Multi-core jobs Ideally may want to run multiple multi-core jobs per worker node Would be good to be able to run short jobs while waiting for slots to become available for multi-core jobs On other batch systems, backfill can do this Time taken for 8 jobs to drain - lots of opportunity to run short jobs 23
24 Next step: enabling backfill ARC CE adds custom attribute to jobs: JobTimeLimit Can have knowledge of job run times Defrag daemon drains worker nodes Multi-core jobs Problem: machine can t run any jobs at this time, including short jobs Alternative idea: Python script (run as a cron) which plays the same role as defrag daemon But doesn t actually drain machines Have a custom attribute on all startds, e.g. NodeDrain Change this instead START expression set so that: If NodeDrain false: allow any jobs to start If NodeDrain true: allow only short jobs under certain conditions, e.g. for a limited time after draining started Provided (some) VOs submit short jobs, should be able to reduce wasted resources due to draining 24
25 Dynamically-provisioned worker nodes 25
26 Prototype cloud StratusLab (based on OpenNebula) iscsi & LVM based persistent disk storage (18 TB) 800 cores No EC2 interface Production cloud (Very) early stage of deployment OpenNebula Private clouds at RAL 900 cores, 3.5 TB RAM, ~1 PB raw storage for Ceph Aims Integrate with batch system, eventually without partitioned resources First step: allow the batch system to expand into the cloud Avoid running additional third-party and/or complex services Use existing functionality in HTCondor as much as possible Should be as simple as possible 26
27 Provisioning worker nodes Use HTCondor s existing power management features Send appropriate offline ClassAd(s) to the collector Hostname used is a random string Represents a type of VM, rather than specific machines condor_rooster Provisions resources Configured to run appropriate command to instantiate a VM When there are idle jobs Negotiator can match jobs to the offline ClassAds condor_rooster daemon notices this match Instantiates a VM Image has HTCondor pre-installed & configured, can join the pool HTCondor on the VM controls the VM s lifetime START expression New jobs allowed to start only for a limited time after VM instantiated HIBERNATE expression VM is shutdown after machine has been idle for too long 27
28 Testing in production Initial test with production HTCondor pool Ran around 11,000 real jobs, including jobs from all LHC VOs Testing Started with 4-core 12GB VMs, then changed to 3-core 9GB VMs Hypervisors enabled Started using 3-core VMs condor_rooster disabled 28
29 Summary Due to scalability problems with Torque/Maui, migrated to HTCondor last year We are happy with the choice we made based on our requirements Confident that the functionality & scalability of HTCondor will meet our needs for the foreseeable future Multi-core jobs working well Looking forward to more VOs submitting multi-core jobs Dynamically-provisioned worker nodes Expect to have in production later this year 29
30 Thank you! 30
HTCondor within the European Grid & in the Cloud
HTCondor within the European Grid & in the Cloud Andrew Lahiff STFC Rutherford Appleton Laboratory HEPiX 2015 Spring Workshop, Oxford The Grid Introduction Computing element requirements Job submission
Virtualisation Cloud Computing at the RAL Tier 1. Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013
Virtualisation Cloud Computing at the RAL Tier 1 Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013 Virtualisation @ RAL Context at RAL Hyper-V Services Platform Scientific Computing Department
Dynamic Resource Provisioning with HTCondor in the Cloud
Dynamic Resource Provisioning with HTCondor in the Cloud Ryan Taylor Frank Berghaus 1 Overview Review of Condor + Cloud Scheduler system Condor job slot configuration Dynamic slot creation Automatic slot
Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb
Batch and Cloud overview Andrew McNab University of Manchester GridPP and LHCb Overview Assumptions Batch systems The Grid Pilot Frameworks DIRAC Virtual Machines Vac Vcycle Tier-2 Evolution Containers
A Year of HTCondor Monitoring. Lincoln Bryant Suchandra Thapa
A Year of HTCondor Monitoring Lincoln Bryant Suchandra Thapa HTCondor Week 2015 May 21, 2015 Analytics vs. Operations Two parallel tracks in mind: o Operations o Analytics Operations needs to: o Observe
Solution for private cloud computing
The CC1 system Solution for private cloud computing 1 Outline What is CC1? Features Technical details Use cases By scientist By HEP experiment System requirements and installation How to get it? 2 What
PES. Batch virtualization and Cloud computing. Part 1: Batch virtualization. Batch virtualization and Cloud computing
Batch virtualization and Cloud computing Batch virtualization and Cloud computing Part 1: Batch virtualization Tony Cass, Sebastien Goasguen, Belmiro Moreira, Ewan Roche, Ulrich Schwickerath, Romain Wartel
U-LITE Network Infrastructure
U-LITE: a proposal for scientific computing at LNGS S. Parlati, P. Spinnato, S. Stalio LNGS 13 Sep. 2011 20 years of Scientific Computing at LNGS Early 90s: highly centralized structure based on VMS cluster
Log managing at PIC. A. Bruno Rodríguez Rodríguez. Port d informació científica Campus UAB, Bellaterra Barcelona. December 3, 2013
Log managing at PIC A. Bruno Rodríguez Rodríguez Port d informació científica Campus UAB, Bellaterra Barcelona December 3, 2013 Bruno Rodríguez (PIC) Log managing at PIC December 3, 2013 1 / 21 What will
Virtualization, Grid, Cloud: Integration Paths for Scientific Computing
Virtualization, Grid, Cloud: Integration Paths for Scientific Computing Or, where and how will my efficient large-scale computing applications be executed? D. Salomoni, INFN Tier-1 Computing Manager [email protected]
Tier0 plans and security and backup policy proposals
Tier0 plans and security and backup policy proposals, CERN IT-PSS CERN - IT Outline Service operational aspects Hardware set-up in 2007 Replication set-up Test plan Backup and security policies CERN Oracle
Evaluation of Nagios for Real-time Cloud Virtual Machine Monitoring
University of Victoria Faculty of Engineering Fall 2009 Work Term Report Evaluation of Nagios for Real-time Cloud Virtual Machine Monitoring Department of Physics University of Victoria Victoria, BC Michael
Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases
NASA Ames NASA Advanced Supercomputing (NAS) Division California, May 24th, 2012 Cloud Computing Architecture with OpenNebula HPC Cloud Use Cases Ignacio M. Llorente Project Director OpenNebula Project.
Monitoring HTCondor with Ganglia
Monitoring HTCondor with Ganglia Ganglia Overview Scalable distributed monitoring for HPC clusters Two daemons gmond every host; collects and send metrics gmetad single host; persists metrics from local
Maximizing SQL Server Virtualization Performance
Maximizing SQL Server Virtualization Performance Michael Otey Senior Technical Director Windows IT Pro SQL Server Pro 1 What this presentation covers Host configuration guidelines CPU, RAM, networking
Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS
Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS Workshop on the Future of Big Data Management 27-28 June 2013 Philip Kershaw Centre for Environmental Data Archival
Evaluation and implementation of CEP mechanisms to act upon infrastructure metrics monitored by Ganglia
Project report CERN Summer Student Programme Evaluation and implementation of CEP mechanisms to act upon infrastructure metrics monitored by Ganglia Author: Martin Adam Supervisors: Cristovao Cordeiro,
Cloud Server. Parallels. Key Features and Benefits. White Paper. www.parallels.com
Parallels Cloud Server White Paper Key Features and Benefits www.parallels.com Table of Contents Introduction... 3 Key Features... 3 Distributed Cloud Storage (Containers and Hypervisors)... 3 Rebootless
Computing in High- Energy-Physics: How Virtualization meets the Grid
Computing in High- Energy-Physics: How Virtualization meets the Grid Yves Kemp Institut für Experimentelle Kernphysik Universität Karlsruhe Yves Kemp Barcelona, 10/23/2006 Outline: Problems encountered
Clusters in the Cloud
Clusters in the Cloud Dr. Paul Coddington, Deputy Director Dr. Shunde Zhang, Compu:ng Specialist eresearch SA October 2014 Use Cases Make the cloud easier to use for compute jobs Par:cularly for users
STeP-IN SUMMIT 2013. June 18 21, 2013 at Bangalore, INDIA. Performance Testing of an IAAS Cloud Software (A CloudStack Use Case)
10 th International Conference on Software Testing June 18 21, 2013 at Bangalore, INDIA by Sowmya Krishnan, Senior Software QA Engineer, Citrix Copyright: STeP-IN Forum and Quality Solutions for Information
Condor: Grid Scheduler and the Cloud
Condor: Grid Scheduler and the Cloud Matthew Farrellee Senior Software Engineer, Red Hat 1 Agenda What is Condor Architecture Condor s ClassAd Language Common Use Cases Virtual Machine management Cloud
SUSE Cloud Installation: Best Practices Using a SMT, Xen and Ceph Storage Environment
Best Practices Guide www.suse.com SUSE Cloud Installation: Best Practices Using a SMT, Xen and Ceph Storage Environment Written by B1 Systems GmbH Table of Contents Introduction...3 Use Case Overview...3
DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group
DSS High performance storage pools for LHC Łukasz Janyst on behalf of the CERN IT-DSS group CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Introduction The goal of EOS is to provide a
Dynamic Slot Tutorial. Condor Project Computer Sciences Department University of Wisconsin-Madison
Dynamic Slot Tutorial Condor Project Computer Sciences Department University of Wisconsin-Madison Outline Why we need partitionable slots How they ve worked since 7.2 What s new in 7.8 What s still left
Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept
Integration of Virtualized Workernodes in Batch Queueing Systems, Dr. Armin Scheurer, Oliver Oberst, Prof. Günter Quast INSTITUT FÜR EXPERIMENTELLE KERNPHYSIK FAKULTÄT FÜR PHYSIK KIT University of the
WHITE PAPER Redefining Monitoring for Today s Modern IT Infrastructures
WHITE PAPER Redefining Monitoring for Today s Modern IT Infrastructures Modern technologies in Zenoss Service Dynamics v5 enable IT organizations to scale out monitoring and scale back costs, avoid service
Interoperating Cloud-based Virtual Farms
Stefano Bagnasco, Domenico Elia, Grazia Luparello, Stefano Piano, Sara Vallero, Massimo Venaruzzo For the STOA-LHC Project Interoperating Cloud-based Virtual Farms The STOA-LHC project 1 Improve the robustness
Database Services for Physics @ CERN
Database Services for Physics @ CERN Deployment and Monitoring Radovan Chytracek CERN IT Department Outline Database services for physics Status today How we do the services tomorrow? Performance tuning
Deploying a distributed data storage system on the UK National Grid Service using federated SRB
Deploying a distributed data storage system on the UK National Grid Service using federated SRB Manandhar A.S., Kleese K., Berrisford P., Brown G.D. CCLRC e-science Center Abstract As Grid enabled applications
Getting Started Hacking on OpenNebula
LinuxTag 2013 Berlin, Germany, May 22nd Getting Started Hacking on OpenNebula Carlos Martín Project Engineer Acknowledgments The research leading to these results has received funding from Comunidad de
How To Run A Cloud Server On A Server Farm (Cloud)
StratusLab: Darn Simple Cloud Charles (Cal) Loomis (CNRS/LAL & SixSq Sàrl) FOSDEM 13: Cloud Devroom (3 February 2013) StratusLab What is it? Complete Infrastructure as a Service (IaaS) cloud distribution
Scaling Graphite Installations
Scaling Graphite Installations Graphite basics Graphite is a web based Graphing program for time series data series plots. Written in Python Consists of multiple separate daemons Has it's own storage backend
SCF/FEF Evaluation of Nagios and Zabbix Monitoring Systems. Ed Simmonds and Jason Harrington 7/20/2009
SCF/FEF Evaluation of Nagios and Zabbix Monitoring Systems Ed Simmonds and Jason Harrington 7/20/2009 Introduction For FEF, a monitoring system must be capable of monitoring thousands of servers and tens
ADAM 5.5. System Requirements
ADAM 5.5 System Requirements 1 1. Overview The schema below shows an overview of the ADAM components that will be installed and set up. ADAM Server: hosts the ADAM core components. You must install the
Deployment of Private, Hybrid & Public Clouds with OpenNebula
EL / LAK (FOSS) 2010 May 14th, 2010 Deployment of Private, Hybrid & Public Clouds with OpenNebula University Complutense of Madrid The Anatomy of an IaaS Cloud Deployment of Private, Hybrid & Public Clouds
Forschungszentrum Karlsruhe in der Helmholtz - Gemeinschaft. Holger Marten. Holger. Marten at iwr. fzk. de www.gridka.de
Tier-2 cloud Holger Marten Holger. Marten at iwr. fzk. de www.gridka.de 1 GridKa associated Tier-2 sites spread over 3 EGEE regions. (4 LHC Experiments, 5 (soon: 6) countries, >20 T2 sites) 2 region DECH
A New Approach to Network Visibility at UBC. Presented by the Network Management Centre and Wireless Infrastructure Teams
A New Approach to Network Visibility at UBC Presented by the Network Management Centre and Wireless Infrastructure Teams Agenda Business Drivers Technical Overview Network Packet Broker Tool Network Monitoring
SURFsara HPC Cloud Workshop
SURFsara HPC Cloud Workshop doc.hpccloud.surfsara.nl UvA workshop 2016-01-25 UvA HPC Course Jan 2016 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current
Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary
Big Science and Big Data Dirk Duellmann, CERN Apache Big Data Europe 28 Sep 2015, Budapest, Hungary 16/02/2015 Real-Time Analytics: Making better and faster business decisions 8 The ATLAS experiment
Next Generation Tier 1 Storage
Next Generation Tier 1 Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies, Matthew Viljoen HEPiX Beijing 16th October 2012 Why are we doing this?
Computing at the HL-LHC
Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory Group ALICE: Pierre Vande Vyvre, Thorsten Kollegger, Predrag Buncic; ATLAS: David Rousseau, Benedetto Gorini,
HPC-Nutzer Informationsaustausch. The Workload Management System LSF
HPC-Nutzer Informationsaustausch The Workload Management System LSF Content Cluster facts Job submission esub messages Scheduling strategies Tools and security Future plans 2 von 10 Some facts about the
www.see-grid-sci.eu Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009
SEE-GRID-SCI Virtualization and Grid Computing with XEN www.see-grid-sci.eu Regional SEE-GRID-SCI Training for Site Administrators Institute of Physics Belgrade March 5-6, 2009 Milan Potocnik University
Das HappyFace Meta-Monitoring Framework
Das HappyFace Meta-Monitoring Framework B. Berge, M. Heinrich, G. Quast, A. Scheurer, M. Zvada, DPG Frühjahrstagung Karlsruhe, 28. März 1. April 2011 KIT University of the State of Baden-Wuerttemberg and
Adaptive Resource Optimizer For Optimal High Performance Compute Resource Utilization
Technical Backgrounder Adaptive Resource Optimizer For Optimal High Performance Compute Resource Utilization July 2015 Introduction In a typical chip design environment, designers use thousands of CPU
ClearPass Policy Manager 6.3
ClearPass Policy Manager 6.3 Tech Note: Installing or Upgrading on a Virtual Machine This document describes the procedures for installing and upgrading ClearPass Policy Manager 6.3 on a Virtual Machine.
Mirantis www.mirantis.com/training
TM Mirantis www.mirantis.com/training Goals Understand OpenStack purpose and use cases Understand OpenStack ecosystem o history o projects Understand OpenStack architecture o logical architecture o components
Datasheet FUJITSU Software ServerView Cloud Monitoring Manager V1.0
Datasheet FUJITSU Software ServerView Cloud Monitoring Manager V1.0 Datasheet FUJITSU Software ServerView Cloud Monitoring Manager V1.0 A Monitoring Cloud Service for Enterprise OpenStack Systems Cloud
Provisioning and Resource Management at Large Scale (Kadeploy and OAR)
Provisioning and Resource Management at Large Scale (Kadeploy and OAR) Olivier Richard Laboratoire d Informatique de Grenoble (LIG) Projet INRIA Mescal 31 octobre 2007 Olivier Richard ( Laboratoire d Informatique
G-Cloud 6 brightsolid Secure Cloud Servers. Service Definition Document
G-Cloud 6 brightsolid Service Definition Document December 2014 Contents 1 Introduction... 3 2 An overview of the G-Cloud Service... 3 3 Information assurance... 3 4 Backup and Disaster Recovery... 3 5
Microsoft Hyper-V chose a Primary Server Virtualization Platform
Roger Shupert, Integration Specialist } Lake Michigan College has been using Microsoft Hyper-V as it s primary server virtualization platform since 2008, in this presentation we will discuss the following;
IaaS Cloud Architectures: Virtualized Data Centers to Federated Cloud Infrastructures
IaaS Cloud Architectures: Virtualized Data Centers to Federated Cloud Infrastructures Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Introduction
OpenNebula Open Souce Solution for DC Virtualization. C12G Labs. Online Webinar
OpenNebula Open Souce Solution for DC Virtualization C12G Labs Online Webinar What is OpenNebula? Multi-tenancy, Elasticity and Automatic Provision on Virtualized Environments I m using virtualization/cloud,
The CMS analysis chain in a distributed environment
The CMS analysis chain in a distributed environment on behalf of the CMS collaboration DESY, Zeuthen,, Germany 22 nd 27 th May, 2005 1 The CMS experiment 2 The CMS Computing Model (1) The CMS collaboration
Running Oracle Databases in a z Systems Cloud environment
Running Oracle Databases in a z Systems Cloud environment Sam Amsavelu [email protected] ISV & Channels Technical Sales - Oracle IBM Advanced Technical Skills (ATS), America Technical University/Symposia
Log Analysis with the ELK Stack (Elasticsearch, Logstash and Kibana) Gary Smith, Pacific Northwest National Laboratory
Log Analysis with the ELK Stack (Elasticsearch, Logstash and Kibana) Gary Smith, Pacific Northwest National Laboratory A Little Context! The Five Golden Principles of Security! Know your system! Principle
SUSE Cloud Installation: Best Practices Using an Existing SMT and KVM Environment
Best Practices Guide www.suse.com SUSE Cloud Installation: Best Practices Using an Existing SMT and KVM Environment Written by B1 Systems GmbH Table of Contents Introduction...3 Use Case Overview...3 Hardware
Enterprise Deployment: Laserfiche 8 in a Virtual Environment. White Paper
Enterprise Deployment: Laserfiche 8 in a Virtual Environment White Paper August 2008 The information contained in this document represents the current view of Compulink Management Center, Inc on the issues
Traditional v/s CONVRGD
Traditional v/s CONVRGD Traditional Virtualization Stack Converged Virtualization Infrastructure with HCE/HSE Data protection software applications PDU Backup Servers + Virtualization Storage Switch HA
Deploying Business Virtual Appliances on Open Source Cloud Computing
International Journal of Computer Science and Telecommunications [Volume 3, Issue 4, April 2012] 26 ISSN 2047-3338 Deploying Business Virtual Appliances on Open Source Cloud Computing Tran Van Lang 1 and
Log management with Logstash and Elasticsearch. Matteo Dessalvi
Log management with Logstash and Elasticsearch Matteo Dessalvi HEPiX 2013 Outline Centralized logging. Logstash: what you can do with it. Logstash + Redis + Elasticsearch. Grok filtering. Elasticsearch
Comparing Open Source Private Cloud (IaaS) Platforms
Comparing Open Source Private Cloud (IaaS) Platforms Lance Albertson OSU Open Source Lab Associate Director of Operations [email protected] / @ramereth About me OSU Open Source Lab Server hosting for Open
Postgres on OpenStack
Postgres on OpenStack Dave Page 18/9/2014 2014 EnterpriseDB Corporation. All rights reserved. 1 Introduction PostgreSQL: Core team member pgadmin lead developer Web/sysadmin teams PGCAC/PGEU board member
An objective comparison test of workload management systems
An objective comparison test of workload management systems Igor Sfiligoi 1 and Burt Holzman 1 1 Fermi National Accelerator Laboratory, Batavia, IL 60510, USA E-mail: [email protected] Abstract. The Grid
Windows Server 2008 R2 Essentials
Windows Server 2008 R2 Essentials Installation, Deployment and Management 2 First Edition 2010 Payload Media. This ebook is provided for personal use only. Unauthorized use, reproduction and/or distribution
How To Use Elasticsearch
Elasticsearch, Logstash, and Kibana (ELK) Dwight Beaver [email protected] Sean Hutchison [email protected] January 2015 2014 Carnegie Mellon University This material is based upon work funded and supported
ATLAS job monitoring in the Dashboard Framework
ATLAS job monitoring in the Dashboard Framework J Andreeva 1, S Campana 1, E Karavakis 1, L Kokoszkiewicz 1, P Saiz 1, L Sargsyan 2, J Schovancova 3, D Tuckett 1 on behalf of the ATLAS Collaboration 1
Building Clouds with OpenNebula 2.2 and StratusLab
Contrail Summer School June 2011, France Building Clouds with OpenNebula 2.2 and StratusLab Javier Fontán Muiños dsa-research.org Distributed Systems Architecture Research Group Universidad Complutense
Hadoop: Embracing future hardware
Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop
Oracle Hyperion Financial Management Virtualization Whitepaper
Oracle Hyperion Financial Management Virtualization Whitepaper Oracle Hyperion Financial Management Virtualization Whitepaper TABLE OF CONTENTS Overview... 3 Benefits... 4 HFM Virtualization testing...
Ubuntu OpenStack on VMware vsphere: A reference architecture for deploying OpenStack while limiting changes to existing infrastructure
TECHNICAL WHITE PAPER Ubuntu OpenStack on VMware vsphere: A reference architecture for deploying OpenStack while limiting changes to existing infrastructure A collaboration between Canonical and VMware
OpenNebula Open Souce Solution for DC Virtualization
13 th LSM 2012 7 th -12 th July, Geneva OpenNebula Open Souce Solution for DC Virtualization Constantino Vázquez Blanco OpenNebula.org What is OpenNebula? Multi-tenancy, Elasticity and Automatic Provision
