Ralf Gerhards presented detailed plans to install and test LSF batch system on the current H1 PC-Farm and make it available to H1 users this summer.

Size: px
Start display at page:

Download "Ralf Gerhards presented detailed plans to install and test LSF batch system on the current H1 PC-Farm and make it available to H1 users this summer."

Transcription

1 Summary of L45 upgrade meeting Alan Campbell Today a small group discussed some ideas on the use of PC farms in H1 for batch processing and online filtering, in the light of a second PC-Farm for which we already have funding. Ralf Gerhards presented detailed plans to install and test LSF batch system on the current H1 PC-Farm and make it available to H1 users this summer. Alan Campbell presented the attached slides. He started to formulate changes in the standard H1 program framework to enable user jobs to use parallel processing. In addition some technolgies which may be usefull in its implementation were mentioned. It was felt that before starting coding we should invite all H1 colleagues ( and friends ) to give comments, ideas, and suggestions and invite participation. Remember: PC-Farm is the future of computing in HEP Overall Goal: Bring this future to H1 Requirements: Offline computing - datafile access, batch Online computing - high reliability, guaranteed performance

2 Alan Campbell A new commodity PC farm for H1 ( using Linux ) The purpose of this talk/document is to initiate discussion on project goals and ask H1 ( and others ) for ideas. Remember: PC-Farm is the future of computing in HEP Overall Goal: Bring this future to H1 Requirements: Offline computing - datafile access, batch Online ( L45 ) computing - high reliability, guaranteed performance "Call for ideas" "Request for Concepts" "BoF Birds of Feather Session" To start this discussion I will present some initial thoughts on the following: An extension to H1 application framework ( BOS/FSEQR/module steering ) to Network communication software and hardware topology

3 An extension to H1 application framework ( BOS/FSEQR/module steering ) to Recall: Current H1 framework BEGJOB open first input file open output file open ntuples book histograms printout BEGRUN access database printout FSEQR/W loop Data File In Data File Out OTHDAT REVENT CALL MODULE1 CALL MODULE2... printout open following input file open following output file ENDRUN printout ENDJOB close ntuple file print histograms printout

4 An extension to H1 application framework ( BOS/FSEQR/module steering ) to Recall: Current L5/PC Farm framework Same as "current H1 framework" except that main input/output is not data file but instead shared memory event pool. Simple to run "current H1 framework" as only event in/out is altered But: multiple database connections multiple printout files multiple histogram files Extra infrastructure required combine histogram files feed events to / from shared memory pool synchronise runs/ sort event numbers Single shared memory pool currently limits all input and output most go via a single machine.

5 An extension to H1 application framework ( BOS/FSEQR/module steering ) to Recall: Current L4 Farm framework FSEQR input stream is fed not only with events but also with database records, end-of-run marker records, runstart record subset, pre-event records. FSEQR output stream contains not only RUNEVENT records but also end-of-run marker records, ( synchronise at run change ) RUNDATA records ( -> TCL job ) SPECIAL stream records histogram records Dummy NDB routine...all database banks automatically present before event processing ALL modules are called BEGRUN true, REVENT false to book histograms Required modules are called with REVENT true for processing No multiple database connections No multiple histogram files Multiple parallel input streams Output stream independant of input Extra infrastructure required database reading ( mdbquefpack ) combine histograms and write to database ( TCL job ) output file writing and dumping ( datalog job )

6 An extension to H1 application framework ( BOS/FSEQR/module steering ) to Proposal for new metacomputer-enabled H1 framework Purpose: Allow all H1 jobs to efficiently use many computers ( including multi-processor computers ) Goal: minimal changes needed to current scheme so that both h1rec modules and user analysis programs can be easily modified Remove completely "specialness" of L4,L5,Reprocessing, H1 batch frameworks -> all jobs run in one scheme -> anyone can maintain -> developments benefit everyone Possible/proposed implementation: Single user application steered by logical flags as present ie BEGJOB, BEGRUN, REVENT... These flags are set by the new application framework according to the records on the input datastream. This application will contain all user code and may be run as single process as present -OR- on metacomputer. In this case the user process will run simultaneously on many computers, and a harness will provide for the data transfer between the processes. New flags are introduced to cause the user application to perform additional tasks ie DBBJOB DBBRUN HISJOB HISRUN TCLANA Fetch job dependant database banks Fetch run dependent database banks Book job dependant histograms Book run dependant histograms Analyse RUNDATA record

7 An extension to H1 application framework ( BOS/FSEQR/module steering ) to Hence the functionality of the current "special tasks" eg mdbquefpack, TCL job is implemented within the user application -> no need for these special tasks -> every user job has access to this functionality histogram collection ntuple collection printout collection are to be implemented within the user application with the aid of new h1 library code histogramming : The harness will supply data on the input stream which causes the "event" loop histogram calls to either fill local histograms and send on output stream at ENDRUN ( or ENDJOB ) -OR- to send on output stream data from histogram filling calls at end of each event processing. The harness will care for the transfer of these data to a copy of the user process so that this copy sums histograms -OR- makes actual histogram filling and is informed of end-run, end-job conditions allowing for writing of histogram files on run and job basis. ntuples: The harness will send on output stream n-tuple entry calls made during event processing to an incarnation of the user process so that ntuple an ntuple file may be written. printout: The new framework will collect all printout made at each stage of processing ie BEGJOB, REVENT event processing etc label it with run/event number and timestamp and send this data on the output stream...to an incarnation of the user process which will write a single printout file. database input: database input information ( eg TCL job ) may be written by the user process during event processing and the harness will feed all such records...to an incarnation of the user process which will have then flag TCLANA set instructing it to analyse this data.

8 An extension to H1 application framework ( BOS/FSEQR/module steering ) to The harness The current L4 system contains what we can consider a prototype harness diski - disk input neto - transfer datastream to processing computers via network, adding end-of-run markers neti - receive datastream from network nodi - transfer datastream to user processes on this computer, adding database records and event steering records ( ie random number seed, event duplicate ) nodo - collect output from user processes on this computer synchronising on end-of-run logo - output this stream to network receiver - collect the output streams from all computers logging - disk output, dump job submission, tcl job submission master - starts some of the tasks ( receiver, logo ) These tasks use PVM for data transfer ( using TCP ) and shared memory for communication with the user application. This harness runs on Lynx-OS/irix and apart from movement to linux has several missing features which will be needed by a general user job harness ie Resource allocation and process management : on which machines should the job be run ( current har coded in master and in configuration files for h1l4iox machines ) - automatic choice according to data file location and network capabilities and system loading Security : user authentication - wee don't want to have to install all user account on all machines Health and Status : monitoring of health and status of system PVM daemon on per user OR per system -> interference between different user jobs PVM master daemon runs on single system -> may become bottleneck Data transfer speed/efficiency : TCP/IP flow control may not be appropriate and has high overhead especially for just round the corner high speed commodity networks eg gigabit ethernet. PVM introduces further data buffering ( and hence copying ) reducing performance. Executable management : construction, caching and location of executables - fast distribution of tasks to all machines in user job Remote data access : fpack causes file staging - but no pre-staging, automatic temporary output file creation and migration on close. -

9 Tigran and I have started to look at what other people are doing and we hope YOU will inform us of systems from which we can learn. ( campbell@desy.de ). Here are some projects I think we can learn from ( or make use of ). Most are research projects and/or work in progress. How should we implement the new harness for PC-Linux? Data transfer software API ( application programmers interface ): PVM ( parallel virtual machine ) MPI ( Message Passing Interface ) may be better than PVM in performance and longer-term support. Various implementation are available for Linux which provide also some of the features missing in MPI cf PVM. Nexus library ( part of Globus computing grid project ) VIA ( virtual interface architecture ) specifically M-VIA for Linux and upcoming MPI for M-VIA. Provides lower latency than TCP even for fast ethernet and for faster networks much higher bandwidth. May be best approach for our system Sockets Shared memory : current harness uses shared memory for inter-machine communication...should we use this approach on Linux or move to eg PVM with UNIX domain sockets or M-VIA

10 How should we implement the new harness for PC-Linux? Resource allocation / Security /... Maybe we can make use of components of the Globus project - the GUSTO Globus Uniquitous Supercomputing Testbed forms a grid of 3600 processors located in 17 sites worldwide ( 2 Tflops/s ). LSF commercial local load sharing facility ( maybe too expensive - RAL ) MOSIX for Linux allows process migration, machine load balancing For L45 application alone a static configuration file approach ( as used by H1 Monte-Carlo or L4 setup on Lynx-OS ) may be sufficient initially. Condor project - load balancing, checkpointing, process migration Easy-LL ( IBM loadleveler ). but is it available for Linux?

11 How should we implement the hardware and network? The nicest PC-Farm setup I found is Sarnoff Research Centre's Cyclone cluster :8000/docs/metacomputing.html Please look at the pictures. The configuration is: 128 dual processor nodes ( 64MB, 3.2GB disk, CDROM, 2 SMC 100Mbit ethernet cards, NO floppy drives, NO graphics cards ) Each node is connected to 2 3COM3c port fast ethernet switches making a public ( for NFS.. ) and a private network ( for application trtaffic only ). The private network uplinks via fibre to a gigabit ethernet switch forming a "fat tree" network. 10 boot CDROMs enable complete reinstallation of all software in 80 minutes. FreeBSD is deployed in addition to Linux Faster than 100Mbit ethernet is currently still expensive eg myrinet, gigabit ethernet, pvic. For the link to North Hall we need more bandwidth than the single FDDI link currently available. Unfortuneately it looks like a gigabit ethernet uplink between bridges in North Hall and DESY computer centre is not on because of the large distance and the special cable which would have to be installed. Perhaps we should consider replacing FDDI by Optical Fast Ethernet between h1l4iox machines and a fast ethernet switch in computer centre. It has to be investigated and tested if the current FDDI cables can be used for this. Otherwise we may have to at least install a second FDDI ring North Hall - Computer centre and network hardware with both FDDI and Fast- or Gigabit ethernet ( also expensive ) -OR- we have to maintain current L4 system for event rejection and live with a single FDDI connection to computer centre ( but this partially destroys the mission of this project ) -OR- we have to find space for the whole installation in the North Hall and live with very limited bandwidth to computer centre -OR-?...we should discuss these matters with computer centre soon. Addendum: During the meeting Ralf reported that it is now clear that many fibres are available which between Hall North and DESY computer centre which are suitable for optical fast ethernet. The suggestion is to install CISCO fast ethernet switchs on both end of these fibres ( say 3 or 4 links between the switches ) and use the rest of the ports on the North Hall side to connect the h1l4iox machines on which we will replace the FDDI interfaces by fast ethernet.

12 How should we proceed now? I propose to announce and publish these slides on H1 WWW and hope for some comments/suggestions from our colleagues. I propose everyone who wishes to actively participate ( ie write code, test code, investigate commercial and public domain software, discuss, test networks, care for PC and/or network hardware, care for Linux installation ) should consider this now and inform me and Ralf Gerhards of the area on which they would like to work and how much time they will may have available. All people associated to H1 are welcome to participate. Ralf and I will then organise a meeting of these volunteers.

A Dataflow Meta-Computing Framework for Event Processing in the H1 experiment

A Dataflow Meta-Computing Framework for Event Processing in the H1 experiment A Dataflow Meta-Computing Framework for Event Processing in the H1 experiment Alan Campbell 1, Ralf Gerhards 1, Christoph Grab 2, Janusz Martyniak 3, Tigran Mkrtchyan 1, Sergey Levonian 1, Jacek Nowak

More information

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago Globus Striped GridFTP Framework and Server Raj Kettimuthu, ANL and U. Chicago Outline Introduction Features Motivation Architecture Globus XIO Experimental Results 3 August 2005 The Ohio State University

More information

Scalability and Classifications

Scalability and Classifications Scalability and Classifications 1 Types of Parallel Computers MIMD and SIMD classifications shared and distributed memory multicomputers distributed shared memory computers 2 Network Topologies static

More information

Building a Linux Cluster

Building a Linux Cluster Building a Linux Cluster CUG Conference May 21-25, 2001 by Cary Whitney Clwhitney@lbl.gov Outline What is PDSF and a little about its history. Growth problems and solutions. Storage Network Hardware Administration

More information

Simplest Scalable Architecture

Simplest Scalable Architecture Simplest Scalable Architecture NOW Network Of Workstations Many types of Clusters (form HP s Dr. Bruce J. Walker) High Performance Clusters Beowulf; 1000 nodes; parallel programs; MPI Load-leveling Clusters

More information

MOSIX: High performance Linux farm

MOSIX: High performance Linux farm MOSIX: High performance Linux farm Paolo Mastroserio [mastroserio@na.infn.it] Francesco Maria Taurino [taurino@na.infn.it] Gennaro Tortone [tortone@na.infn.it] Napoli Index overview on Linux farm farm

More information

E-Business Technologies

E-Business Technologies E-Business Technologies Craig Van Slyke and France Bélanger John Wiley & Sons, Inc. Slides by Fred Niederman 7-1 Client/Server Technologies for E-Business Chapter 7 7-2 Key Ideas E-commerce applications

More information

Scheduling and Resource Management in Computational Mini-Grids

Scheduling and Resource Management in Computational Mini-Grids Scheduling and Resource Management in Computational Mini-Grids July 1, 2002 Project Description The concept of grid computing is becoming a more and more important one in the high performance computing

More information

Hadoop Architecture. Part 1

Hadoop Architecture. Part 1 Hadoop Architecture Part 1 Node, Rack and Cluster: A node is simply a computer, typically non-enterprise, commodity hardware for nodes that contain data. Consider we have Node 1.Then we can add more nodes,

More information

- An Essential Building Block for Stable and Reliable Compute Clusters

- An Essential Building Block for Stable and Reliable Compute Clusters Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative

More information

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN 1 PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN Introduction What is cluster computing? Classification of Cluster Computing Technologies: Beowulf cluster Construction

More information

SAN Conceptual and Design Basics

SAN Conceptual and Design Basics TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer

More information

A Comparison on Current Distributed File Systems for Beowulf Clusters

A Comparison on Current Distributed File Systems for Beowulf Clusters A Comparison on Current Distributed File Systems for Beowulf Clusters Rafael Bohrer Ávila 1 Philippe Olivier Alexandre Navaux 2 Yves Denneulin 3 Abstract This paper presents a comparison on current file

More information

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

CMS Tier-3 cluster at NISER. Dr. Tania Moulik CMS Tier-3 cluster at NISER Dr. Tania Moulik What and why? Grid computing is a term referring to the combination of computer resources from multiple administrative domains to reach common goal. Grids tend

More information

Chapter 1 - Web Server Management and Cluster Topology

Chapter 1 - Web Server Management and Cluster Topology Objectives At the end of this chapter, participants will be able to understand: Web server management options provided by Network Deployment Clustered Application Servers Cluster creation and management

More information

A Flexible Cluster Infrastructure for Systems Research and Software Development

A Flexible Cluster Infrastructure for Systems Research and Software Development Award Number: CNS-551555 Title: CRI: Acquisition of an InfiniBand Cluster with SMP Nodes Institution: Florida State University PIs: Xin Yuan, Robert van Engelen, Kartik Gopalan A Flexible Cluster Infrastructure

More information

Grid on Blades. Basil Smith 7/2/2005. 2003 IBM Corporation

Grid on Blades. Basil Smith 7/2/2005. 2003 IBM Corporation Grid on Blades Basil Smith 7/2/2005 2003 IBM Corporation What is the problem? Inefficient utilization of resources (MIPS, Memory, Storage, Bandwidth) Fundamentally resources are being wasted due to wide

More information

Microsoft Windows Compute Cluster Server 2003 Getting Started Guide

Microsoft Windows Compute Cluster Server 2003 Getting Started Guide Microsoft Windows Compute Cluster Server 2003 Getting Started Guide Part Number 434709-003 March 2007 (Third Edition) Copyright 2006, 2007 Hewlett-Packard Development Company, L.P. The information contained

More information

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010 Best Practices for Data Sharing in a Grid Distributed SAS Environment Updated July 2010 B E S T P R A C T I C E D O C U M E N T Table of Contents 1 Abstract... 2 1.1 Storage performance is critical...

More information

Recommended hardware system configurations for ANSYS users

Recommended hardware system configurations for ANSYS users Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range

More information

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one

More information

RevoScaleR Speed and Scalability

RevoScaleR Speed and Scalability EXECUTIVE WHITE PAPER RevoScaleR Speed and Scalability By Lee Edlefsen Ph.D., Chief Scientist, Revolution Analytics Abstract RevoScaleR, the Big Data predictive analytics library included with Revolution

More information

Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang

Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang Distributed RAID Architectures for Cluster I/O Computing Kai Hwang Internet and Cluster Computing Lab. University of Southern California 1 Presentation Outline : Scalable Cluster I/O The RAID-x Architecture

More information

Network Attached Storage. Jinfeng Yang Oct/19/2015

Network Attached Storage. Jinfeng Yang Oct/19/2015 Network Attached Storage Jinfeng Yang Oct/19/2015 Outline Part A 1. What is the Network Attached Storage (NAS)? 2. What are the applications of NAS? 3. The benefits of NAS. 4. NAS s performance (Reliability

More information

Cisco Application Networking Manager Version 2.0

Cisco Application Networking Manager Version 2.0 Cisco Application Networking Manager Version 2.0 Cisco Application Networking Manager (ANM) software enables centralized configuration, operations, and monitoring of Cisco data center networking equipment

More information

Communicating with devices

Communicating with devices Introduction to I/O Where does the data for our CPU and memory come from or go to? Computers communicate with the outside world via I/O devices. Input devices supply computers with data to operate on.

More information

How To Monitor And Test An Ethernet Network On A Computer Or Network Card

How To Monitor And Test An Ethernet Network On A Computer Or Network Card 3. MONITORING AND TESTING THE ETHERNET NETWORK 3.1 Introduction The following parameters are covered by the Ethernet performance metrics: Latency (delay) the amount of time required for a frame to travel

More information

Distribution One Server Requirements

Distribution One Server Requirements Distribution One Server Requirements Introduction Welcome to the Hardware Configuration Guide. The goal of this guide is to provide a practical approach to sizing your Distribution One application and

More information

Implementing Network Attached Storage. Ken Fallon Bill Bullers Impactdata

Implementing Network Attached Storage. Ken Fallon Bill Bullers Impactdata Implementing Network Attached Storage Ken Fallon Bill Bullers Impactdata Abstract The Network Peripheral Adapter (NPA) is an intelligent controller and optimized file server that enables network-attached

More information

Building a Highly Available and Scalable Web Farm

Building a Highly Available and Scalable Web Farm Page 1 of 10 MSDN Home > MSDN Library > Deployment Rate this page: 10 users 4.9 out of 5 Building a Highly Available and Scalable Web Farm Duwamish Online Paul Johns and Aaron Ching Microsoft Developer

More information

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components

Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components Welcome to the unit of Hadoop Fundamentals on Hadoop architecture. I will begin with a terminology review and then cover the major components of Hadoop. We will see what types of nodes can exist in a Hadoop

More information

Deploying a distributed data storage system on the UK National Grid Service using federated SRB

Deploying a distributed data storage system on the UK National Grid Service using federated SRB Deploying a distributed data storage system on the UK National Grid Service using federated SRB Manandhar A.S., Kleese K., Berrisford P., Brown G.D. CCLRC e-science Center Abstract As Grid enabled applications

More information

Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking

Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking Burjiz Soorty School of Computing and Mathematical Sciences Auckland University of Technology Auckland, New Zealand

More information

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration Solutions Integrated Storage Appliances Management Optimized Storage & Migration Archive Data Retention & Compliance Services Global Installation & Support SECURING THE FUTURE OF YOUR DATA w w w.q sta

More information

Grid Scheduling Dictionary of Terms and Keywords

Grid Scheduling Dictionary of Terms and Keywords Grid Scheduling Dictionary Working Group M. Roehrig, Sandia National Laboratories W. Ziegler, Fraunhofer-Institute for Algorithms and Scientific Computing Document: Category: Informational June 2002 Status

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

An Oracle White Paper July 2011. Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide

An Oracle White Paper July 2011. Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide An Oracle White Paper July 2011 1 Disclaimer The following is intended to outline our general product direction.

More information

SYSTEM SETUP FOR SPE PLATFORMS

SYSTEM SETUP FOR SPE PLATFORMS BEST PRACTICE SYSTEM SETUP FOR SPE PLATFORMS Product Snow License Manager Version 7.0 Content System requirements SQL Server configuration Maintenance Test environment Document date 2015-10-15 ABOUT THIS

More information

EMC Invista: The Easy to Use Storage Manager

EMC Invista: The Easy to Use Storage Manager EMC s Invista SAN Virtualization System Tested Feb. 2006 Page 1 of 13 EMC Invista: The Easy to Use Storage Manager Invista delivers centrally managed LUN Virtualization, Data Mobility, and Copy Services

More information

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This

More information

White Paper. Intrusion Detection Deploying the Shomiti Century Tap

White Paper. Intrusion Detection Deploying the Shomiti Century Tap White Paper Intrusion Detection Deploying the Shomiti Century Tap . Shomiti Tap Deployment Purpose of this Paper The scalability of Intrusion Detection Systems (IDS) is often an issue when deploying an

More information

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck

Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Sockets vs. RDMA Interface over 1-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji Hemal V. Shah D. K. Panda Network Based Computing Lab Computer Science and Engineering

More information

IBM Systems Director Navigator for i5/os New Web console for i5, Fast, Easy, Ready

IBM Systems Director Navigator for i5/os New Web console for i5, Fast, Easy, Ready Agenda Key: Session Number: 35CA 540195 IBM Systems Director Navigator for i5/os New Web console for i5, Fast, Easy, Ready 8 Copyright IBM Corporation, 2008. All Rights Reserved. This publication may refer

More information

Multi-Channel Clustered Web Application Servers

Multi-Channel Clustered Web Application Servers THE AMERICAN UNIVERSITY IN CAIRO SCHOOL OF SCIENCES AND ENGINEERING Multi-Channel Clustered Web Application Servers A Masters Thesis Department of Computer Science and Engineering Status Report Seminar

More information

Module I-7410 Advanced Linux FS-11 Part1: Virtualization with KVM

Module I-7410 Advanced Linux FS-11 Part1: Virtualization with KVM Bern University of Applied Sciences Engineering and Information Technology Module I-7410 Advanced Linux FS-11 Part1: Virtualization with KVM By Franz Meyer Version 1.0 February 2011 Virtualization Architecture

More information

Virtualization Infrastructure at Karlsruhe

Virtualization Infrastructure at Karlsruhe Virtualization Infrastructure at Karlsruhe HEPiX Fall 2007 Volker Buege 1),2), Ariel Garcia 1), Marcus Hardt 1), Fabian Kulla 1),Marcel Kunze 1), Oliver Oberst 1),2), Günter Quast 2), Christophe Saout

More information

How To Design A Data Centre

How To Design A Data Centre DATA CENTRE TECHNOLOGIES & SERVICES RE-Solution Data Ltd Reach Recruit Resolve Refine 170 Greenford Road Harrow Middlesex HA1 3QX T +44 (0) 8450 031323 EXECUTIVE SUMMARY The purpose of a data centre is

More information

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking Cluster Grid Interconects Tony Kay Chief Architect Enterprise Grid and Networking Agenda Cluster Grid Interconnects The Upstart - Infiniband The Empire Strikes Back - Myricom Return of the King 10G Gigabit

More information

M6310 USB Flash Drive Tester/Duplicator

M6310 USB Flash Drive Tester/Duplicator NEW... USB Production Tester/ Duplicator from the Industry Leader Simple to Use Fast Reliable Versatile Features!!! High Quality Production USB Tester/Duplicator for USB Sticks USB Hard Drives MP3 Players

More information

How To Connect To Bloomerg.Com With A Network Card From A Powerline To A Powerpoint Terminal On A Microsoft Powerbook (Powerline) On A Blackberry Or Ipnet (Powerbook) On An Ipnet Box On

How To Connect To Bloomerg.Com With A Network Card From A Powerline To A Powerpoint Terminal On A Microsoft Powerbook (Powerline) On A Blackberry Or Ipnet (Powerbook) On An Ipnet Box On Transport and Security Specification 15 July 2015 Version: 5.9 Contents Overview 3 Standard network requirements 3 Source and Destination Ports 3 Configuring the Connection Wizard 4 Private Bloomberg Network

More information

The GRID and the Linux Farm at the RCF

The GRID and the Linux Farm at the RCF The GRID and the Linux Farm at the RCF A. Chan, R. Hogue, C. Hollowell, O. Rind, J. Smith, T. Throwe, T. Wlodek, D. Yu Brookhaven National Laboratory, NY 11973, USA The emergence of the GRID architecture

More information

Network Station - Thin Client Computing - Overview

Network Station - Thin Client Computing - Overview Network Station - Thin Client Computing - Overview Overview The objective of this document is to help develop an understanding of a Server Based Computing/Thin-Client environment using MS Windows NT 4.0,

More information

- Behind The Cloud -

- Behind The Cloud - - Behind The Cloud - Infrastructure and Technologies used for Cloud Computing Alexander Huemer, 0025380 Johann Taferl, 0320039 Florian Landolt, 0420673 Seminar aus Informatik, University of Salzburg Overview

More information

DB2 Connect for NT and the Microsoft Windows NT Load Balancing Service

DB2 Connect for NT and the Microsoft Windows NT Load Balancing Service DB2 Connect for NT and the Microsoft Windows NT Load Balancing Service Achieving Scalability and High Availability Abstract DB2 Connect Enterprise Edition for Windows NT provides fast and robust connectivity

More information

Gigabit Ethernet Design

Gigabit Ethernet Design Gigabit Ethernet Design Laura Jeanne Knapp Network Consultant 1-919-254-8801 laura@lauraknapp.com www.lauraknapp.com Tom Hadley Network Consultant 1-919-301-3052 tmhadley@us.ibm.com HSEdes_ 010 ed and

More information

Client/Server Computing Distributed Processing, Client/Server, and Clusters

Client/Server Computing Distributed Processing, Client/Server, and Clusters Client/Server Computing Distributed Processing, Client/Server, and Clusters Chapter 13 Client machines are generally single-user PCs or workstations that provide a highly userfriendly interface to the

More information

Open Text Archive Server and Microsoft Windows Azure Storage

Open Text Archive Server and Microsoft Windows Azure Storage Open Text Archive Server and Microsoft Windows Azure Storage Whitepaper Open Text December 23nd, 2009 2 Microsoft W indows Azure Platform W hite Paper Contents Executive Summary / Introduction... 4 Overview...

More information

Shared Parallel File System

Shared Parallel File System Shared Parallel File System Fangbin Liu fliu@science.uva.nl System and Network Engineering University of Amsterdam Shared Parallel File System Introduction of the project The PVFS2 parallel file system

More information

Client/Server and Distributed Computing

Client/Server and Distributed Computing Adapted from:operating Systems: Internals and Design Principles, 6/E William Stallings CS571 Fall 2010 Client/Server and Distributed Computing Dave Bremer Otago Polytechnic, N.Z. 2008, Prentice Hall Traditional

More information

Data Management. Network transfers

Data Management. Network transfers Data Management Network transfers Network data transfers Not everyone needs to transfer large amounts of data on and off a HPC service Sometimes data is created and consumed on the same service. If you

More information

Owner of the content within this article is www.msexchange.org Written by Marc Grote www.it-training-grote.de

Owner of the content within this article is www.msexchange.org Written by Marc Grote www.it-training-grote.de Owner of the content within this article is www.msexchange.org Written by Marc Grote www.it-training-grote.de Exchange 2003 Implementing iscsi for Exchange Server 2003 Written by Marc Grote - mailto:grotem@it-training-grote.de

More information

Solution for private cloud computing

Solution for private cloud computing The CC1 system Solution for private cloud computing 1 Outline What is CC1? Features Technical details Use cases By scientist By HEP experiment System requirements and installation How to get it? 2 What

More information

Terms of Reference Microsoft Exchange and Domain Controller/ AD implementation

Terms of Reference Microsoft Exchange and Domain Controller/ AD implementation Terms of Reference Microsoft Exchange and Domain Controller/ AD implementation Overview Maldivian Red Crescent will implement it s first Microsoft Exchange server and replace it s current Domain Controller

More information

IBM LoadLeveler for Linux delivers job scheduling for IBM pseries and IBM xseries platforms running Linux

IBM LoadLeveler for Linux delivers job scheduling for IBM pseries and IBM xseries platforms running Linux Software Announcement May 11, 2004 IBM LoadLeveler for Linux delivers job scheduling for IBM pseries and IBM xseries platforms running Linux Overview LoadLeveler for Linux is a versatile workload management

More information

Using Linux Clusters as VoD Servers

Using Linux Clusters as VoD Servers HAC LUCE Using Linux Clusters as VoD Servers Víctor M. Guĺıas Fernández gulias@lfcia.org Computer Science Department University of A Corunha funded by: Outline Background: The Borg Cluster Video on Demand.

More information

Chapter 11 I/O Management and Disk Scheduling

Chapter 11 I/O Management and Disk Scheduling Operating Systems: Internals and Design Principles, 6/E William Stallings Chapter 11 I/O Management and Disk Scheduling Dave Bremer Otago Polytechnic, NZ 2008, Prentice Hall I/O Devices Roadmap Organization

More information

Analysis and Implementation of Cluster Computing Using Linux Operating System

Analysis and Implementation of Cluster Computing Using Linux Operating System IOSR Journal of Computer Engineering (IOSRJCE) ISSN: 2278-0661 Volume 2, Issue 3 (July-Aug. 2012), PP 06-11 Analysis and Implementation of Cluster Computing Using Linux Operating System Zinnia Sultana

More information

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a

More information

Configuration Maximums VMware vsphere 4.0

Configuration Maximums VMware vsphere 4.0 Topic Configuration s VMware vsphere 4.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.0. The limits presented in the

More information

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller White Paper From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller The focus of this paper is on the emergence of the converged network interface controller

More information

VIA CONNECT PRO Deployment Guide

VIA CONNECT PRO Deployment Guide VIA CONNECT PRO Deployment Guide www.true-collaboration.com Infinite Ways to Collaborate CONTENTS Introduction... 3 User Experience... 3 Pre-Deployment Planning... 3 Connectivity... 3 Network Addressing...

More information

LinuxWorld Conference & Expo Server Farms and XML Web Services

LinuxWorld Conference & Expo Server Farms and XML Web Services LinuxWorld Conference & Expo Server Farms and XML Web Services Jorgen Thelin, CapeConnect Chief Architect PJ Murray, Product Manager Cape Clear Software Objectives What aspects must a developer be aware

More information

PATROL Console Server and RTserver Getting Started

PATROL Console Server and RTserver Getting Started PATROL Console Server and RTserver Getting Started Supporting PATROL Console Server 7.5.00 RTserver 6.6.00 February 14, 2005 Contacting BMC Software You can access the BMC Software website at http://www.bmc.com.

More information

Cluster, Grid, Cloud Concepts

Cluster, Grid, Cloud Concepts Cluster, Grid, Cloud Concepts Kalaiselvan.K Contents Section 1: Cluster Section 2: Grid Section 3: Cloud Cluster An Overview Need for a Cluster Cluster categorizations A computer cluster is a group of

More information

Building Clusters for Gromacs and other HPC applications

Building Clusters for Gromacs and other HPC applications Building Clusters for Gromacs and other HPC applications Erik Lindahl lindahl@cbr.su.se CBR Outline: Clusters Clusters vs. small networks of machines Why do YOU need a cluster? Computer hardware Network

More information

SIDN Server Measurements

SIDN Server Measurements SIDN Server Measurements Yuri Schaeffer 1, NLnet Labs NLnet Labs document 2010-003 July 19, 2010 1 Introduction For future capacity planning SIDN would like to have an insight on the required resources

More information

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver 1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution

More information

Overlapping Data Transfer With Application Execution on Clusters

Overlapping Data Transfer With Application Execution on Clusters Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm reid@cs.toronto.edu stumm@eecg.toronto.edu Department of Computer Science Department of Electrical and Computer

More information

Rcore Embedded Software Platform

Rcore Embedded Software Platform Industrial Computing Rcore Embedded Software Platform The high reliability and stability of Moxa s embedded computers give our customers an edge over the competition. Moxa s Rcore embedded software platform

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described

More information

CSE-E5430 Scalable Cloud Computing Lecture 2

CSE-E5430 Scalable Cloud Computing Lecture 2 CSE-E5430 Scalable Cloud Computing Lecture 2 Keijo Heljanko Department of Computer Science School of Science Aalto University keijo.heljanko@aalto.fi 14.9-2015 1/36 Google MapReduce A scalable batch processing

More information

High Performance Cluster Support for NLB on Window

High Performance Cluster Support for NLB on Window High Performance Cluster Support for NLB on Window [1]Arvind Rathi, [2] Kirti, [3] Neelam [1]M.Tech Student, Department of CSE, GITM, Gurgaon Haryana (India) arvindrathi88@gmail.com [2]Asst. Professor,

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

Selling Compellent NAS: File & Block Level in the Same System Chad Thibodeau

Selling Compellent NAS: File & Block Level in the Same System Chad Thibodeau Selling Compellent NAS: File & Block Level in the Same System Chad Thibodeau Agenda Session Objectives Feature Overview Technology Overview Compellent Differentiators Competition Available Resources Questions

More information

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment

The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment The Advantages of Multi-Port Network Adapters in an SWsoft Virtual Environment Introduction... 2 Virtualization addresses key challenges facing IT today... 2 Introducing Virtuozzo... 2 A virtualized environment

More information

IMPLEMENTING GREEN IT

IMPLEMENTING GREEN IT Saint Petersburg State University of Information Technologies, Mechanics and Optics Department of Telecommunication Systems IMPLEMENTING GREEN IT APPROACH FOR TRANSFERRING BIG DATA OVER PARALLEL DATA LINK

More information

IP SAN Fundamentals: An Introduction to IP SANs and iscsi

IP SAN Fundamentals: An Introduction to IP SANs and iscsi IP SAN Fundamentals: An Introduction to IP SANs and iscsi Updated April 2007 Sun Microsystems, Inc. 2007 Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, CA 95054 USA All rights reserved. This

More information

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) ( TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx

More information

Configuration Maximums

Configuration Maximums Topic Configuration s VMware vsphere 5.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 5.1. The limits presented in the

More information

VMWARE WHITE PAPER 1

VMWARE WHITE PAPER 1 1 VMWARE WHITE PAPER Introduction This paper outlines the considerations that affect network throughput. The paper examines the applications deployed on top of a virtual infrastructure and discusses the

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...

More information

Lecture 2 Parallel Programming Platforms

Lecture 2 Parallel Programming Platforms Lecture 2 Parallel Programming Platforms Flynn s Taxonomy In 1966, Michael Flynn classified systems according to numbers of instruction streams and the number of data stream. Data stream Single Multiple

More information

Dell PowerEdge Blades Outperform Cisco UCS in East-West Network Performance

Dell PowerEdge Blades Outperform Cisco UCS in East-West Network Performance Dell PowerEdge Blades Outperform Cisco UCS in East-West Network Performance This white paper compares the performance of blade-to-blade network traffic between two enterprise blade solutions: the Dell

More information

Guideline for stresstest Page 1 of 6. Stress test

Guideline for stresstest Page 1 of 6. Stress test Guideline for stresstest Page 1 of 6 Stress test Objective: Show unacceptable problems with high parallel load. Crash, wrong processing, slow processing. Test Procedure: Run test cases with maximum number

More information

Plug-and-play Virtual Appliance Clusters Running Hadoop. Dr. Renato Figueiredo ACIS Lab - University of Florida

Plug-and-play Virtual Appliance Clusters Running Hadoop. Dr. Renato Figueiredo ACIS Lab - University of Florida Plug-and-play Virtual Appliance Clusters Running Hadoop Dr. Renato Figueiredo ACIS Lab - University of Florida Advanced Computing and Information Systems laboratory Introduction You have so far learned

More information

Parallel Computing. Benson Muite. benson.muite@ut.ee http://math.ut.ee/ benson. https://courses.cs.ut.ee/2014/paralleel/fall/main/homepage

Parallel Computing. Benson Muite. benson.muite@ut.ee http://math.ut.ee/ benson. https://courses.cs.ut.ee/2014/paralleel/fall/main/homepage Parallel Computing Benson Muite benson.muite@ut.ee http://math.ut.ee/ benson https://courses.cs.ut.ee/2014/paralleel/fall/main/homepage 3 November 2014 Hadoop, Review Hadoop Hadoop History Hadoop Framework

More information

Distributed Operating Systems. Cluster Systems

Distributed Operating Systems. Cluster Systems Distributed Operating Systems Cluster Systems Ewa Niewiadomska-Szynkiewicz ens@ia.pw.edu.pl Institute of Control and Computation Engineering Warsaw University of Technology E&IT Department, WUT 1 1. Cluster

More information

MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM?

MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? Ashutosh Shinde Performance Architect ashutosh_shinde@hotmail.com Validating if the workload generated by the load generating tools is applied

More information

Monitoring high-speed networks using ntop. Luca Deri <deri@ntop.org>

Monitoring high-speed networks using ntop. Luca Deri <deri@ntop.org> Monitoring high-speed networks using ntop Luca Deri 1 Project History Started in 1997 as monitoring application for the Univ. of Pisa 1998: First public release v 0.4 (GPL2) 1999-2002:

More information

RCL: Software Prototype

RCL: Software Prototype Business Continuity as a Service ICT FP7-609828 RCL: Software Prototype D3.2.1 June 2014 Document Information Scheduled delivery 30.06.2014 Actual delivery 30.06.2014 Version 1.0 Responsible Partner IBM

More information