Using Reservations to Implement Fixed Duration Node Allotment with PBS Professional
|
|
- Horatio George
- 8 years ago
- Views:
Transcription
1 Using Reservations to Implement Fixed Duration Node Allotment with PBS Professional Brajesh Pande Senior Computer Engineer Computer Centre IIT Kanpur Kanpur, UP India Manoj Soni Technical Consultant Altair India Chintels Techo Park, Kailash Colony New Delhi , India Dario Dorella PBS Works Technical Specialist Altair Engineering c/o Environment Park,Via Livorno Torino,Italy Abbreviations: HPC High Performance Computing, CPU Central Processing Unit, ILO Integrated lights out, GPU Graphical Processing Unit, VLAN Virtual Local Area Network, GbPS Giga bits per second, GBPS Giga bytes per second, CMU Cluster Management Utility, HP Hewlett Packard, SAN storage area network Keywords: High Performance Computing, Schedulers, PBS Professional, reservations, resource allocation, node allotment Abstract Indian Institute of Technology Kanpur has a High Performance Computing Facility that was in the top 500 list [1] of June PBS professional, a part of PBS-Works [2] solution from Altair, finds interesting use in implementing a sequential queue on this facility. A user who submits a job to this queue is given the non-interactive use of an 8-core-compute-node for a fixed period of time. The envisioned use of this allotment is to allow the user to run any number of sequential jobs on the given node with the condition that the latest ending time of all such jobs be the same. It is expected that this queue be used only for sequential jobs yet clever parallel use of the queue are not ruled out. The desired functionality is achieved by creating a fixed duration reservation for the user after the first job is submitted to the queue. The first and subsequent jobs of the same user on this queue are run on the same reservation. We detail this implementation and touch upon some generalizations of this concept. We also briefly describe the facility, its queuing structure and other provisions that leverage PBS Professional functionality to ensure the smooth functioning of the facility Introduction A High Performance Computing (HPC) machine typically consists of several compute nodes connected together through some interconnect. The compute nodes could be multiple CPU / Core machines or even GPU based machines while interconnects could be the memory paths or the network with varying speeds or a combination of both. Multiple CPUs / CORE machines use the memory paths for intra-node communication and the network for inter-node communication. Additional hardware and various complex tools become evermore necessary as the number of nodes in the facility increase. Thus these machines also have service processors for ILO management, System Deployment Tools, Tools for Monitoring and Management of Nodes, some Head nodes for access and utility services apart from the necessary CPUs, software and operating systems. These machines also have a scheduler, the focus of this paper that manages job distribution on the nodes. Schedulers view the machine as a collection of resources. A user uses the services of a scheduler to request for resources satisfying some desired criteria such as the number of compute nodes required, the amount of memory on the nodes, the duration for which the nodes are to be given and the licenses for software that are needed. Resources can not only be mapped to physical objects but can also be virtualized and defined during the configuration of the scheduler and can also be dynamically discovered. If the requested resources are available, and the policy configured on the scheduler permits allocation, the available resources are released to the user. The scheduler restricts the usage of resources by configured policy dictates. Schedulers maintain job queues with attributes that can be linked to the resources. The job queues and the linkage between the resources and the queues are defined during the configuration of the queues. Jobs that are submitted by the user (resource requests) to a queue are either finally scheduled on some resources by the scheduler based on the tunable internal algorithms that take the policies into account or else wait in the queue. 1
2 Since configuring the policies, the queues, and the resources on a scheduler can be a complex task, not all desired functionality from the scheduler can be accomplished by simple specifications. Good schedulers recognize this and provide hooks and interfaces that allow the scheduler to meet, with varying degrees of satisfaction, the user requirements after some programming and effort. Such functionality is driven by user needs. If such needs are general enough they can be over a period of time be incorporated within the framework of specifications of the scheduler. This paper focuses on a functionality desired for the HPC facility at IIT Kanpur that uses PBS Professional as its scheduler. It describes the HPC facility in brief, talks about its queue structures and the regular use of PBS Professional in the facility. It then goes on to enunciate the need for the more complex functionality. This functionality implements what the authors have named a sequential queue. A user who submits a job to this queue is given the non-interactive use of an 8-core-compute-node for a fixed period of time. The envisioned use of this allotment is to allow the user to run any number of sequential jobs on the given node with the condition that the latest ending time of all such jobs be the same. This paper also provides a solution for the implementation of this functionality through the use of interfaces provides by PBS Professional. Main HPC Facility at IIT Kanpur Figure 1, shows the architecture of the HPC facility at IIT Kanpur [3]. This facility has 368 compute nodes and 4 management nodes and 4 utility nodes. The blade type compute nodes are 2 Nehalem Intel-CPUs with 4 cores each on a BL-280c-G6 server and 48 GB of RAM. Each node has a 146GB small form factor Serial AT Attachment (SATA) disk for holding the boot images, operating systems and some necessary software. These nodes are connected together by several networks. The first network that connects these nodes is a boot network that is primarily used for booting the systems, and managerial functionality. The second network connecting them is a managerial network that connects ILO system and the console of various machines. These networks are represented by a single VLAN shown in the figure as Ethernet & ILO. All the nodes are also connected by a QDR-Infiniband network from QLogic that has a speed of 40 GbPS. This is a very fast interconnect and separation of management concerns from user I/O concerns makes for a fast through put of the users of the HPC system. Three file systems that use Modular Smart Array of disks (MSA) P2000 G3 FC from HP are also present in the entire facility. These disks use the Fibre-Channel Network to connect to SAN switches that finally server data for the file systems. The first file system of 7.4 TB is a NFS file system that is used for administrative purpose such as holding boot images and for miscellaneous storage purposes. The second file system is a Lustre based file system that has a high read and write throughput (around 3.4 GBPS read and 2.4 GBPS write throughput). This file system is called the scratch file system and has 40 TB of capacity. There is a slower home file system of 60 TB again on Lustre (1.7 GBPS read and 1.3 GBPS write throughput). Two different Lustre file systems separate out user concerns between unlimited scratch-space during compute operations versus quota based slower home file systems. 2
3 Figure1: The Architecture of HPC Facility at IIT Kanpur The operating system on the machines is RHEL5.4. The cluster uses CMU tool from HP for management and monitoring of the systems and deployment of operating system and software on the nodes. It uses PBS Professional as the scheduler for jobs submitted by the user. PBS Analytics and Compute Manager have also been deployed on the cluster The machine is mostly used by students, faculty and researchers of IIT Kanpur. Some users from other academic institutions also use the facility. Typically users come from almost all departments of the institute and use the facility for computational purposes through home grown software, libraries and compilers as well as available third party software that include but are not limited to: Gaussian, TurboMole, Ansys, Matlab, Accelrys, CPMD, WRF, Metis, MedeaVasp, Laamps, Tinker, NAG parallel libraries and compilers, GAMS, gcc, INTEL suite of compilers, pgi compiler from PORTLAND, many MPI libraries variations of mvapich2 available from QLogic, openmp, and flavors of mpich2. This machine has a peak theoretical performance of TFLOP and a maximum performance of TFLOP and was ranked 369 in the Top 500 list of June A specific usage scenario of the machine is shown in Figure 2 that was prepared by pulling out data with the help of PBS-Analytics. It shows that some reservations apart from the main queues are also in use and that the average wait time for jobs is increasing indicating a need for more compute nodes. Use of reservations in this facility is discussed more later. 3
4 Figure: 2 Typical usage of the Queues of the HPC Facility Use of PBS Professional in the HPC Facility: Currently the system uses PBS Professional version 11.2 which was upgraded from an earlier 10.4 release. A PBS failover server is also deployed for high availability. The system is divided into several queues as shown in the Table 1. These will be discussed in brief till the focus shifts to the main queue under discussion that is the sequential queue labeled seq in the table. 4
5 TABLE I QUEUES OF THE HPC FACILITY AT IIT KANPUR Queue Total Available Min Max Nodes Min Max Cores Allowed CPU / Nodes Allowed Allowed Wall Time large Max / 72 hrs medium Max / 96 hrs small Max/ 120 hrs workq hrs/48 hrs test 3 any any Max / Max seq 16 1 any Max / 120 hrs Table 1 shows the queues, total nodes available in each queue, minimum and maximum nodes allowed to a for a job, minimum and maximum cores allowed for a job and the maximum CPU and wall time allowed on the node. The idea behind these restrictions stems from the fact that given a reasonably sized problem, most HPC problems do not scale up beyond a certain number of nodes. Restrictions on wall time are necessary for giving fair access of the queues to users. There are no special privileges for any user of the cluster currently. Provision for allowing jobs with more number of cores or wall-time is possible but is reserved with the policy making body for HPC usage. The large queue allows for more number of cores but lesser wall time, the medium allows lesser cores but more wall time compared to the large queue and the small queue allows even lesser cores but more wall time than the medium queue. This arrangement is well suited for scaling studies. Work queue (workq) is a special queue that is allows both interactive as well as batch jobs. The idea behind this queue is that users would land on this queue for day to day activities and would be placed in a round robin fashion on the nodes of the work queue. This load balances the interactive usage of the queue. This is achieved by defining a static consumable resource on each node that is used when the node is in use. Allocation of the node of the work queue is done after sorting on the available numbers of this resource on a node. The round robin allocation of the batch processing in the work queue is a bit trickier and is implemented by a runtime hook that also makes the work queue an interactive queue. The implementation of the work queue more or less follows standard implementation examples in the PBS Professional manual with the necessary programming for the round robin allocation. The total number of round robin resources associated with the work queue decide the amount to which this queue can be loaded. The test queue is reserved for testing of new software and implementations. The sequential queue (seq) is the main focus of this paper and is discussed in more detail in the next section. The Sequential Queue requirements The need for the sequential queue arises from the fact that though HPC machines are mostly used for parallel applications sequential needs of users cannot be ruled out especially if the users belong to diverse computing environments. There may be legacy code as well as some code that is inherently sequential. Thus provision has to be made for such a queue. An obvious solution to this is to assign a core to any user who wants to use the sequential queue. The problem with this simplistic allocation policy is that the assigned core belongs to a machine that has memory which is shared by the other cores. If the job that has been allotted the single core uses large amounts of memory it would ultimately unfairly use the memory resources allowed to a job on another core on the same node. This is clearly an undesirable situation and would ultimately slow down all jobs that contend for memory on this node. In nodes that are severely affected almost all the memory and even the swap space would be used up and this would result in hanging the node. A user using memory within limits would also be penalized with non performance in this scenario. One of the ways out of this dilemma would be to restrict the memory use of a single job on the sequential queue. If we do a calculation with 48 GB of memory amongst 8 cores and also take into account the memory reserved for the operating system then the amount of memory that could be allowed in such a scenario would be less than 6GB per core. This seems a reasonable amount of memory for a sequential job. However situations where sequential jobs need more memory do exist in high performance computing environments. This is especially true for some of the more complex research problems that are investigated in the HPC facility. Another, solution which seems to suggest itself is to allocate the entire node to a 5
6 sequential job. However, this would be a waste of computing cores if the job uses only limited amount of memory. A reasonable optimization between these conflicting goals was need. This conflict of core versus memory utilization would still arise in parallel submissions but the severity would be less as it is likely that at least one of the resources would be fully utilized. Using Reservations for Sequential Queue Implementation This section details the implementation of the sequential queue at the HPC facility of IIT Kanpur. As brought out in the previous section, conflicting issues of core usage versus memory usages were observed in trying to come up with any scheme of allotting nodes / cores to the sequential queue. The trick in engineering an optimal solution, suited to the needs of the HPC environment, lay in taking the solution of assigning the entire node to the sequential job for a fixed period of time and also allowing the same user to submit more jobs to the sequential queue with the condition that all new jobs would run on the same node. This ensures that that the latest ending time of all jobs submitted to the sequential queue is the same. This was the first solution that was acceptable and was implemented with the wall time restriction on the allocation of the node. Better solutions to this problem could be possible with better placement of jobs and optimization of memory usages. Having outlined the usage policy it was now the task to implement it. There were no ready made mechanisms in PBS Professional that could provide for the exact functionality requested. Thus an implementation for this kind of a queue lay in the use of the mechanics of reservations that are provided by PBS. Reservations allow the user to book resources that would be used subsequently by the user. Once a reservation is granted the user can move some of the jobs into the reservation. Reservations themselves are queues and can be in several states such as Confirmed, Running and Unconfirmed. In a sense reservations once granted are a set of resources that can be thought of as a queue into which a job could be submitted or even moved from another queue. Thus to implement the desired functionality of sequential queue a reservation for a fixed duration is created and jobs are moved into the reservation from the queue if it is in a confirmed state and the job belongs to the user who has created the reservation. The algorithm for doing this is sketched in Algorithm 1. Algorithm: 1 In every Scheduling Cycle or through a periodic cron-job DO 1. REPEAT for a job in the sequential queue Find the user who has submitted the job IF user has an existing reservation THEN IF reservation is in Confirmed or Running State THEN Move the job into the reservation from the sequential queue ENDIF ELSE IF user does not have an existing reservation THEN Append the user to a list of users with no reservations if the user is not already present in the list ENDIF UNTIL all jobs in the sequential queue have been processed 2. Create a reservation of a fixed duration for all users in the list of users with no reservations It is worth noting that for users without reservations only the necessary reservations are created without attempting to move the job into the respective reservations. This task is reserved for the next execution of Algorithm 1 which can be through a cron-job or in a scheduling cycle. Also, for a clean separation of concerns the task is divided 6
7 into two parts i) that of moving jobs into reservations and ii) of creating reservations. This also minimizes the risk of out of order job movements into reservations. The actual code implementing the above algorithm takes into account efficiency of implementation, error checking and logging needs. It can be a bit of a challenge to code this algorithm efficiently. An algorithm closely tied to Algorithm 1 above is that of deleting idle reservations. This simply deletes any reservation that was idle for some k previous execution of the algorithm and is also currently idle. This algorithm can also be tied to the scheduling cycle or to a periodic cron-job. This has also been implemented with a current value of k as 1. In the current implementation all algorithms have been implemented through shell scripts using sed and awk and standard commands provided by PBS Professional for manipulation of reservations and queues. Post Implementation Observations and Generalizations for Future Studies Though several prototypes were made before arriving at the algorithms above, the algorithms outlined above seem to be stable in the current HPC environment and provide the desired functionality. There are enough hooks and interfaces in PBS to smoothly implement the algorithms. In the implementation it is also possible to limit the maximum number of jobs that are moved into the reservation though no upper limit needs be kept on the sequential queue itself. Though the original algorithms were designed keeping in mind that users would use it for submissions of sequential jobs it is observed that users can also submit a parallel job through a single submission of a sequential job that internally uses openmp or even uses mpich that works up-to 8 cores. Is this a violation of the desired functionality? Well the answer is both yes and no as there seems to be no method to force the submissions to be strictly sequential and also allow the entire memory to be used as well as allow all the cores to be used if memory permits. However, this implementation seems to meet the conflicting objectives and works satisfactorily. Clever solutions to the problem posed are not ruled out. It would also be interesting to generalize the problem a bit more. Supposing that the nodes allotted through reservation were more than one per user then it would be interesting to speculate as to how the moving of the jobs from the sequential queue to the reservations happens. Would jobs be moved in a round robin fashion amongst reservations or would one reservation be packed to a job limit before the next reservation is used? Could an additional memory based heuristic help deciding this scheduling problem? Could a best fit of memory requirements be used in conjunction with the reservation packing algorithms? These questions may require careful investigation and further study. Conclusions This paper describes the HPC facility at IIT Kanpur. It states the problem of core utilization versus memory utilization in node allocation to users of sequential services. This problem could still arise in parallel submissions but would be less severe. A solution to this problem is presented through the use of fixed duration reservations. The mechanisms of reservations serves the current goals of the HPC environment adequately enough, however, better solutions to the problem could be possible. The problem becomes interesting when an attempt is made to generalize the specification. Several clever generalization and specifications of the problem may be possible. Acknowledgements The authors would like to thank Scott, Sam G and Rajiv Jaisankar of Altair Engineering for valuable inputs to the PBS Pro based solutions implemented in the HPC facility of IIT Kanpur. Department of Science and Technology and IIT Kanpur have funded the HPC cluster. They would also like to thank Prof. Amalendu Chandra the Head of the Computer Centre of IIT Kanpur. Help of Vairavan Mani for a short duration at IIT Kanpur is duly acknowledged. The user community of the HPC facility at IIT Kanpur also deserves due thanks for their patience while the various versions of the solutions were being implemented. The authors also wish to thank engineers Ram Mohan Shrivastava and Abhishek Kesarwani from HP for their support to this activity. REFERENCES [1] [2] [3] Poster Presentation by hpcsupport at the Symposium on HPC Applications March at IIT Kanpur 7
8 8
HPC Growing Pains. Lessons learned from building a Top500 supercomputer
HPC Growing Pains Lessons learned from building a Top500 supercomputer John L. Wofford Center for Computational Biology & Bioinformatics Columbia University I. What is C2B2? Outline Lessons learned from
More informationRunning applications on the Cray XC30 4/12/2015
Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes
More informationHow to Run Parallel Jobs Efficiently
How to Run Parallel Jobs Efficiently Shao-Ching Huang High Performance Computing Group UCLA Institute for Digital Research and Education May 9, 2013 1 The big picture: running parallel jobs on Hoffman2
More informationRecommended hardware system configurations for ANSYS users
Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range
More information- An Essential Building Block for Stable and Reliable Compute Clusters
Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative
More informationCluster Computing at HRI
Cluster Computing at HRI J.S.Bagla Harish-Chandra Research Institute, Chhatnag Road, Jhunsi, Allahabad 211019. E-mail: jasjeet@mri.ernet.in 1 Introduction and some local history High performance computing
More informationHow To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
More informationVirtual desktops made easy
Product test: DataCore Virtual Desktop Server 2.0 Virtual desktops made easy Dr. Götz Güttich The Virtual Desktop Server 2.0 allows administrators to launch and maintain virtual desktops with relatively
More informationInternational Journal of Advance Research in Computer Science and Management Studies
Volume 3, Issue 6, June 2015 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online
More informationLoad Testing and Monitoring Web Applications in a Windows Environment
OpenDemand Systems, Inc. Load Testing and Monitoring Web Applications in a Windows Environment Introduction An often overlooked step in the development and deployment of Web applications on the Windows
More informationThe Asterope compute cluster
The Asterope compute cluster ÅA has a small cluster named asterope.abo.fi with 8 compute nodes Each node has 2 Intel Xeon X5650 processors (6-core) with a total of 24 GB RAM 2 NVIDIA Tesla M2050 GPGPU
More informationHigh Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates
High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of
More informationMartinos Center Compute Clusters
Intro What are the compute clusters How to gain access Housekeeping Usage Log In Submitting Jobs Queues Request CPUs/vmem Email Status I/O Interactive Dependencies Daisy Chain Wrapper Script In Progress
More informationCORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER
CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER Tender Notice No. 3/2014-15 dated 29.12.2014 (IIT/CE/ENQ/COM/HPC/2014-15/569) Tender Submission Deadline Last date for submission of sealed bids is extended
More informationHPSA Agent Characterization
HPSA Agent Characterization Product HP Server Automation (SA) Functional Area Managed Server Agent Release 9.0 Page 1 HPSA Agent Characterization Quick Links High-Level Agent Characterization Summary...
More informationLLamasoft K2 Enterprise 8.1 System Requirements
Overview... 3 RAM... 3 Cores and CPU Speed... 3 Local System for Operating Supply Chain Guru... 4 Applying Supply Chain Guru Hardware in K2 Enterprise... 5 Example... 6 Determining the Correct Number of
More informationWorking with HPC and HTC Apps. Abhinav Thota Research Technologies Indiana University
Working with HPC and HTC Apps Abhinav Thota Research Technologies Indiana University Outline What are HPC apps? Working with typical HPC apps Compilers - Optimizations and libraries Installation Modules
More informationsupercomputing. simplified.
supercomputing. simplified. INTRODUCING WINDOWS HPC SERVER 2008 R2 SUITE Windows HPC Server 2008 R2, Microsoft s third-generation HPC solution, provides a comprehensive and costeffective solution for harnessing
More informationGrid Computing Approach for Dynamic Load Balancing
International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-1 E-ISSN: 2347-2693 Grid Computing Approach for Dynamic Load Balancing Kapil B. Morey 1*, Sachin B. Jadhav
More informationWindows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...
More informationPurchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers
Information Technology Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers Effective for FY2016 Purpose This document summarizes High Performance Computing
More informationPerformance Analysis of Remote Desktop Virtualization based on Hyper-V versus Remote Desktop Services
MACRo 2015-5 th International Conference on Recent Achievements in Mechatronics, Automation, Computer Science and Robotics Performance Analysis of Remote Desktop Virtualization based on Hyper-V versus
More informationCluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
More informationChapter 2: Getting Started
Chapter 2: Getting Started Once Partek Flow is installed, Chapter 2 will take the user to the next stage and describes the user interface and, of note, defines a number of terms required to understand
More informationThe PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver
1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution
More informationAgenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
More informationDragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers
Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers This section includes system requirements for DMENE Network configurations that utilize virtual
More informationHPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014
HPC Cluster Decisions and ANSYS Configuration Best Practices Diana Collier Lead Systems Support Specialist Houston UGM May 2014 1 Agenda Introduction Lead Systems Support Specialist Cluster Decisions Job
More informationWindows Server Performance Monitoring
Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly
More informationQuantum StorNext. Product Brief: Distributed LAN Client
Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without
More informationConfiguration Maximums
Topic Configuration s VMware vsphere 5.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 5.0. The limits presented in the
More informationAnalysis of VDI Storage Performance During Bootstorm
Analysis of VDI Storage Performance During Bootstorm Introduction Virtual desktops are gaining popularity as a more cost effective and more easily serviceable solution. The most resource-dependent process
More informationNEC HPC-Linux-Cluster
NEC HPC-Linux-Cluster Hardware configuration: 4 Front-end servers: each with SandyBridge-EP processors: 16 cores per node 128 GB memory 134 compute nodes: 112 nodes with SandyBridge-EP processors (16 cores
More informationIntroduction to Supercomputing with Janus
Introduction to Supercomputing with Janus Shelley Knuth shelley.knuth@colorado.edu Peter Ruprecht peter.ruprecht@colorado.edu www.rc.colorado.edu Outline Who is CU Research Computing? What is a supercomputer?
More informationClusters: Mainstream Technology for CAE
Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
More informationMicrosoft HPC. V 1.0 José M. Cámara (checam@ubu.es)
Microsoft HPC V 1.0 José M. Cámara (checam@ubu.es) Introduction Microsoft High Performance Computing Package addresses computing power from a rather different approach. It is mainly focused on commodity
More informationThe Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist
The Top Six Advantages of CUDA-Ready Clusters Ian Lumb Bright Evangelist GTC Express Webinar January 21, 2015 We scientists are time-constrained, said Dr. Yamanaka. Our priority is our research, not managing
More informationAdvanced Techniques with Newton. Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011
Advanced Techniques with Newton Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011 Workshop Goals Gain independence Executing your work Finding Information Fixing Problems Optimizing Effectiveness
More informationAn introduction to Fyrkat
Cluster Computing May 25, 2011 How to get an account https://fyrkat.grid.aau.dk/useraccount How to get help https://fyrkat.grid.aau.dk/wiki What is a Cluster Anyway It is NOT something that does any of
More informationPriority Pro v17: Hardware and Supporting Systems
Introduction Priority Pro v17: Hardware and Supporting Systems The following provides minimal system configuration requirements for Priority with respect to three types of installations: On-premise Priority
More informationAn Introduction to High Performance Computing in the Department
An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software
More informationHow To Live Migrate In Hyperv On Windows Server 22 (Windows) (Windows V) (Hyperv) (Powerpoint) (For A Hyperv Virtual Machine) (Virtual Machine) And (Hyper V) Vhd (Virtual Hard Disk
Poster Companion Reference: Hyper-V Virtual Machine Mobility Live Migration Without Shared Storage Storage Migration Live Migration with SMB Shared Storage Live Migration with Failover Clusters Copyright
More informationOverview of HPC systems and software available within
Overview of HPC systems and software available within Overview Available HPC Systems Ba Cy-Tera Available Visualization Facilities Software Environments HPC System at Bibliotheca Alexandrina SUN cluster
More informationPARALLELS CLOUD STORAGE
PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...
More informationStreamline Computing Linux Cluster User Training. ( Nottingham University)
1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running
More informationContact for all enquiries Phone: +61 2 8006 9730. Email: info@recordpoint.com.au. Page 2. RecordPoint Release Notes V3.8 for SharePoint 2013
Release Notes V3.8 Notice This document contains confidential and trade secret information of RecordPoint Software ( RPS ). RecordPoint Software has prepared this document for use solely with RecordPoint.
More informationBuilding a Linux Cluster
Building a Linux Cluster CUG Conference May 21-25, 2001 by Cary Whitney Clwhitney@lbl.gov Outline What is PDSF and a little about its history. Growth problems and solutions. Storage Network Hardware Administration
More informationCloud Computing through Virtualization and HPC technologies
Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC
More informationWork Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015
Work Environment David Tur HPC Expert HPC Users Training September, 18th 2015 1. Atlas Cluster: Accessing and using resources 2. Software Overview 3. Job Scheduler 1. Accessing Resources DIPC technicians
More informationDeploying and Optimizing SQL Server for Virtual Machines
Deploying and Optimizing SQL Server for Virtual Machines Deploying and Optimizing SQL Server for Virtual Machines Much has been written over the years regarding best practices for deploying Microsoft SQL
More informationAssetTrack. Overview: AMI AssetTrack Integration for HP Asset Manager. Mobile Forms/Scanners. Web Forms CMYK: 62 42 73 25
Overview: AMI AssetTrack Integration for HP Asset Manager AssetTrack is an enterprise barcode asset tracking solution for HP Asset Manager that enables Asset Manager customers to use handheld barcode scanners
More informationQuantcast Petabyte Storage at Half Price with QFS!
9-131 Quantcast Petabyte Storage at Half Price with QFS Presented by Silvius Rus, Director, Big Data Platforms September 2013 Quantcast File System (QFS) A high performance alternative to the Hadoop Distributed
More informationLCMON Network Traffic Analysis
LCMON Network Traffic Analysis Adam Black Centre for Advanced Internet Architectures, Technical Report 79A Swinburne University of Technology Melbourne, Australia adamblack@swin.edu.au Abstract The Swinburne
More informationAvoiding Performance Bottlenecks in Hyper-V
Avoiding Performance Bottlenecks in Hyper-V Identify and eliminate capacity related performance bottlenecks in Hyper-V while placing new VMs for optimal density and performance Whitepaper by Chris Chesley
More informationGC3: Grid Computing Competence Center Cluster computing, I Batch-queueing systems
GC3: Grid Computing Competence Center Cluster computing, I Batch-queueing systems Riccardo Murri, Sergio Maffioletti Grid Computing Competence Center, Organisch-Chemisches Institut, University of Zurich
More informationInternational Engineering Journal For Research & Development
Evolution Of Operating System And Open Source Android Application Nilesh T.Gole 1, Amit Manikrao 2, Niraj Kanot 3,Mohan Pande 4 1,M.tech(CSE)JNTU, 2 M.tech(CSE)SGBAU, 3 M.tech(CSE),JNTU, Hyderabad 1 sheyanilu@gmail.com,
More informationParallel Computing with MATLAB
Parallel Computing with MATLAB Scott Benway Senior Account Manager Jiro Doke, Ph.D. Senior Application Engineer 2013 The MathWorks, Inc. 1 Acceleration Strategies Applied in MATLAB Approach Options Best
More informationUsing Parallel Computing to Run Multiple Jobs
Beowulf Training Using Parallel Computing to Run Multiple Jobs Jeff Linderoth August 5, 2003 August 5, 2003 Beowulf Training Running Multiple Jobs Slide 1 Outline Introduction to Scheduling Software The
More informationCertification: HP ATA Servers & Storage
HP ExpertONE Competency Model Certification: HP ATA Servers & Storage Overview Achieving an HP certification provides relevant skills that can lead to a fulfilling career in Information Technology. HP
More informationOptimizing Shared Resource Contention in HPC Clusters
Optimizing Shared Resource Contention in HPC Clusters Sergey Blagodurov Simon Fraser University Alexandra Fedorova Simon Fraser University Abstract Contention for shared resources in HPC clusters occurs
More informationLecture 1: the anatomy of a supercomputer
Where a calculator on the ENIAC is equipped with 18,000 vacuum tubes and weighs 30 tons, computers of the future may have only 1,000 vacuum tubes and perhaps weigh 1½ tons. Popular Mechanics, March 1949
More informationImproving Compute Farm Efficiency for EDA
Improving Compute Farm Efficiency for EDA Many IT managers report that the average utilization of their compute farms is just 50-60%. Neel Desai, product marketing manager, Lynx Design System, explains
More informationConfiguration Maximums VMware vsphere 4.0
Topic Configuration s VMware vsphere 4.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.0. The limits presented in the
More informationAn Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing
An Alternative Storage Solution for MapReduce Eric Lomascolo Director, Solutions Marketing MapReduce Breaks the Problem Down Data Analysis Distributes processing work (Map) across compute nodes and accumulates
More informationFLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
More informationGetting Started with HPC
Getting Started with HPC An Introduction to the Minerva High Performance Computing Resource 17 Sep 2013 Outline of Topics Introduction HPC Accounts Logging onto the HPC Clusters Common Linux Commands Storage
More informationUsing WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014
Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan
More informationConfiguration Maximums VMware vsphere 4.1
Topic Configuration s VMware vsphere 4.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.1. The limits presented in the
More informationUsing the Windows Cluster
Using the Windows Cluster Christian Terboven terboven@rz.rwth aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster
More informationAppro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
More informationDeploying Exchange Server 2007 SP1 on Windows Server 2008
Deploying Exchange Server 2007 SP1 on Windows Server 2008 Product Group - Enterprise Dell White Paper By Ananda Sankaran Andrew Bachler April 2008 Contents Introduction... 3 Deployment Considerations...
More informationAgenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.
Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance
More informationSystem Software for High Performance Computing. Joe Izraelevitz
System Software for High Performance Computing Joe Izraelevitz Agenda Overview of Supercomputers Blue Gene/Q System LoadLeveler Job Scheduler General Parallel File System HPC at UR What is a Supercomputer?
More informationAchieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
More informationImprove Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database
WHITE PAPER Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive
More informationSRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center
SRNWP Workshop HP Solutions and Activities in Climate & Weather Research Michael Riedmann European Performance Center Agenda A bit of marketing: HP Solutions for HPC A few words about recent Met deals
More informationSolution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware
Introduction By leveraging the inherent benefits of a virtualization based platform, a Microsoft Exchange Server 2007 deployment on VMware Infrastructure 3 offers a variety of availability and recovery
More informationWindows Server 2008 R2 Hyper-V Live Migration
Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described
More informationEMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi
EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Applied Technology Abstract Microsoft SQL Server includes a powerful capability to protect active databases by using either
More informationEnergy Constrained Resource Scheduling for Cloud Environment
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
More informationRecommendations for Performance Benchmarking
Recommendations for Performance Benchmarking Shikhar Puri Abstract Performance benchmarking of applications is increasingly becoming essential before deployment. This paper covers recommendations and best
More informationCloud Management: Knowing is Half The Battle
Cloud Management: Knowing is Half The Battle Raouf BOUTABA David R. Cheriton School of Computer Science University of Waterloo Joint work with Qi Zhang, Faten Zhani (University of Waterloo) and Joseph
More informationParallel Programming for Multi-Core, Distributed Systems, and GPUs Exercises
Parallel Programming for Multi-Core, Distributed Systems, and GPUs Exercises Pierre-Yves Taunay Research Computing and Cyberinfrastructure 224A Computer Building The Pennsylvania State University University
More informationPBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007
PBS Tutorial Fangrui Ma Universit of Nebraska-Lincoln October 26th, 2007 Abstract In this tutorial we gave a brief introduction to using PBS Pro. We gave examples on how to write control script, and submit
More informationBig Data and Cloud Computing for GHRSST
Big Data and Cloud Computing for GHRSST Jean-Francois Piollé (jfpiolle@ifremer.fr) Frédéric Paul, Olivier Archer CERSAT / Institut Français de Recherche pour l Exploitation de la Mer Facing data deluge
More informationWITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What
More informationData processing goes big
Test report: Integration Big Data Edition Data processing goes big Dr. Götz Güttich Integration is a powerful set of tools to access, transform, move and synchronize data. With more than 450 connectors,
More informationMPI / ClusterTools Update and Plans
HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski
More informationChapter 2: OS Overview
Chapter 2: OS Overview CmSc 335 Operating Systems 1. Operating system objectives and functions Operating systems control and support the usage of computer systems. a. usage users of a computer system:
More informationx64 Servers: Do you want 64 or 32 bit apps with that server?
TMurgent Technologies x64 Servers: Do you want 64 or 32 bit apps with that server? White Paper by Tim Mangan TMurgent Technologies February, 2006 Introduction New servers based on what is generally called
More informationHigh Availability of the Polarion Server
Polarion Software CONCEPT High Availability of the Polarion Server Installing Polarion in a high availability environment Europe, Middle-East, Africa: Polarion Software GmbH Hedelfinger Straße 60 70327
More informationPerformance Optimization Guide
Performance Optimization Guide Publication Date: July 06, 2016 Copyright Metalogix International GmbH, 2001-2016. All Rights Reserved. This software is protected by copyright law and international treaties.
More informationCisco Unified Computing System and EMC VNX5300 Unified Storage Platform
Cisco Unified Computing System and EMC VNX5300 Unified Storage Platform Implementing an Oracle Data Warehouse Test Workload White Paper January 2011, Revision 1.0 Contents Executive Summary... 3 Cisco
More informationEstimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010
Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 This document is provided as-is. Information and views expressed in this document, including URL and other Internet
More informationStep by Step Guide To vstorage Backup Server (Proxy) Sizing
Tivoli Storage Manager for Virtual Environments V6.3 Step by Step Guide To vstorage Backup Server (Proxy) Sizing 12 September 2012 1.1 Author: Dan Wolfe, Tivoli Software Advanced Technology Page 1 of 18
More informationSetup for Failover Clustering and Microsoft Cluster Service
Setup for Failover Clustering and Microsoft Cluster Service Update 1 ESX 4.0 ESXi 4.0 vcenter Server 4.0 This document supports the version of each product listed and supports all subsequent versions until
More informationHP high availability solutions for Microsoft SQL Server Fast Track Data Warehouse using SQL Server 2012 failover clustering
Technical white paper HP high availability solutions for Microsoft SQL Server Fast Track Data Warehouse using SQL Server 2012 failover clustering Table of contents Executive summary 2 Fast Track reference
More informationBenchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
More information