Using Reservations to Implement Fixed Duration Node Allotment with PBS Professional

Size: px
Start display at page:

Download "Using Reservations to Implement Fixed Duration Node Allotment with PBS Professional"

Transcription

1 Using Reservations to Implement Fixed Duration Node Allotment with PBS Professional Brajesh Pande Senior Computer Engineer Computer Centre IIT Kanpur Kanpur, UP India Manoj Soni Technical Consultant Altair India Chintels Techo Park, Kailash Colony New Delhi , India Dario Dorella PBS Works Technical Specialist Altair Engineering c/o Environment Park,Via Livorno Torino,Italy Abbreviations: HPC High Performance Computing, CPU Central Processing Unit, ILO Integrated lights out, GPU Graphical Processing Unit, VLAN Virtual Local Area Network, GbPS Giga bits per second, GBPS Giga bytes per second, CMU Cluster Management Utility, HP Hewlett Packard, SAN storage area network Keywords: High Performance Computing, Schedulers, PBS Professional, reservations, resource allocation, node allotment Abstract Indian Institute of Technology Kanpur has a High Performance Computing Facility that was in the top 500 list [1] of June PBS professional, a part of PBS-Works [2] solution from Altair, finds interesting use in implementing a sequential queue on this facility. A user who submits a job to this queue is given the non-interactive use of an 8-core-compute-node for a fixed period of time. The envisioned use of this allotment is to allow the user to run any number of sequential jobs on the given node with the condition that the latest ending time of all such jobs be the same. It is expected that this queue be used only for sequential jobs yet clever parallel use of the queue are not ruled out. The desired functionality is achieved by creating a fixed duration reservation for the user after the first job is submitted to the queue. The first and subsequent jobs of the same user on this queue are run on the same reservation. We detail this implementation and touch upon some generalizations of this concept. We also briefly describe the facility, its queuing structure and other provisions that leverage PBS Professional functionality to ensure the smooth functioning of the facility Introduction A High Performance Computing (HPC) machine typically consists of several compute nodes connected together through some interconnect. The compute nodes could be multiple CPU / Core machines or even GPU based machines while interconnects could be the memory paths or the network with varying speeds or a combination of both. Multiple CPUs / CORE machines use the memory paths for intra-node communication and the network for inter-node communication. Additional hardware and various complex tools become evermore necessary as the number of nodes in the facility increase. Thus these machines also have service processors for ILO management, System Deployment Tools, Tools for Monitoring and Management of Nodes, some Head nodes for access and utility services apart from the necessary CPUs, software and operating systems. These machines also have a scheduler, the focus of this paper that manages job distribution on the nodes. Schedulers view the machine as a collection of resources. A user uses the services of a scheduler to request for resources satisfying some desired criteria such as the number of compute nodes required, the amount of memory on the nodes, the duration for which the nodes are to be given and the licenses for software that are needed. Resources can not only be mapped to physical objects but can also be virtualized and defined during the configuration of the scheduler and can also be dynamically discovered. If the requested resources are available, and the policy configured on the scheduler permits allocation, the available resources are released to the user. The scheduler restricts the usage of resources by configured policy dictates. Schedulers maintain job queues with attributes that can be linked to the resources. The job queues and the linkage between the resources and the queues are defined during the configuration of the queues. Jobs that are submitted by the user (resource requests) to a queue are either finally scheduled on some resources by the scheduler based on the tunable internal algorithms that take the policies into account or else wait in the queue. 1

2 Since configuring the policies, the queues, and the resources on a scheduler can be a complex task, not all desired functionality from the scheduler can be accomplished by simple specifications. Good schedulers recognize this and provide hooks and interfaces that allow the scheduler to meet, with varying degrees of satisfaction, the user requirements after some programming and effort. Such functionality is driven by user needs. If such needs are general enough they can be over a period of time be incorporated within the framework of specifications of the scheduler. This paper focuses on a functionality desired for the HPC facility at IIT Kanpur that uses PBS Professional as its scheduler. It describes the HPC facility in brief, talks about its queue structures and the regular use of PBS Professional in the facility. It then goes on to enunciate the need for the more complex functionality. This functionality implements what the authors have named a sequential queue. A user who submits a job to this queue is given the non-interactive use of an 8-core-compute-node for a fixed period of time. The envisioned use of this allotment is to allow the user to run any number of sequential jobs on the given node with the condition that the latest ending time of all such jobs be the same. This paper also provides a solution for the implementation of this functionality through the use of interfaces provides by PBS Professional. Main HPC Facility at IIT Kanpur Figure 1, shows the architecture of the HPC facility at IIT Kanpur [3]. This facility has 368 compute nodes and 4 management nodes and 4 utility nodes. The blade type compute nodes are 2 Nehalem Intel-CPUs with 4 cores each on a BL-280c-G6 server and 48 GB of RAM. Each node has a 146GB small form factor Serial AT Attachment (SATA) disk for holding the boot images, operating systems and some necessary software. These nodes are connected together by several networks. The first network that connects these nodes is a boot network that is primarily used for booting the systems, and managerial functionality. The second network connecting them is a managerial network that connects ILO system and the console of various machines. These networks are represented by a single VLAN shown in the figure as Ethernet & ILO. All the nodes are also connected by a QDR-Infiniband network from QLogic that has a speed of 40 GbPS. This is a very fast interconnect and separation of management concerns from user I/O concerns makes for a fast through put of the users of the HPC system. Three file systems that use Modular Smart Array of disks (MSA) P2000 G3 FC from HP are also present in the entire facility. These disks use the Fibre-Channel Network to connect to SAN switches that finally server data for the file systems. The first file system of 7.4 TB is a NFS file system that is used for administrative purpose such as holding boot images and for miscellaneous storage purposes. The second file system is a Lustre based file system that has a high read and write throughput (around 3.4 GBPS read and 2.4 GBPS write throughput). This file system is called the scratch file system and has 40 TB of capacity. There is a slower home file system of 60 TB again on Lustre (1.7 GBPS read and 1.3 GBPS write throughput). Two different Lustre file systems separate out user concerns between unlimited scratch-space during compute operations versus quota based slower home file systems. 2

3 Figure1: The Architecture of HPC Facility at IIT Kanpur The operating system on the machines is RHEL5.4. The cluster uses CMU tool from HP for management and monitoring of the systems and deployment of operating system and software on the nodes. It uses PBS Professional as the scheduler for jobs submitted by the user. PBS Analytics and Compute Manager have also been deployed on the cluster The machine is mostly used by students, faculty and researchers of IIT Kanpur. Some users from other academic institutions also use the facility. Typically users come from almost all departments of the institute and use the facility for computational purposes through home grown software, libraries and compilers as well as available third party software that include but are not limited to: Gaussian, TurboMole, Ansys, Matlab, Accelrys, CPMD, WRF, Metis, MedeaVasp, Laamps, Tinker, NAG parallel libraries and compilers, GAMS, gcc, INTEL suite of compilers, pgi compiler from PORTLAND, many MPI libraries variations of mvapich2 available from QLogic, openmp, and flavors of mpich2. This machine has a peak theoretical performance of TFLOP and a maximum performance of TFLOP and was ranked 369 in the Top 500 list of June A specific usage scenario of the machine is shown in Figure 2 that was prepared by pulling out data with the help of PBS-Analytics. It shows that some reservations apart from the main queues are also in use and that the average wait time for jobs is increasing indicating a need for more compute nodes. Use of reservations in this facility is discussed more later. 3

4 Figure: 2 Typical usage of the Queues of the HPC Facility Use of PBS Professional in the HPC Facility: Currently the system uses PBS Professional version 11.2 which was upgraded from an earlier 10.4 release. A PBS failover server is also deployed for high availability. The system is divided into several queues as shown in the Table 1. These will be discussed in brief till the focus shifts to the main queue under discussion that is the sequential queue labeled seq in the table. 4

5 TABLE I QUEUES OF THE HPC FACILITY AT IIT KANPUR Queue Total Available Min Max Nodes Min Max Cores Allowed CPU / Nodes Allowed Allowed Wall Time large Max / 72 hrs medium Max / 96 hrs small Max/ 120 hrs workq hrs/48 hrs test 3 any any Max / Max seq 16 1 any Max / 120 hrs Table 1 shows the queues, total nodes available in each queue, minimum and maximum nodes allowed to a for a job, minimum and maximum cores allowed for a job and the maximum CPU and wall time allowed on the node. The idea behind these restrictions stems from the fact that given a reasonably sized problem, most HPC problems do not scale up beyond a certain number of nodes. Restrictions on wall time are necessary for giving fair access of the queues to users. There are no special privileges for any user of the cluster currently. Provision for allowing jobs with more number of cores or wall-time is possible but is reserved with the policy making body for HPC usage. The large queue allows for more number of cores but lesser wall time, the medium allows lesser cores but more wall time compared to the large queue and the small queue allows even lesser cores but more wall time than the medium queue. This arrangement is well suited for scaling studies. Work queue (workq) is a special queue that is allows both interactive as well as batch jobs. The idea behind this queue is that users would land on this queue for day to day activities and would be placed in a round robin fashion on the nodes of the work queue. This load balances the interactive usage of the queue. This is achieved by defining a static consumable resource on each node that is used when the node is in use. Allocation of the node of the work queue is done after sorting on the available numbers of this resource on a node. The round robin allocation of the batch processing in the work queue is a bit trickier and is implemented by a runtime hook that also makes the work queue an interactive queue. The implementation of the work queue more or less follows standard implementation examples in the PBS Professional manual with the necessary programming for the round robin allocation. The total number of round robin resources associated with the work queue decide the amount to which this queue can be loaded. The test queue is reserved for testing of new software and implementations. The sequential queue (seq) is the main focus of this paper and is discussed in more detail in the next section. The Sequential Queue requirements The need for the sequential queue arises from the fact that though HPC machines are mostly used for parallel applications sequential needs of users cannot be ruled out especially if the users belong to diverse computing environments. There may be legacy code as well as some code that is inherently sequential. Thus provision has to be made for such a queue. An obvious solution to this is to assign a core to any user who wants to use the sequential queue. The problem with this simplistic allocation policy is that the assigned core belongs to a machine that has memory which is shared by the other cores. If the job that has been allotted the single core uses large amounts of memory it would ultimately unfairly use the memory resources allowed to a job on another core on the same node. This is clearly an undesirable situation and would ultimately slow down all jobs that contend for memory on this node. In nodes that are severely affected almost all the memory and even the swap space would be used up and this would result in hanging the node. A user using memory within limits would also be penalized with non performance in this scenario. One of the ways out of this dilemma would be to restrict the memory use of a single job on the sequential queue. If we do a calculation with 48 GB of memory amongst 8 cores and also take into account the memory reserved for the operating system then the amount of memory that could be allowed in such a scenario would be less than 6GB per core. This seems a reasonable amount of memory for a sequential job. However situations where sequential jobs need more memory do exist in high performance computing environments. This is especially true for some of the more complex research problems that are investigated in the HPC facility. Another, solution which seems to suggest itself is to allocate the entire node to a 5

6 sequential job. However, this would be a waste of computing cores if the job uses only limited amount of memory. A reasonable optimization between these conflicting goals was need. This conflict of core versus memory utilization would still arise in parallel submissions but the severity would be less as it is likely that at least one of the resources would be fully utilized. Using Reservations for Sequential Queue Implementation This section details the implementation of the sequential queue at the HPC facility of IIT Kanpur. As brought out in the previous section, conflicting issues of core usage versus memory usages were observed in trying to come up with any scheme of allotting nodes / cores to the sequential queue. The trick in engineering an optimal solution, suited to the needs of the HPC environment, lay in taking the solution of assigning the entire node to the sequential job for a fixed period of time and also allowing the same user to submit more jobs to the sequential queue with the condition that all new jobs would run on the same node. This ensures that that the latest ending time of all jobs submitted to the sequential queue is the same. This was the first solution that was acceptable and was implemented with the wall time restriction on the allocation of the node. Better solutions to this problem could be possible with better placement of jobs and optimization of memory usages. Having outlined the usage policy it was now the task to implement it. There were no ready made mechanisms in PBS Professional that could provide for the exact functionality requested. Thus an implementation for this kind of a queue lay in the use of the mechanics of reservations that are provided by PBS. Reservations allow the user to book resources that would be used subsequently by the user. Once a reservation is granted the user can move some of the jobs into the reservation. Reservations themselves are queues and can be in several states such as Confirmed, Running and Unconfirmed. In a sense reservations once granted are a set of resources that can be thought of as a queue into which a job could be submitted or even moved from another queue. Thus to implement the desired functionality of sequential queue a reservation for a fixed duration is created and jobs are moved into the reservation from the queue if it is in a confirmed state and the job belongs to the user who has created the reservation. The algorithm for doing this is sketched in Algorithm 1. Algorithm: 1 In every Scheduling Cycle or through a periodic cron-job DO 1. REPEAT for a job in the sequential queue Find the user who has submitted the job IF user has an existing reservation THEN IF reservation is in Confirmed or Running State THEN Move the job into the reservation from the sequential queue ENDIF ELSE IF user does not have an existing reservation THEN Append the user to a list of users with no reservations if the user is not already present in the list ENDIF UNTIL all jobs in the sequential queue have been processed 2. Create a reservation of a fixed duration for all users in the list of users with no reservations It is worth noting that for users without reservations only the necessary reservations are created without attempting to move the job into the respective reservations. This task is reserved for the next execution of Algorithm 1 which can be through a cron-job or in a scheduling cycle. Also, for a clean separation of concerns the task is divided 6

7 into two parts i) that of moving jobs into reservations and ii) of creating reservations. This also minimizes the risk of out of order job movements into reservations. The actual code implementing the above algorithm takes into account efficiency of implementation, error checking and logging needs. It can be a bit of a challenge to code this algorithm efficiently. An algorithm closely tied to Algorithm 1 above is that of deleting idle reservations. This simply deletes any reservation that was idle for some k previous execution of the algorithm and is also currently idle. This algorithm can also be tied to the scheduling cycle or to a periodic cron-job. This has also been implemented with a current value of k as 1. In the current implementation all algorithms have been implemented through shell scripts using sed and awk and standard commands provided by PBS Professional for manipulation of reservations and queues. Post Implementation Observations and Generalizations for Future Studies Though several prototypes were made before arriving at the algorithms above, the algorithms outlined above seem to be stable in the current HPC environment and provide the desired functionality. There are enough hooks and interfaces in PBS to smoothly implement the algorithms. In the implementation it is also possible to limit the maximum number of jobs that are moved into the reservation though no upper limit needs be kept on the sequential queue itself. Though the original algorithms were designed keeping in mind that users would use it for submissions of sequential jobs it is observed that users can also submit a parallel job through a single submission of a sequential job that internally uses openmp or even uses mpich that works up-to 8 cores. Is this a violation of the desired functionality? Well the answer is both yes and no as there seems to be no method to force the submissions to be strictly sequential and also allow the entire memory to be used as well as allow all the cores to be used if memory permits. However, this implementation seems to meet the conflicting objectives and works satisfactorily. Clever solutions to the problem posed are not ruled out. It would also be interesting to generalize the problem a bit more. Supposing that the nodes allotted through reservation were more than one per user then it would be interesting to speculate as to how the moving of the jobs from the sequential queue to the reservations happens. Would jobs be moved in a round robin fashion amongst reservations or would one reservation be packed to a job limit before the next reservation is used? Could an additional memory based heuristic help deciding this scheduling problem? Could a best fit of memory requirements be used in conjunction with the reservation packing algorithms? These questions may require careful investigation and further study. Conclusions This paper describes the HPC facility at IIT Kanpur. It states the problem of core utilization versus memory utilization in node allocation to users of sequential services. This problem could still arise in parallel submissions but would be less severe. A solution to this problem is presented through the use of fixed duration reservations. The mechanisms of reservations serves the current goals of the HPC environment adequately enough, however, better solutions to the problem could be possible. The problem becomes interesting when an attempt is made to generalize the specification. Several clever generalization and specifications of the problem may be possible. Acknowledgements The authors would like to thank Scott, Sam G and Rajiv Jaisankar of Altair Engineering for valuable inputs to the PBS Pro based solutions implemented in the HPC facility of IIT Kanpur. Department of Science and Technology and IIT Kanpur have funded the HPC cluster. They would also like to thank Prof. Amalendu Chandra the Head of the Computer Centre of IIT Kanpur. Help of Vairavan Mani for a short duration at IIT Kanpur is duly acknowledged. The user community of the HPC facility at IIT Kanpur also deserves due thanks for their patience while the various versions of the solutions were being implemented. The authors also wish to thank engineers Ram Mohan Shrivastava and Abhishek Kesarwani from HP for their support to this activity. REFERENCES [1] [2] [3] Poster Presentation by hpcsupport at the Symposium on HPC Applications March at IIT Kanpur 7

8 8

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

HPC Growing Pains. Lessons learned from building a Top500 supercomputer HPC Growing Pains Lessons learned from building a Top500 supercomputer John L. Wofford Center for Computational Biology & Bioinformatics Columbia University I. What is C2B2? Outline Lessons learned from

More information

Running applications on the Cray XC30 4/12/2015

Running applications on the Cray XC30 4/12/2015 Running applications on the Cray XC30 4/12/2015 1 Running on compute nodes By default, users do not log in and run applications on the compute nodes directly. Instead they launch jobs on compute nodes

More information

How to Run Parallel Jobs Efficiently

How to Run Parallel Jobs Efficiently How to Run Parallel Jobs Efficiently Shao-Ching Huang High Performance Computing Group UCLA Institute for Digital Research and Education May 9, 2013 1 The big picture: running parallel jobs on Hoffman2

More information

Recommended hardware system configurations for ANSYS users

Recommended hardware system configurations for ANSYS users Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range

More information

- An Essential Building Block for Stable and Reliable Compute Clusters

- An Essential Building Block for Stable and Reliable Compute Clusters Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative

More information

Cluster Computing at HRI

Cluster Computing at HRI Cluster Computing at HRI J.S.Bagla Harish-Chandra Research Institute, Chhatnag Road, Jhunsi, Allahabad 211019. E-mail: jasjeet@mri.ernet.in 1 Introduction and some local history High performance computing

More information

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) ( TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx

More information

Virtual desktops made easy

Virtual desktops made easy Product test: DataCore Virtual Desktop Server 2.0 Virtual desktops made easy Dr. Götz Güttich The Virtual Desktop Server 2.0 allows administrators to launch and maintain virtual desktops with relatively

More information

International Journal of Advance Research in Computer Science and Management Studies

International Journal of Advance Research in Computer Science and Management Studies Volume 3, Issue 6, June 2015 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online

More information

Load Testing and Monitoring Web Applications in a Windows Environment

Load Testing and Monitoring Web Applications in a Windows Environment OpenDemand Systems, Inc. Load Testing and Monitoring Web Applications in a Windows Environment Introduction An often overlooked step in the development and deployment of Web applications on the Windows

More information

The Asterope compute cluster

The Asterope compute cluster The Asterope compute cluster ÅA has a small cluster named asterope.abo.fi with 8 compute nodes Each node has 2 Intel Xeon X5650 processors (6-core) with a total of 24 GB RAM 2 NVIDIA Tesla M2050 GPGPU

More information

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of

More information

Martinos Center Compute Clusters

Martinos Center Compute Clusters Intro What are the compute clusters How to gain access Housekeeping Usage Log In Submitting Jobs Queues Request CPUs/vmem Email Status I/O Interactive Dependencies Daisy Chain Wrapper Script In Progress

More information

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER Tender Notice No. 3/2014-15 dated 29.12.2014 (IIT/CE/ENQ/COM/HPC/2014-15/569) Tender Submission Deadline Last date for submission of sealed bids is extended

More information

HPSA Agent Characterization

HPSA Agent Characterization HPSA Agent Characterization Product HP Server Automation (SA) Functional Area Managed Server Agent Release 9.0 Page 1 HPSA Agent Characterization Quick Links High-Level Agent Characterization Summary...

More information

LLamasoft K2 Enterprise 8.1 System Requirements

LLamasoft K2 Enterprise 8.1 System Requirements Overview... 3 RAM... 3 Cores and CPU Speed... 3 Local System for Operating Supply Chain Guru... 4 Applying Supply Chain Guru Hardware in K2 Enterprise... 5 Example... 6 Determining the Correct Number of

More information

Working with HPC and HTC Apps. Abhinav Thota Research Technologies Indiana University

Working with HPC and HTC Apps. Abhinav Thota Research Technologies Indiana University Working with HPC and HTC Apps Abhinav Thota Research Technologies Indiana University Outline What are HPC apps? Working with typical HPC apps Compilers - Optimizations and libraries Installation Modules

More information

supercomputing. simplified.

supercomputing. simplified. supercomputing. simplified. INTRODUCING WINDOWS HPC SERVER 2008 R2 SUITE Windows HPC Server 2008 R2, Microsoft s third-generation HPC solution, provides a comprehensive and costeffective solution for harnessing

More information

Grid Computing Approach for Dynamic Load Balancing

Grid Computing Approach for Dynamic Load Balancing International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-1 E-ISSN: 2347-2693 Grid Computing Approach for Dynamic Load Balancing Kapil B. Morey 1*, Sachin B. Jadhav

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...

More information

Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers

Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers Information Technology Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers Effective for FY2016 Purpose This document summarizes High Performance Computing

More information

Performance Analysis of Remote Desktop Virtualization based on Hyper-V versus Remote Desktop Services

Performance Analysis of Remote Desktop Virtualization based on Hyper-V versus Remote Desktop Services MACRo 2015-5 th International Conference on Recent Achievements in Mechatronics, Automation, Computer Science and Robotics Performance Analysis of Remote Desktop Virtualization based on Hyper-V versus

More information

Cluster Implementation and Management; Scheduling

Cluster Implementation and Management; Scheduling Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /

More information

Chapter 2: Getting Started

Chapter 2: Getting Started Chapter 2: Getting Started Once Partek Flow is installed, Chapter 2 will take the user to the next stage and describes the user interface and, of note, defines a number of terms required to understand

More information

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver 1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution

More information

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical

More information

Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers

Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers This section includes system requirements for DMENE Network configurations that utilize virtual

More information

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014

HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014 HPC Cluster Decisions and ANSYS Configuration Best Practices Diana Collier Lead Systems Support Specialist Houston UGM May 2014 1 Agenda Introduction Lead Systems Support Specialist Cluster Decisions Job

More information

Windows Server Performance Monitoring

Windows Server Performance Monitoring Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly

More information

Quantum StorNext. Product Brief: Distributed LAN Client

Quantum StorNext. Product Brief: Distributed LAN Client Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without

More information

Configuration Maximums

Configuration Maximums Topic Configuration s VMware vsphere 5.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 5.0. The limits presented in the

More information

Analysis of VDI Storage Performance During Bootstorm

Analysis of VDI Storage Performance During Bootstorm Analysis of VDI Storage Performance During Bootstorm Introduction Virtual desktops are gaining popularity as a more cost effective and more easily serviceable solution. The most resource-dependent process

More information

NEC HPC-Linux-Cluster

NEC HPC-Linux-Cluster NEC HPC-Linux-Cluster Hardware configuration: 4 Front-end servers: each with SandyBridge-EP processors: 16 cores per node 128 GB memory 134 compute nodes: 112 nodes with SandyBridge-EP processors (16 cores

More information

Introduction to Supercomputing with Janus

Introduction to Supercomputing with Janus Introduction to Supercomputing with Janus Shelley Knuth shelley.knuth@colorado.edu Peter Ruprecht peter.ruprecht@colorado.edu www.rc.colorado.edu Outline Who is CU Research Computing? What is a supercomputer?

More information

Clusters: Mainstream Technology for CAE

Clusters: Mainstream Technology for CAE Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux

More information

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance 11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu

More information

Microsoft HPC. V 1.0 José M. Cámara (checam@ubu.es)

Microsoft HPC. V 1.0 José M. Cámara (checam@ubu.es) Microsoft HPC V 1.0 José M. Cámara (checam@ubu.es) Introduction Microsoft High Performance Computing Package addresses computing power from a rather different approach. It is mainly focused on commodity

More information

The Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist

The Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist The Top Six Advantages of CUDA-Ready Clusters Ian Lumb Bright Evangelist GTC Express Webinar January 21, 2015 We scientists are time-constrained, said Dr. Yamanaka. Our priority is our research, not managing

More information

Advanced Techniques with Newton. Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011

Advanced Techniques with Newton. Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011 Advanced Techniques with Newton Gerald Ragghianti Advanced Newton workshop Sept. 22, 2011 Workshop Goals Gain independence Executing your work Finding Information Fixing Problems Optimizing Effectiveness

More information

An introduction to Fyrkat

An introduction to Fyrkat Cluster Computing May 25, 2011 How to get an account https://fyrkat.grid.aau.dk/useraccount How to get help https://fyrkat.grid.aau.dk/wiki What is a Cluster Anyway It is NOT something that does any of

More information

Priority Pro v17: Hardware and Supporting Systems

Priority Pro v17: Hardware and Supporting Systems Introduction Priority Pro v17: Hardware and Supporting Systems The following provides minimal system configuration requirements for Priority with respect to three types of installations: On-premise Priority

More information

An Introduction to High Performance Computing in the Department

An Introduction to High Performance Computing in the Department An Introduction to High Performance Computing in the Department Ashley Ford & Chris Jewell Department of Statistics University of Warwick October 30, 2012 1 Some Background 2 How is Buster used? 3 Software

More information

How To Live Migrate In Hyperv On Windows Server 22 (Windows) (Windows V) (Hyperv) (Powerpoint) (For A Hyperv Virtual Machine) (Virtual Machine) And (Hyper V) Vhd (Virtual Hard Disk

How To Live Migrate In Hyperv On Windows Server 22 (Windows) (Windows V) (Hyperv) (Powerpoint) (For A Hyperv Virtual Machine) (Virtual Machine) And (Hyper V) Vhd (Virtual Hard Disk Poster Companion Reference: Hyper-V Virtual Machine Mobility Live Migration Without Shared Storage Storage Migration Live Migration with SMB Shared Storage Live Migration with Failover Clusters Copyright

More information

Overview of HPC systems and software available within

Overview of HPC systems and software available within Overview of HPC systems and software available within Overview Available HPC Systems Ba Cy-Tera Available Visualization Facilities Software Environments HPC System at Bibliotheca Alexandrina SUN cluster

More information

PARALLELS CLOUD STORAGE

PARALLELS CLOUD STORAGE PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...

More information

Streamline Computing Linux Cluster User Training. ( Nottingham University)

Streamline Computing Linux Cluster User Training. ( Nottingham University) 1 Streamline Computing Linux Cluster User Training ( Nottingham University) 3 User Training Agenda System Overview System Access Description of Cluster Environment Code Development Job Schedulers Running

More information

Contact for all enquiries Phone: +61 2 8006 9730. Email: info@recordpoint.com.au. Page 2. RecordPoint Release Notes V3.8 for SharePoint 2013

Contact for all enquiries Phone: +61 2 8006 9730. Email: info@recordpoint.com.au. Page 2. RecordPoint Release Notes V3.8 for SharePoint 2013 Release Notes V3.8 Notice This document contains confidential and trade secret information of RecordPoint Software ( RPS ). RecordPoint Software has prepared this document for use solely with RecordPoint.

More information

Building a Linux Cluster

Building a Linux Cluster Building a Linux Cluster CUG Conference May 21-25, 2001 by Cary Whitney Clwhitney@lbl.gov Outline What is PDSF and a little about its history. Growth problems and solutions. Storage Network Hardware Administration

More information

Cloud Computing through Virtualization and HPC technologies

Cloud Computing through Virtualization and HPC technologies Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC

More information

Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015

Work Environment. David Tur HPC Expert. HPC Users Training September, 18th 2015 Work Environment David Tur HPC Expert HPC Users Training September, 18th 2015 1. Atlas Cluster: Accessing and using resources 2. Software Overview 3. Job Scheduler 1. Accessing Resources DIPC technicians

More information

Deploying and Optimizing SQL Server for Virtual Machines

Deploying and Optimizing SQL Server for Virtual Machines Deploying and Optimizing SQL Server for Virtual Machines Deploying and Optimizing SQL Server for Virtual Machines Much has been written over the years regarding best practices for deploying Microsoft SQL

More information

AssetTrack. Overview: AMI AssetTrack Integration for HP Asset Manager. Mobile Forms/Scanners. Web Forms CMYK: 62 42 73 25

AssetTrack. Overview: AMI AssetTrack Integration for HP Asset Manager. Mobile Forms/Scanners. Web Forms CMYK: 62 42 73 25 Overview: AMI AssetTrack Integration for HP Asset Manager AssetTrack is an enterprise barcode asset tracking solution for HP Asset Manager that enables Asset Manager customers to use handheld barcode scanners

More information

Quantcast Petabyte Storage at Half Price with QFS!

Quantcast Petabyte Storage at Half Price with QFS! 9-131 Quantcast Petabyte Storage at Half Price with QFS Presented by Silvius Rus, Director, Big Data Platforms September 2013 Quantcast File System (QFS) A high performance alternative to the Hadoop Distributed

More information

LCMON Network Traffic Analysis

LCMON Network Traffic Analysis LCMON Network Traffic Analysis Adam Black Centre for Advanced Internet Architectures, Technical Report 79A Swinburne University of Technology Melbourne, Australia adamblack@swin.edu.au Abstract The Swinburne

More information

Avoiding Performance Bottlenecks in Hyper-V

Avoiding Performance Bottlenecks in Hyper-V Avoiding Performance Bottlenecks in Hyper-V Identify and eliminate capacity related performance bottlenecks in Hyper-V while placing new VMs for optimal density and performance Whitepaper by Chris Chesley

More information

GC3: Grid Computing Competence Center Cluster computing, I Batch-queueing systems

GC3: Grid Computing Competence Center Cluster computing, I Batch-queueing systems GC3: Grid Computing Competence Center Cluster computing, I Batch-queueing systems Riccardo Murri, Sergio Maffioletti Grid Computing Competence Center, Organisch-Chemisches Institut, University of Zurich

More information

International Engineering Journal For Research & Development

International Engineering Journal For Research & Development Evolution Of Operating System And Open Source Android Application Nilesh T.Gole 1, Amit Manikrao 2, Niraj Kanot 3,Mohan Pande 4 1,M.tech(CSE)JNTU, 2 M.tech(CSE)SGBAU, 3 M.tech(CSE),JNTU, Hyderabad 1 sheyanilu@gmail.com,

More information

Parallel Computing with MATLAB

Parallel Computing with MATLAB Parallel Computing with MATLAB Scott Benway Senior Account Manager Jiro Doke, Ph.D. Senior Application Engineer 2013 The MathWorks, Inc. 1 Acceleration Strategies Applied in MATLAB Approach Options Best

More information

Using Parallel Computing to Run Multiple Jobs

Using Parallel Computing to Run Multiple Jobs Beowulf Training Using Parallel Computing to Run Multiple Jobs Jeff Linderoth August 5, 2003 August 5, 2003 Beowulf Training Running Multiple Jobs Slide 1 Outline Introduction to Scheduling Software The

More information

Certification: HP ATA Servers & Storage

Certification: HP ATA Servers & Storage HP ExpertONE Competency Model Certification: HP ATA Servers & Storage Overview Achieving an HP certification provides relevant skills that can lead to a fulfilling career in Information Technology. HP

More information

Optimizing Shared Resource Contention in HPC Clusters

Optimizing Shared Resource Contention in HPC Clusters Optimizing Shared Resource Contention in HPC Clusters Sergey Blagodurov Simon Fraser University Alexandra Fedorova Simon Fraser University Abstract Contention for shared resources in HPC clusters occurs

More information

Lecture 1: the anatomy of a supercomputer

Lecture 1: the anatomy of a supercomputer Where a calculator on the ENIAC is equipped with 18,000 vacuum tubes and weighs 30 tons, computers of the future may have only 1,000 vacuum tubes and perhaps weigh 1½ tons. Popular Mechanics, March 1949

More information

Improving Compute Farm Efficiency for EDA

Improving Compute Farm Efficiency for EDA Improving Compute Farm Efficiency for EDA Many IT managers report that the average utilization of their compute farms is just 50-60%. Neel Desai, product marketing manager, Lynx Design System, explains

More information

Configuration Maximums VMware vsphere 4.0

Configuration Maximums VMware vsphere 4.0 Topic Configuration s VMware vsphere 4.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.0. The limits presented in the

More information

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing An Alternative Storage Solution for MapReduce Eric Lomascolo Director, Solutions Marketing MapReduce Breaks the Problem Down Data Analysis Distributes processing work (Map) across compute nodes and accumulates

More information

FLOW-3D Performance Benchmark and Profiling. September 2012

FLOW-3D Performance Benchmark and Profiling. September 2012 FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute

More information

Getting Started with HPC

Getting Started with HPC Getting Started with HPC An Introduction to the Minerva High Performance Computing Resource 17 Sep 2013 Outline of Topics Introduction HPC Accounts Logging onto the HPC Clusters Common Linux Commands Storage

More information

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014

Using WestGrid. Patrick Mann, Manager, Technical Operations Jan.15, 2014 Using WestGrid Patrick Mann, Manager, Technical Operations Jan.15, 2014 Winter 2014 Seminar Series Date Speaker Topic 5 February Gino DiLabio Molecular Modelling Using HPC and Gaussian 26 February Jonathan

More information

Configuration Maximums VMware vsphere 4.1

Configuration Maximums VMware vsphere 4.1 Topic Configuration s VMware vsphere 4.1 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.1. The limits presented in the

More information

Using the Windows Cluster

Using the Windows Cluster Using the Windows Cluster Christian Terboven terboven@rz.rwth aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster

More information

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007

More information

Deploying Exchange Server 2007 SP1 on Windows Server 2008

Deploying Exchange Server 2007 SP1 on Windows Server 2008 Deploying Exchange Server 2007 SP1 on Windows Server 2008 Product Group - Enterprise Dell White Paper By Ananda Sankaran Andrew Bachler April 2008 Contents Introduction... 3 Deployment Considerations...

More information

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance. Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance

More information

System Software for High Performance Computing. Joe Izraelevitz

System Software for High Performance Computing. Joe Izraelevitz System Software for High Performance Computing Joe Izraelevitz Agenda Overview of Supercomputers Blue Gene/Q System LoadLeveler Job Scheduler General Parallel File System HPC at UR What is a Supercomputer?

More information

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance

More information

Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database

Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database WHITE PAPER Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive

More information

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center SRNWP Workshop HP Solutions and Activities in Climate & Weather Research Michael Riedmann European Performance Center Agenda A bit of marketing: HP Solutions for HPC A few words about recent Met deals

More information

Solution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware

Solution Brief Availability and Recovery Options: Microsoft Exchange Solutions on VMware Introduction By leveraging the inherent benefits of a virtualization based platform, a Microsoft Exchange Server 2007 deployment on VMware Infrastructure 3 offers a variety of availability and recovery

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described

More information

EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi

EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Celerra Unified Storage Platforms Using iscsi EMC Business Continuity for Microsoft SQL Server Enabled by SQL DB Mirroring Applied Technology Abstract Microsoft SQL Server includes a powerful capability to protect active databases by using either

More information

Energy Constrained Resource Scheduling for Cloud Environment

Energy Constrained Resource Scheduling for Cloud Environment Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering

More information

Recommendations for Performance Benchmarking

Recommendations for Performance Benchmarking Recommendations for Performance Benchmarking Shikhar Puri Abstract Performance benchmarking of applications is increasingly becoming essential before deployment. This paper covers recommendations and best

More information

Cloud Management: Knowing is Half The Battle

Cloud Management: Knowing is Half The Battle Cloud Management: Knowing is Half The Battle Raouf BOUTABA David R. Cheriton School of Computer Science University of Waterloo Joint work with Qi Zhang, Faten Zhani (University of Waterloo) and Joseph

More information

Parallel Programming for Multi-Core, Distributed Systems, and GPUs Exercises

Parallel Programming for Multi-Core, Distributed Systems, and GPUs Exercises Parallel Programming for Multi-Core, Distributed Systems, and GPUs Exercises Pierre-Yves Taunay Research Computing and Cyberinfrastructure 224A Computer Building The Pennsylvania State University University

More information

PBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007

PBS Tutorial. Fangrui Ma Universit of Nebraska-Lincoln. October 26th, 2007 PBS Tutorial Fangrui Ma Universit of Nebraska-Lincoln October 26th, 2007 Abstract In this tutorial we gave a brief introduction to using PBS Pro. We gave examples on how to write control script, and submit

More information

Big Data and Cloud Computing for GHRSST

Big Data and Cloud Computing for GHRSST Big Data and Cloud Computing for GHRSST Jean-Francois Piollé (jfpiolle@ifremer.fr) Frédéric Paul, Olivier Archer CERSAT / Institut Français de Recherche pour l Exploitation de la Mer Facing data deluge

More information

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What

More information

Data processing goes big

Data processing goes big Test report: Integration Big Data Edition Data processing goes big Dr. Götz Güttich Integration is a powerful set of tools to access, transform, move and synchronize data. With more than 450 connectors,

More information

MPI / ClusterTools Update and Plans

MPI / ClusterTools Update and Plans HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski

More information

Chapter 2: OS Overview

Chapter 2: OS Overview Chapter 2: OS Overview CmSc 335 Operating Systems 1. Operating system objectives and functions Operating systems control and support the usage of computer systems. a. usage users of a computer system:

More information

x64 Servers: Do you want 64 or 32 bit apps with that server?

x64 Servers: Do you want 64 or 32 bit apps with that server? TMurgent Technologies x64 Servers: Do you want 64 or 32 bit apps with that server? White Paper by Tim Mangan TMurgent Technologies February, 2006 Introduction New servers based on what is generally called

More information

High Availability of the Polarion Server

High Availability of the Polarion Server Polarion Software CONCEPT High Availability of the Polarion Server Installing Polarion in a high availability environment Europe, Middle-East, Africa: Polarion Software GmbH Hedelfinger Straße 60 70327

More information

Performance Optimization Guide

Performance Optimization Guide Performance Optimization Guide Publication Date: July 06, 2016 Copyright Metalogix International GmbH, 2001-2016. All Rights Reserved. This software is protected by copyright law and international treaties.

More information

Cisco Unified Computing System and EMC VNX5300 Unified Storage Platform

Cisco Unified Computing System and EMC VNX5300 Unified Storage Platform Cisco Unified Computing System and EMC VNX5300 Unified Storage Platform Implementing an Oracle Data Warehouse Test Workload White Paper January 2011, Revision 1.0 Contents Executive Summary... 3 Cisco

More information

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 This document is provided as-is. Information and views expressed in this document, including URL and other Internet

More information

Step by Step Guide To vstorage Backup Server (Proxy) Sizing

Step by Step Guide To vstorage Backup Server (Proxy) Sizing Tivoli Storage Manager for Virtual Environments V6.3 Step by Step Guide To vstorage Backup Server (Proxy) Sizing 12 September 2012 1.1 Author: Dan Wolfe, Tivoli Software Advanced Technology Page 1 of 18

More information

Setup for Failover Clustering and Microsoft Cluster Service

Setup for Failover Clustering and Microsoft Cluster Service Setup for Failover Clustering and Microsoft Cluster Service Update 1 ESX 4.0 ESXi 4.0 vcenter Server 4.0 This document supports the version of each product listed and supports all subsequent versions until

More information

HP high availability solutions for Microsoft SQL Server Fast Track Data Warehouse using SQL Server 2012 failover clustering

HP high availability solutions for Microsoft SQL Server Fast Track Data Warehouse using SQL Server 2012 failover clustering Technical white paper HP high availability solutions for Microsoft SQL Server Fast Track Data Warehouse using SQL Server 2012 failover clustering Table of contents Executive summary 2 Fast Track reference

More information

Benchmarking Hadoop & HBase on Violin

Benchmarking Hadoop & HBase on Violin Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages

More information