Kashif Iqbal - PhD [email protected]
|
|
|
- Diane Charles
- 10 years ago
- Views:
Transcription
1 HPC/HTC vs. Cloud Benchmarking An empirical evalua.on of the performance and cost implica.ons Kashif Iqbal - PhD [email protected] ICHEC, NUI Galway, Ireland With acknowledgment to Michele MicheloDo INFN for their kind support for the HTC Benchmarking + Resources
2 Outline Benchmarking - Why, which benchmark? HPC and HTC Benchmarking Benchmarks (NPB, HEPSPEC06) Environment Setup Results Concluding Remarks 2
3 Overview RaYonale Diverse compuyng infrastructures (HPC. HTC, Cloud) Diverse workloads (SMEs, academics, app. Domains) Methodology System benchmarking for: Comparison of HPC and HTC systems vs. Cloud offerings Comparison of parallelism techniques (e.g. MPI/OMP) ***To calculate the performance variayon factor*** For Cost analysis and infrastructures comparison Performance overhead as indirect costs Inclusion of an addi-onal weight factor in the cost model 3
4 Scope What we are not doing? Benchmarking individual components (e.g. memory, network or I/O bandwidth) Why? Because resources cost money Aim is to compare HPC, HTC and Cloud infrastructures to idenyfy ranges for performance variayons 4
5 Benchmarks LINPACK Top 500 SPEC06 CPU intensive benchmark HEP- SPEC06 (for HTC vs. Cloud instances) HPC Challenge (HPCC) Linpack, DGEMM, STREAM, FFT Graph 500 MPPtest MPI performance NAS Parallel Benchmark (NPB) (for HPC vs. Cloud HPC instances) 5
6 NAS Parallel Benchmark Open- source and free CFD benchmark Performance evaluayon of commonly used parallelism techniques Serial, MPI, OpenMP, OpenMP+MPI, Java, HPF Customisable for different problem sizes Classes S: small for quick tests Class W: workstayon size Classes A, B, C: standard test problems Classes D, E, F: large test problems Performance metric Kernel execuyon Yme (in sec) 6
7 NPB Kernels Kernel Descrip-on Problem Size Memory (Mw) EP MG CG FT Monte Carlo kernel to compute the soluyon of an integral Embarrassingly parallel MulY- grid kernel to compute the soluyon of the 3D Poisson equayon Kernel to compute the smallest eigenvalue of a symmetric posiyve definite matrix Kernel to solve a 3D paryal difference equayon using an FFT based method 2 30 random- num pairs grid size no. of rows 512x256x256 grid size IS Parallel sort kernel based on bucket sort 2 25 no. of keys 114 LU SP ComputaYonal Fluid Dynamics (CFD) applicayon using symmetric successive over relaxayon CFD applicayon using the Beam- Warming approximate factorisayon method grid size grid size 22 BT CFD applicayon using an implicit soluyon method grid size 96 7
8 Cloud Cluster Setup EC2 instance management StarCluster Toolkit hdp://web.mit.edu/star/cluster/ StarCluster AMIs Amazon Machine Image Resource manager plugin Login vs. compute instances EC2 small instance as login node File system shared via NFS across nodes 8
9 Cloud vs. HPC Compute Node Amazon EC2 23 GB of memory, 2 x Intel Xeon X5570, quad- core Nehalem (8 cores X 4 Nodes) Stokes HPC 24 GB memory, 2 x Intel Xeon E5650, hex- core Westmere (12 cores X 3 Nodes) Connec-vity 10 Gigabit Ethernet ConnectX Infiniband (DDR) OS Ubuntu, 64- bit planorm Open- SUSE, 64- bit planorm Resource manager Sun Grid Engine Torque Compilers & libraries Intel C, Intel Fortran, Intel MKL, Intel MVAPICH2 Non- trivial to replicate runyme environments Large variayons in performance possible Intel C, Intel Fortran, Intel MKL, Intel MVAPICH2 Logical vs. Physical cores HT/SMT Hyper or Simultaneous MulY- Threading (i.e. 2 X Physical Cores) 9
10 NPB MPI 90 BT and SP using 16 cores, rest using 32 cores (22 runs) Time in Seconds Stokes HPC Amazon EC BT CG EP FT IS LU MG SP NPB3.3- MPI - Class B The average performance loss ~ 48.42% (ranging from 1.02% to 67.76%). 10
11 NPB - OpenMP 8 cores with 8 OMP Threads (22 runs) 160 Time in Seconds Stokes EC BT CG EP FT IS LU MG SP UA NPB3.3- OMP - Class B The average performance loss ~ 37.26% (ranging from % 11
12 Cost USD J ~100 % uylisayon Compute cluster $1.300 per Hour Small $0.080 per Hour Other useful facts: On- Demand instances Overheads (performance, I/O, setup) Data transfer costs and Yme 12
13 HEPSPEC Benchmark HEP Benchmark to measure CPU performance Based on all_cpp bset of SPEC CPU2006 Fair distribuyon of SPECint and SPECfp Real workload Performance Metric: HEPSPEC Score 32- bit binaries Can be compiled using 64- bit mode ~ for beder results 13
14 Benchmark Environment Compute Nodes Amazon EC2 Medium: 2 ECU Large: 4 ECU XL: 8 ECU M3 XL: 13 ECU M3 2 XL: 26 ECU 1 ECU = GHz HTC resource at INFN Intel(R) Xeon(R) CPU 2.2 GHz, 2 X 8 cores, 64 GB memory AMD Opteron 6272 (aka 2.1 GHz, 2 X 16 cores M instance Single- core VM L instance Dual- core VM XL Instance Quad- core VM M3 XL Instance Quad- core VM M3 Double XL Instance Eight Core VM OS SL6.2, 64- bit planorm SL6.2, 64- bit planorm Memory 3.75 GB, 7.5 GB and 15 GB 64 GB for both Intel and AMD Hyper- Threading Enabled Enabled (for Intel) ~ 32 logical cores Compilers GCC GCC 14
15 Benchmark VariaYons Bare- metal vs. EC2 imply: *Cloud migrayon performance implicayons* 1 VM vs. Bare- metal: *VirtualisaYon effect* n VMs + 1 loaded (HS06) vs. Bare- metal: *VirtualisaYon + muly- tenancy effect with minimal workload* - Best Case Scenario! n VMs + n loaded *VirtualisaYon + muly- tenancy effect with full workload* - Worst Case Scenario! 15
16 HS06 for Medium HS06 Score No. of VMs = 32 Bare Metal 1 VM n VMs + n VMs + Fully Minimal load loaded SPEC score < with the > no. of VMs VirtualisaYon + MulY- Tenancy (MT) effect on performance ~ 3.28% to 58.48% 47.95% Performance loss for Cloud migrayon HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current 1 VM + idle N VMs + minimal load Future N VMs + Fully loaded Amazon EC2 - Possible varia4ons 1 VM + idle Unlikely in EC2 N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! 16
17 HS06 for Medium HS06 Score No. of VMs = 32 Bare Metal 1 VM n VMs + n VMs + Fully Minimal load loaded HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current 1 VM + idle N VMs + minimal load Future N VMs + Fully loaded Amazon EC2 - Possible varia4ons 1 VM + idle Unlikely in EC2 N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! VirtualisaYon + MT effect on performance ~ 3.77% to 47.89% Performance loss for Cloud migrayon 17
18 HS06 for Large HS06 Score No. of VMs = 16 Bare Metal 1 VM n VMs + n VMs + Fully Minimal load loaded HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current 1 VM + idle N VMs + minimal load Future N VMs + Fully loaded Amazon EC2 - Possible varia4ons 1 VM + idle Unlikely in EC2 N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! VirtualisaYon + MT effect on performance ~ 9.49% to 57.47% Note the minimal effect on performance with > no. of VMs 58.79% Performance loss for Cloud migrayon 18
19 HS06 for Large HS06 Score No. of VMs = 16 Bare Metal 1 VM n VMs + n VMs + Fully Minimal load loaded VirtualisaYon + MT effect on performance ~ 9.04% to 48.88% 30.75% Performance loss for Cloud migrayon HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current 1 VM + idle N VMs + minimal load Future N VMs + Fully loaded Amazon EC2 - Possible varia4ons 1 VM + idle Unlikely in EC2 N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! 19
20 HS06 for Xlarge HS06 Score No. of VMs = 8 Bare Metal 1 VM n VMs + n VMs + Fully Minimal load loaded HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current 1 VM + idle N VMs + minimal load Future N VMs + Fully loaded Amazon EC2 - Possible varia4ons 1 VM + idle Unlikely in EC2 N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! VirtualisaYon + MT effect on performance ~ 8.14% to 55.84% Note the minimal effect of > no. of VMs 46.21% Performance loss for Cloud migrayon 20
21 M3 X- Large No. of VMs = HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current N VMs + minimal load N VMs + Fully loaded Bare Metal VMs 4 HS06 Amazon EC2 - Possible varia4ons N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! VirtualisaYon + MT effect on performance ~ % Performance loss for Cloud migrayon 21
22 M3 Double X- Large No. of VMs = 4 Bare Metal VMs 4 HS06 HTC EC2 NGIs Possible varia4ons Bare Metal NGIs current N VMs + minimal load N VMs + Fully loaded Amazon EC2 - Possible varia4ons N VMs + minimal load Best- case! N VMs + Fully loaded Worst- case! VirtualisaYon + MT effect on performance ~ % 37.79% Performance loss for Cloud migrayon 22
23 InteresYng Finding No Overcomitment No Overcomitment Overcomitment Memory: 28GB of 64GB Memory: 48GB of 64GB Memory: 120GB of 64 GB 0 Bare Metal VMs 4 HS06 Over- commitment of resources vs. Cost effecyveness That is more reasonably why 1 ECU = GHz 2007 Xeon or Opteron!! 23
24 Conclusions HPC vs. EC2 As expected a purpose built HPC cluster outperforms EC2 cluster for same number of OMP threads Average performance loss over all NPB tests: ~37% Similarly so for when comparing 10GigE versus Infiniband networking fabrics Average performance loss over all NPB test: ~48% Even at a modest problem size the differences in performances between systems is highlighted. 24
25 Conclusions HTC vs. EC2 VirtualisaYon overhead is much less than the MulY- Tenancy effect What others are running will have a direct effect! Standard deviayon with pre- launched VMs in EC2 is significantly low! Results with M3 L instance: 46.92, 46.61, HS06 Scores variayons in the order of 40-48% 25
26 Thank you for your adenyon! QuesYons??
benchmarking Amazon EC2 for high-performance scientific computing
Edward Walker benchmarking Amazon EC2 for high-performance scientific computing Edward Walker is a Research Scientist with the Texas Advanced Computing Center at the University of Texas at Austin. He received
Multicore Parallel Computing with OpenMP
Multicore Parallel Computing with OpenMP Tan Chee Chiang (SVU/Academic Computing, Computer Centre) 1. OpenMP Programming The death of OpenMP was anticipated when cluster systems rapidly replaced large
Performance Evaluation of Amazon EC2 for NASA HPC Applications!
National Aeronautics and Space Administration Performance Evaluation of Amazon EC2 for NASA HPC Applications! Piyush Mehrotra!! J. Djomehri, S. Heistand, R. Hood, H. Jin, A. Lazanoff,! S. Saini, R. Biswas!
Cloud Computing through Virtualization and HPC technologies
Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC
Toward a practical HPC Cloud : Performance tuning of a virtualized HPC cluster
Toward a practical HPC Cloud : Performance tuning of a virtualized HPC cluster Ryousei Takano Information Technology Research Institute, National Institute of Advanced Industrial Science and Technology
Workshop on Parallel and Distributed Scientific and Engineering Computing, Shanghai, 25 May 2012
Scientific Application Performance on HPC, Private and Public Cloud Resources: A Case Study Using Climate, Cardiac Model Codes and the NPB Benchmark Suite Peter Strazdins (Research School of Computer Science),
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi ICPP 6 th International Workshop on Parallel Programming Models and Systems Software for High-End Computing October 1, 2013 Lyon, France
RDMA over Ethernet - A Preliminary Study
RDMA over Ethernet - A Preliminary Study Hari Subramoni, Miao Luo, Ping Lai and Dhabaleswar. K. Panda Computer Science & Engineering Department The Ohio State University Outline Introduction Problem Statement
SR-IOV In High Performance Computing
SR-IOV In High Performance Computing Hoot Thompson & Dan Duffy NASA Goddard Space Flight Center Greenbelt, MD 20771 [email protected] [email protected] www.nccs.nasa.gov Focus on the research side
How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications
Amazon Cloud Performance Compared David Adams Amazon EC2 performance comparison How does EC2 compare to traditional supercomputer for scientific applications? "Performance Analysis of High Performance
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 Agenda v Mapping clients needs to cloud technologies v Addressing your pain
Can High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
Enabling Technologies for Distributed and Cloud Computing
Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Multi-core CPUs and Multithreading
Amazon EC2 XenApp Scalability Analysis
WHITE PAPER Citrix XenApp Amazon EC2 XenApp Scalability Analysis www.citrix.com Table of Contents Introduction...3 Results Summary...3 Detailed Results...4 Methods of Determining Results...4 Amazon EC2
Parallel Programming Survey
Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory
Cloud Computing. Alex Crawford Ben Johnstone
Cloud Computing Alex Crawford Ben Johnstone Overview What is cloud computing? Amazon EC2 Performance Conclusions What is the Cloud? A large cluster of machines o Economies of scale [1] Customers use a
SR-IOV: Performance Benefits for Virtualized Interconnects!
SR-IOV: Performance Benefits for Virtualized Interconnects! Glenn K. Lockwood! Mahidhar Tatineni! Rick Wagner!! July 15, XSEDE14, Atlanta! Background! High Performance Computing (HPC) reaching beyond traditional
StACC: St Andrews Cloud Computing Co laboratory. A Performance Comparison of Clouds. Amazon EC2 and Ubuntu Enterprise Cloud
StACC: St Andrews Cloud Computing Co laboratory A Performance Comparison of Clouds Amazon EC2 and Ubuntu Enterprise Cloud Jonathan S Ward StACC (pronounced like 'stack') is a research collaboration launched
Running Scientific Codes on Amazon EC2: a Performance Analysis of Five High-end Instances
JCS&T Vol. 13 No. 3 December 213 Running Scientific Codes on Amazon EC2: a Performance Analysis of Five High-end Instances Roberto R. Expósito, Guillermo L. Taboada, Xoán C. Pardo, Juan Touriño and Ramón
Introduction to High Performance Cluster Computing. Cluster Training for UCL Part 1
Introduction to High Performance Cluster Computing Cluster Training for UCL Part 1 What is HPC HPC = High Performance Computing Includes Supercomputing HPCC = High Performance Cluster Computing Note: these
Enabling Technologies for Distributed Computing
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies
A quick tutorial on Intel's Xeon Phi Coprocessor
A quick tutorial on Intel's Xeon Phi Coprocessor www.cism.ucl.ac.be [email protected] Architecture Setup Programming The beginning of wisdom is the definition of terms. * Name Is a... As opposed
A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures
11 th International LS-DYNA Users Conference Computing Technology A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures Yih-Yih Lin Hewlett-Packard Company Abstract In this paper, the
1 Bull, 2011 Bull Extreme Computing
1 Bull, 2011 Bull Extreme Computing Table of Contents HPC Overview. Cluster Overview. FLOPS. 2 Bull, 2011 Bull Extreme Computing HPC Overview Ares, Gerardo, HPC Team HPC concepts HPC: High Performance
Performance and Scalability of the NAS Parallel Benchmarks in Java
Performance and Scalability of the NAS Parallel Benchmarks in Java Michael A. Frumkin, Matthew Schultz, Haoqiang Jin, and Jerry Yan NASA Advanced Supercomputing (NAS) Division NASA Ames Research Center,
PERFORMANCE CONSIDERATIONS FOR NETWORK SWITCH FABRICS ON LINUX CLUSTERS
PERFORMANCE CONSIDERATIONS FOR NETWORK SWITCH FABRICS ON LINUX CLUSTERS Philip J. Sokolowski Department of Electrical and Computer Engineering Wayne State University 55 Anthony Wayne Dr. Detroit, MI 822
FLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
OpenACC Parallelization and Optimization of NAS Parallel Benchmarks
OpenACC Parallelization and Optimization of NAS Parallel Benchmarks Presented by Rengan Xu GTC 2014, S4340 03/26/2014 Rengan Xu, Xiaonan Tian, Sunita Chandrasekaran, Yonghong Yan, Barbara Chapman HPC Tools
Neptune. A Domain Specific Language for Deploying HPC Software on Cloud Platforms. Chris Bunch Navraj Chohan Chandra Krintz Khawaja Shams
Neptune A Domain Specific Language for Deploying HPC Software on Cloud Platforms Chris Bunch Navraj Chohan Chandra Krintz Khawaja Shams ScienceCloud 2011 @ San Jose, CA June 8, 2011 Cloud Computing Three
DELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering
DELL Virtual Desktop Infrastructure Study END-TO-END COMPUTING Dell Enterprise Solutions Engineering 1 THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL
1 DCSC/AU: HUGE. DeIC Sekretariat 2013-03-12/RB. Bilag 1. DeIC (DCSC) Scientific Computing Installations
Bilag 1 2013-03-12/RB DeIC (DCSC) Scientific Computing Installations DeIC, previously DCSC, currently has a number of scientific computing installations, distributed at five regional operating centres.
Assessing the Performance of OpenMP Programs on the Intel Xeon Phi
Assessing the Performance of OpenMP Programs on the Intel Xeon Phi Dirk Schmidl, Tim Cramer, Sandra Wienke, Christian Terboven, and Matthias S. Müller [email protected] Rechen- und Kommunikationszentrum
High Performance Computing in CST STUDIO SUITE
High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver
System Requirements Table of contents
Table of contents 1 Introduction... 2 2 Knoa Agent... 2 2.1 System Requirements...2 2.2 Environment Requirements...4 3 Knoa Server Architecture...4 3.1 Knoa Server Components... 4 3.2 Server Hardware Setup...5
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING José Daniel García Sánchez ARCOS Group University Carlos III of Madrid Contents 2 The ARCOS Group. Expand motivation. Expand
Improved LS-DYNA Performance on Sun Servers
8 th International LS-DYNA Users Conference Computing / Code Tech (2) Improved LS-DYNA Performance on Sun Servers Youn-Seo Roh, Ph.D. And Henry H. Fong Sun Microsystems, Inc. Abstract Current Sun platforms
Maximize Performance and Scalability of RADIOSS* Structural Analysis Software on Intel Xeon Processor E7 v2 Family-Based Platforms
Maximize Performance and Scalability of RADIOSS* Structural Analysis Software on Family-Based Platforms Executive Summary Complex simulations of structural and systems performance, such as car crash simulations,
Using the Windows Cluster
Using the Windows Cluster Christian Terboven [email protected] aachen.de Center for Computing and Communication RWTH Aachen University Windows HPC 2008 (II) September 17, RWTH Aachen Agenda o Windows Cluster
Scaling in a Hypervisor Environment
Scaling in a Hypervisor Environment Richard McDougall Chief Performance Architect VMware VMware ESX Hypervisor Architecture Guest Monitor Guest TCP/IP Monitor (BT, HW, PV) File System CPU is controlled
1000Mbps Ethernet Performance Test Report 2014.4
1000Mbps Ethernet Performance Test Report 2014.4 Test Setup: Test Equipment Used: Lenovo ThinkPad T420 Laptop Intel Core i5-2540m CPU - 2.60 GHz 4GB DDR3 Memory Intel 82579LM Gigabit Ethernet Adapter CentOS
Benchmark Hadoop and Mars: MapReduce on cluster versus on GPU
Benchmark Hadoop and Mars: MapReduce on cluster versus on GPU Heshan Li, Shaopeng Wang The Johns Hopkins University 3400 N. Charles Street Baltimore, Maryland 21218 {heshanli, shaopeng}@cs.jhu.edu 1 Overview
Scalability evaluation of barrier algorithms for OpenMP
Scalability evaluation of barrier algorithms for OpenMP Ramachandra Nanjegowda, Oscar Hernandez, Barbara Chapman and Haoqiang H. Jin High Performance Computing and Tools Group (HPCTools) Computer Science
LS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
International Journal of Computer & Organization Trends Volume20 Number1 May 2015
Performance Analysis of Various Guest Operating Systems on Ubuntu 14.04 Prof. (Dr.) Viabhakar Pathak 1, Pramod Kumar Ram 2 1 Computer Science and Engineering, Arya College of Engineering, Jaipur, India.
Parallel Algorithm Engineering
Parallel Algorithm Engineering Kenneth S. Bøgh PhD Fellow Based on slides by Darius Sidlauskas Outline Background Current multicore architectures UMA vs NUMA The openmp framework Examples Software crisis
GPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
Integration of Virtualized Workernodes in Batch Queueing Systems The ViBatch Concept
Integration of Virtualized Workernodes in Batch Queueing Systems, Dr. Armin Scheurer, Oliver Oberst, Prof. Günter Quast INSTITUT FÜR EXPERIMENTELLE KERNPHYSIK FAKULTÄT FÜR PHYSIK KIT University of the
GRIDCENTRIC VMS TECHNOLOGY VDI PERFORMANCE STUDY
GRIDCENTRIC VMS TECHNOLOGY VDI PERFORMANCE STUDY TECHNICAL WHITE PAPER MAY 1 ST, 2012 GRIDCENTRIC S VIRTUAL MEMORY STREAMING (VMS) TECHNOLOGY SIGNIFICANTLY IMPROVES THE COST OF THE CLASSIC VIRTUAL MACHINE
Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software
WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications
Benchmarking Amazon s EC2 Cloud Platform
Benchmarking Amazon s EC2 Cloud Platform Goal of the project: The goal of this project is to analyze the performance of an 8-node cluster in Amazon s Elastic Compute Cloud (EC2). The cluster will be benchmarked
The CNMS Computer Cluster
The CNMS Computer Cluster This page describes the CNMS Computational Cluster, how to access it, and how to use it. Introduction (2014) The latest block of the CNMS Cluster (2010) Previous blocks of the
Building a Private Cloud with Eucalyptus
Building a Private Cloud with Eucalyptus 5th IEEE International Conference on e-science Oxford December 9th 2009 Christian Baun, Marcel Kunze KIT The cooperation of Forschungszentrum Karlsruhe GmbH und
Performance Comparison of ISV Simulation Codes on Microsoft Windows HPC Server 2008 and SUSE Linux Enterprise Server 10.2
Fraunhofer Institute for Algorithms and Scientific Computing SCAI Performance Comparison of ISV Simulation Codes on Microsoft HPC Server 28 and SUSE Enterprise Server 1.2 Karsten Reineck und Horst Schwichtenberg
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
NAVAL POSTGRADUATE SCHOOL THESIS
NAVAL POSTGRADUATE SCHOOL MONTEREY, CALIFORNIA THESIS FEASIBILITY OF VIRTUAL MACHINE AND CLOUD COMPUTING TECHNOLOGIES FOR HIGH PERFORMANCE COMPUTING by Richard Chad Hutchins December 2013 Thesis Co-Advisors:
ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009
ECLIPSE Best Practices Performance, Productivity, Efficiency March 29 ECLIPSE Performance, Productivity, Efficiency The following research was performed under the HPC Advisory Council activities HPC Advisory
Benchmarking Large Scale Cloud Computing in Asia Pacific
2013 19th IEEE International Conference on Parallel and Distributed Systems ing Large Scale Cloud Computing in Asia Pacific Amalina Mohamad Sabri 1, Suresh Reuben Balakrishnan 1, Sun Veer Moolye 1, Chung
WHITE PAPER FUJITSU PRIMERGY SERVERS PERFORMANCE REPORT PRIMERGY BX620 S6
WHITE PAPER PERFORMANCE REPORT PRIMERGY BX620 S6 WHITE PAPER FUJITSU PRIMERGY SERVERS PERFORMANCE REPORT PRIMERGY BX620 S6 This document contains a summary of the benchmarks executed for the PRIMERGY BX620
- An Essential Building Block for Stable and Reliable Compute Clusters
Ferdinand Geier ParTec Cluster Competence Center GmbH, V. 1.4, March 2005 Cluster Middleware - An Essential Building Block for Stable and Reliable Compute Clusters Contents: Compute Clusters a Real Alternative
How To Test Cloud Stack On A Microsoft Powerbook 2.5 (Amd64) On A Linux Computer (Amd86) On An Ubuntu) Or Windows Xp (Amd66) On Windows Xp 2.2.2 (Amd65
Function and Performance Test of pen Source CloudStack Platform for HPC Service Function and Performance Test of pen Source CloudStack Platform for HPC Service 1 Jaegyoon Hahm, 2 Sang Boem Lim, 3 Guohua
OpenMP Programming on ScaleMP
OpenMP Programming on ScaleMP Dirk Schmidl [email protected] Rechen- und Kommunikationszentrum (RZ) MPI vs. OpenMP MPI distributed address space explicit message passing typically code redesign
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
CloudHarmony Performance Benchmark: Select High-Performing Public Cloud to Increase Economic Benefits
What You Will Learn Even small differences in cloud performance can affect costs, customer satisfaction, and time to market. This white paper, intended for enterprises evaluating public cloud services,
An Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics
An Oracle White Paper November 2010 Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics 1 Introduction New applications such as web searches, recommendation engines,
Performance of the NAS Parallel Benchmarks on Grid Enabled Clusters
Performance of the NAS Parallel Benchmarks on Grid Enabled Clusters Philip J. Sokolowski Dept. of Electrical and Computer Engineering Wayne State University 55 Anthony Wayne Dr., Detroit, MI 4822 [email protected]
A Performance and Cost Analysis of the Amazon Elastic Compute Cloud (EC2) Cluster Compute Instance
A Performance and Cost Analysis of the Amazon Elastic Compute Cloud (EC2) Cluster Compute Instance Michael Fenn ([email protected]), Jason Holmes ([email protected]), Jeffrey Nucciarone ([email protected]) Research
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
CPU Benchmarks Over 600,000 CPUs Benchmarked
Shopping cart Search Home Software Hardware Benchmarks Services Store Support Forums About Us Home» CPU Benchmarks» Multiple CPU Systems CPU Benchmarks Video Card Benchmarks Hard Drive Benchmarks RAM PC
PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN
1 PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN Introduction What is cluster computing? Classification of Cluster Computing Technologies: Beowulf cluster Construction
A Micro-benchmark Suite for Evaluating Hadoop RPC on High-Performance Networks
A Micro-benchmark Suite for Evaluating Hadoop RPC on High-Performance Networks Xiaoyi Lu, Md. Wasi- ur- Rahman, Nusrat Islam, and Dhabaleswar K. (DK) Panda Network- Based Compu2ng Laboratory Department
MPI / ClusterTools Update and Plans
HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski
PERFORMANCE ANALYSIS OF KERNEL-BASED VIRTUAL MACHINE
PERFORMANCE ANALYSIS OF KERNEL-BASED VIRTUAL MACHINE Sudha M 1, Harish G M 2, Nandan A 3, Usha J 4 1 Department of MCA, R V College of Engineering, Bangalore : 560059, India [email protected] 2 Department
Oracle Database Reliability, Performance and scalability on Intel Xeon platforms Mitch Shults, Intel Corporation October 2011
Oracle Database Reliability, Performance and scalability on Intel platforms Mitch Shults, Intel Corporation October 2011 1 Intel Processor E7-8800/4800/2800 Product Families Up to 10 s and 20 Threads 30MB
ACANO SOLUTION VIRTUALIZED DEPLOYMENTS. White Paper. Simon Evans, Acano Chief Scientist
ACANO SOLUTION VIRTUALIZED DEPLOYMENTS White Paper Simon Evans, Acano Chief Scientist Updated April 2015 CONTENTS Introduction... 3 Host Requirements... 5 Sizing a VM... 6 Call Bridge VM... 7 Acano Edge
PARALLELS CLOUD SERVER
PARALLELS CLOUD SERVER Performance and Scalability 1 Table of Contents Executive Summary... Error! Bookmark not defined. LAMP Stack Performance Evaluation... Error! Bookmark not defined. Background...
An Oracle White Paper March 2013. Load Testing Best Practices for Oracle E- Business Suite using Oracle Application Testing Suite
An Oracle White Paper March 2013 Load Testing Best Practices for Oracle E- Business Suite using Oracle Application Testing Suite Executive Overview... 1 Introduction... 1 Oracle Load Testing Setup... 2
Virtualization for Cloud Computing
Virtualization for Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF CLOUD COMPUTING On demand provision of computational resources
DISTRIBUTED SYSTEMS [COMP9243] Lecture 9a: Cloud Computing WHAT IS CLOUD COMPUTING? 2
DISTRIBUTED SYSTEMS [COMP9243] Lecture 9a: Cloud Computing Slide 1 Slide 3 A style of computing in which dynamically scalable and often virtualized resources are provided as a service over the Internet.
Comparing the performance of the Landmark Nexus reservoir simulator on HP servers
WHITE PAPER Comparing the performance of the Landmark Nexus reservoir simulator on HP servers Landmark Software & Services SOFTWARE AND ASSET SOLUTIONS Comparing the performance of the Landmark Nexus
The Green Index: A Metric for Evaluating System-Wide Energy Efficiency in HPC Systems
202 IEEE 202 26th IEEE International 26th International Parallel Parallel and Distributed and Distributed Processing Processing Symposium Symposium Workshops Workshops & PhD Forum The Green Index: A Metric
GUEST OPERATING SYSTEM BASED PERFORMANCE COMPARISON OF VMWARE AND XEN HYPERVISOR
GUEST OPERATING SYSTEM BASED PERFORMANCE COMPARISON OF VMWARE AND XEN HYPERVISOR ANKIT KUMAR, SAVITA SHIWANI 1 M. Tech Scholar, Software Engineering, Suresh Gyan Vihar University, Rajasthan, India, Email:
Java for High Performance Cloud Computing
Java for Computer Architecture Group University of A Coruña (Spain) [email protected] October 31st, 2012, Future Networks and Distributed SystemS (FUNDS), Manchester Metropolitan University, Manchester (UK)
Keys to node-level performance analysis and threading in HPC applications
Keys to node-level performance analysis and threading in HPC applications Thomas GUILLET (Intel; Exascale Computing Research) IFERC seminar, 18 March 2015 Legal Disclaimer & Optimization Notice INFORMATION
Variations in Performance and Scalability when Migrating n-tier Applications to Different Clouds
Variations in Performance and Scalability when Migrating n-tier Applications to Different Clouds Deepal Jayasinghe, Simon Malkowski, Qingyang Wang, Jack Li, Pengcheng Xiong, Calton Pu Outline Motivation
Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820
Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820 This white paper discusses the SQL server workload consolidation capabilities of Dell PowerEdge R820 using Virtualization.
MOSIX: High performance Linux farm
MOSIX: High performance Linux farm Paolo Mastroserio [[email protected]] Francesco Maria Taurino [[email protected]] Gennaro Tortone [[email protected]] Napoli Index overview on Linux farm farm
Table of Contents. P a g e 2
Solution Guide Balancing Graphics Performance, User Density & Concurrency with NVIDIA GRID Virtual GPU Technology (vgpu ) for Autodesk AutoCAD Power Users V1.0 P a g e 2 Table of Contents The GRID vgpu
