ECDF Infrastructure Refresh - Requirements Consultation Document
|
|
- Chrystal Manning
- 8 years ago
- Views:
Transcription
1 Edinburgh Compute & Data Facility - December 2014 ECDF Infrastructure Refresh - Requirements Consultation Document Introduction In order to sustain the University s central research data and computing service, Information Services will be investing in a compute service refresh this financial year. This document has been produced for circulation throughout the University's research community to help inform the requirements consultation which will take place in January Detail on providing feedback is given at the end of this document. Overview The refresh will deliver: A refresh of the (ECDF) Eddie compute cluster service. A path for migration of existing parallel computational workloads, such as MPI, from the current service to the national HPC service, Archer. A computational cloud service. Infrastructure improvements to further support Data Science, such as an implementation of the Hadoop framework. An extendable service, to allow for direct researcher capital investment. This complements the substantial investment in data storage and management services that has already been made in 2013 and Roadmap Consultation: January Procurement commences: February Equipment delivered and installed: May Equipment acceptance testing & validation: June/July Eddie refresh available: August Cloud service early adopters: from October The funding allocated is for the University's financial year, and thus the procurement aspect of the project must be completed before 31st July Deliverables Eddie Refresh The current incarnation of Eddie, the University's research computing cluster, was purchased in 2010 and The refresh of the service will deliver:
2 Performance improvements in the underlying infrastructure, such as a more modern generation of processor architecture; a faster network backbone for improved and more efficient migration and processing of large research data sets. An expandable infrastructure core, allowing for easier integration of custom hardware and bespoke research solutions. A number of options relating to the memory-per-core available to better accommodate the diverse memory requirements across the University. High-performance disk storage tightly coupled with the compute cluster to store and render data on request. Improved resource management and prioritisation of resource on the compute cluster through the gridengine job scheduler. Details of the specific technical and configuration improvements the upgrade will provide can be found in Appendix A. Parallel computational workloads The intention is to cease provision of low-latency compute node interconnects with this refresh of Eddie. Tasks that require multi-node low-latency communication will be transitioned to the national HPC service, Archer, on which the University has guaranteed institutional capacity. This transition will be confirmed during the service consultation process. Cloud services An OpenStack cloud service will be included in the project, supported by the equipment supplier. This computational cloud will open up the infrastructure to more researchers and to be accessed by different use-cases. In traditional compute cluster services, a researcher s specific software environment needs to be integrated into the cluster environment, e.g. porting a new software package onto an existing operating system and compiler environment. With a computational cloud the researcher s software environment can be encapsulated within a virtual machine in which they have complete control of the environment. This makes it simpler for a researcher to manage their own analysis environment and it allows for pre-packaged environments to be easily run on common infrastructure. For example, Biolinux is a popular Linux distribution which already has a number of common Bioinformatics and Biological data processing tools available. This refresh will deliver a service which can deploy such bespoke environments via a computational cloud, which will enhance usability and open up the service to more researchers. We aim for this service to also allow for user-driven self-provisioning of compute environments, with a simple user interface to allow for rapid deployment. Supporting Data Science The Eddie service has always had a strong focus on data services, and this will be further developed with the move to a "Data Centric Computing" service model. This will integrate the Eddie storage and the University's DataStore service, allowing for easy access to data held on DataStore from the Eddie service, and vice versa. Eddie's
3 local file systems will be tuned for performance, rather than capacity. They will effectively act as a cache of data held on the DataStore service, with the system automatically moving data to Eddie on demand, and copying all new data from Eddie back to the DataStore. This will offer many advantages, including the ability to securely share data between research groups and the ability to make data available on multiple different systems. The service refresh will also look to deploy the appropriate architectures for the emerging needs of the University s Data Science community. There is existing use of the Hadoop framework for data analysis within the institution, and the most appropriate way to sustain this will be considered. The consultation process will also investigate the infrastructure requirements for newer technologies such as Apache Spark and for large scale structured data. Extension capability Core to the new research computing infrastructure will be the ability for research groups to add their own purchased equipment to the service. To facilitate this, vacant serviced rack space will be provisioned, pre-cabled with power and ethernet connectivity. The cluster provisioning environment will be given a significant overhaul, moving towards a virtualised provisioning environment to maximise the flexibility in hardware compatibility and deployment modes. This will allow for rapid commissioning of new equipment - making each new node available in the cloud environment on the same day it is removed from its packaging. A simple service costing model will be developed for this type of provisioning to make this attractive for Principal Investigators who hold capital only grant funding. A flexible and capable provisioning system will be included in the service, allowing new pieces of equipment to be rapidly deployed and installed to a working state, without overly onerous processes, licenses or configuration being required. Consultation and feedback In January 2015 we will engage in a consultation with the University's research community to capture specific requirements for this refresh. We intend for this document to be widely circulated and hope that it will act as input for the consultation process. We intend to arrange an open meeting within HSS, CMVM and each school in CSE within the first three weeks of January, and complete the consultation process by the end of January. Please provide feedback and thoughts to either your nominated school/college contact (please see Appendix B), or directly to the ECDF through either Nicholas Moir or Kenton D'Mellow, via the service consultation mailing list: ecdf-refresh@mlist.ed.ac.uk
4 Appendix A Technical detail Processors The current Eddie nodes use Intel Westmere processors, with DDR3 memory. The Intel processors available on the market now include Ivy Bridge (with DDR3 memory) and Haswell (with DDR4 memory). For large memory systems, DDR3 currently offers significantly more capacity per pound than DDR4, and so we can expect that it may make economic sense to use the Ivy Bridge processors in such systems. Ivy Bridge processors offer approximately 2.5x the performance of Westmere processors, and Haswell offers approximately 3x the performance of Westmere processors, so use of either technology will represent a significant boost in node performance. Hyperthreading will be enabled on the majority of serial workload nodes, allowing more slots per node to be available giving better utilisation of CPU resources. Memory Eddie currently provides 2GB of memory per processor core. Over the lifetime of Eddie, this amount has become insufficient for some use cases which emerged over that time (particularly in life sciences). However, many use cases still require substantially less than that. Furthermore, very large (1-2TB) memory is required for a small number of use cases. To accommodate the variety of use cases, the new Eddie cluster will be specified with a range of memory options, and the service configured to make best use of the available configurations. Specifically, a balance of approximately the following memory-per-core will be specified for the compute nodes: 4GB per core: 75% 12GB per core: 25% 2TB total system memory: 2 nodes The service will be reshaped to allow smaller memory slots to be available, thus freeing up larger slots on systems with the lowest memory per core. For example, a 16-core node with 4GB of memory per core will have 64GB of RAM, and this could be divided into service slots as: 4x 1GB 6x 2GB 6x 8GB The exact profile of memory allocation will be determined operationally in response to observed and predicted use.
5 Specialist hardware The refresh will include appropriate accelerator nodes. The existing two Phi nodes provide a multi integrated core architecture and will be maintained. The use of GPGPUs remains an important part of some niche use cases. The Eddie service has provided a small GPGPU cluster environment to allow for initial proof of concept work and small scale production work, and this will be extended with the inclusion of two GPGPU nodes (each with two latest generation NVidia cards) in the cluster. The existing GPGPU nodes will remain available for use while the hardware remains functional. Networking The primary use of Eddie has always been for non-mpi batch processing tasks. The requirement for running multi-node distributed memory jobs has diminished greatly with the increase in node cpu core count, along with many use cases targeting a "commodity" compute environment. Alongside this, the University now has access to the Archer HPC service, which is designed explicitly to service multi-node distributed memory workloads. Therefore, the Eddie service will no longer be provisioned with a network targeted specifically at "MPI" tasks. Instead, the general purpose network will be upgraded from 1GE to 10GE to provide improved large data handling, and also provide some opportunity for multi-node distributed memory jobs (albeit with a potentially lower performance than if using a dedicated infiniband network for this purpose). Pending confirmation from the University's research community, tasks which require multi-node communication will be moved to the Archer service. The core of the Eddie network will be upgraded to 40GE to accommodate the increased edge bandwidth of 10GE, and the GPFS storage servers will be connected directly to this 40GE core, offering improved storage bandwidth. The storage network will also be upgraded to balance this. Storage The disk infrastructure will be refreshed with cost effective, performance driven storage, providing sufficient I/O to drive the enhanced compute cluster. The "iops per node" ratio will be increased to account for the increased core density, and new file systems will be created with the latest performance enhancements implemented (in particular - a large filesystem block size will be used). The current /exports/work filesystem data will be migrated over to the new file system by August GPFS will continue to be the file system used on Eddie, having proven itself to be highly capable and functional over the life of the Eddie service, and proving to be market competitive. By using GPFS to integrate the HPC storage with the DataStore service (see "Supporting Data Science" previously), researchers will no longer need to store large volumes of unused data on the HPC file systems. HPC storage becomes more cost effective as a result, with adequate space only needed for an appropriately sized, user defined "working data set", rather than for all data for a given group. Data will also be easily available from other systems.
6 Extension Capability Two vacant 42U racks will be provisioned, configured with with 1/10GE networking and rack PDU's sufficient for connecting at least 24 hosted systems per rack. Space will be mapped out to allow for further expansion - both of further vacant rack space and of full-rack/multi-rack solutions. If required, per-socket metered PDU's will be installed to allow for exact measurement of power consumption by each hosted device. Appendix B Nominated School/College contacts CSE Biology- Alastair Kerr Physics- Bob Mann Geosciences- Mike Mineter Chemistry- Carole Morrison Mathematics- Steve Law Engineering- Antonis Giannopoulos Informatics- Jim Bednar Alastair.Kerr@ed.ac.uk rgm@roe.ac.uk M.Mineter@ed.ac.uk C.Morrison@ed.ac.uk S.Law@ed.ac.uk A.Giannopoulos@ed.ac.uk jbednar@inf.ed.ac.uk HSS Fraser Muir F.Muir@ed.ac.uk CMVM Paul Clark Paul.Clark@ed.ac.uk
Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers
Information Technology Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers Effective for FY2016 Purpose This document summarizes High Performance Computing
More informationPlatfora Big Data Analytics
Platfora Big Data Analytics ISV Partner Solution Case Study and Cisco Unified Computing System Platfora, the leading enterprise big data analytics platform built natively on Hadoop and Spark, delivers
More informationCloud Computing through Virtualization and HPC technologies
Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC
More informationAgenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
More informationHigh Performance Computing in CST STUDIO SUITE
High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver
More informationSun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
More informationMapR Enterprise Edition & Enterprise Database Edition
MapR Enterprise Edition & Enterprise Database Edition Reference Architecture A PSSC Labs Reference Architecture Guide June 2015 Introduction PSSC Labs continues to bring innovative compute server and cluster
More informationBuilding a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
More informationConfiguration Maximums VMware Infrastructure 3
Technical Note Configuration s VMware Infrastructure 3 When you are selecting and configuring your virtual and physical equipment, you must stay at or below the maximums supported by VMware Infrastructure
More information3 Red Hat Enterprise Linux 6 Consolidation
Whitepaper Consolidation EXECUTIVE SUMMARY At this time of massive and disruptive technological changes where applications must be nimbly deployed on physical, virtual, and cloud infrastructure, Red Hat
More informationOracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers
Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers Masood Ahmed EMEA Infrastructure Solutions Oracle/SAP Relationship Overview First SAP R/3 release
More informationWhite Paper Solarflare High-Performance Computing (HPC) Applications
Solarflare High-Performance Computing (HPC) Applications 10G Ethernet: Now Ready for Low-Latency HPC Applications Solarflare extends the benefits of its low-latency, high-bandwidth 10GbE server adapters
More informationCisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage
Cisco for SAP HANA Scale-Out Solution Solution Brief December 2014 With Intelligent Intel Xeon Processors Highlights Scale SAP HANA on Demand Scale-out capabilities, combined with high-performance NetApp
More informationRemoving Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
More informationEnabling High performance Big Data platform with RDMA
Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery
More informationHUAWEI TECHNOLOGIES CO., LTD. HUAWEI FusionServer X6800 Data Center Server
HUAWEI TECHNOLOGIES CO., LTD. HUAWEI FusionServer X6800 Data Center Server HUAWEI FusionServer X6800 Data Center Server Data Center Cloud Internet App Big Data HPC As the IT infrastructure changes with
More informationHETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
More informationDell Reference Configuration for Hortonworks Data Platform
Dell Reference Configuration for Hortonworks Data Platform A Quick Reference Configuration Guide Armando Acosta Hadoop Product Manager Dell Revolutionary Cloud and Big Data Group Kris Applegate Solution
More informationThe Benefits of POWER7+ and PowerVM over Intel and an x86 Hypervisor
The Benefits of POWER7+ and PowerVM over Intel and an x86 Hypervisor Howard Anglin rhbear@us.ibm.com IBM Competitive Project Office May 2013 Abstract...3 Virtualization and Why It Is Important...3 Resiliency
More informationHow To Speed Up A Flash Flash Storage System With The Hyperq Memory Router
HyperQ Hybrid Flash Storage Made Easy White Paper Parsec Labs, LLC. 7101 Northland Circle North, Suite 105 Brooklyn Park, MN 55428 USA 1-763-219-8811 www.parseclabs.com info@parseclabs.com sales@parseclabs.com
More informationHPC Growing Pains. Lessons learned from building a Top500 supercomputer
HPC Growing Pains Lessons learned from building a Top500 supercomputer John L. Wofford Center for Computational Biology & Bioinformatics Columbia University I. What is C2B2? Outline Lessons learned from
More informationHow To Build A Cisco Ukcsob420 M3 Blade Server
Data Sheet Cisco UCS B420 M3 Blade Server Product Overview The Cisco Unified Computing System (Cisco UCS ) combines Cisco UCS B-Series Blade Servers and C-Series Rack Servers with networking and storage
More informationClusters: Mainstream Technology for CAE
Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux
More informationStovepipes to Clouds. Rick Reid Principal Engineer SGI Federal. 2013 by SGI Federal. Published by The Aerospace Corporation with permission.
Stovepipes to Clouds Rick Reid Principal Engineer SGI Federal 2013 by SGI Federal. Published by The Aerospace Corporation with permission. Agenda Stovepipe Characteristics Why we Built Stovepipes Cluster
More informationCluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
More informationHortonworks Data Platform Reference Architecture
Hortonworks Data Platform Reference Architecture A PSSC Labs Reference Architecture Guide December 2014 Introduction PSSC Labs continues to bring innovative compute server and cluster platforms to market.
More informationHP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief
Technical white paper HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief Scale-up your Microsoft SQL Server environment to new heights Table of contents Executive summary... 2 Introduction...
More informationWrangler: A New Generation of Data-intensive Supercomputing. Christopher Jordan, Siva Kulasekaran, Niall Gaffney
Wrangler: A New Generation of Data-intensive Supercomputing Christopher Jordan, Siva Kulasekaran, Niall Gaffney Project Partners Academic partners: TACC Primary system design, deployment, and operations
More informationIntroduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution
Arista 10 Gigabit Ethernet Switch Lab-Tested with Panasas ActiveStor Parallel Storage System Delivers Best Results for High-Performance and Low Latency for Scale-Out Cloud Storage Applications Introduction
More informationAnnouncing the product launch of a Mitel Virtual Mitel Communication Director (Virtual MCD)
Announcing the product launch of a Mitel Virtual Mitel Communication Director (Virtual MCD) In today's global economy, businesses are - more than ever - looking to reduce capital and operating expenses.
More informationHow To Build A Cloud Computer
Introducing the Singlechip Cloud Computer Exploring the Future of Many-core Processors White Paper Intel Labs Jim Held Intel Fellow, Intel Labs Director, Tera-scale Computing Research Sean Koehl Technology
More informationHadoop: Embracing future hardware
Hadoop: Embracing future hardware Suresh Srinivas @suresh_m_s Page 1 About Me Architect & Founder at Hortonworks Long time Apache Hadoop committer and PMC member Designed and developed many key Hadoop
More informationEnabling Technologies for Distributed and Cloud Computing
Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Multi-core CPUs and Multithreading
More information10GBASE T for Broad 10_Gigabit Adoption in the Data Center
10GBASE T for Broad 10_Gigabit Adoption in the Data Center Contributors Carl G. Hansen, Intel Carrie Higbie, Siemon Yinglin (Frank) Yang, Commscope, Inc 1 Table of Contents 10Gigabit Ethernet: Drivers
More informationGPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
More informationComputational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar
Computational infrastructure for NGS data analysis José Carbonell Caballero Pablo Escobar Computational infrastructure for NGS Cluster definition: A computer cluster is a group of linked computers, working
More informationEnabling Technologies for Distributed Computing
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies
More informationWhite Paper on Consolidation Ratios for VDI implementations
White Paper on Consolidation Ratios for VDI implementations Executive Summary TecDem have produced this white paper on consolidation ratios to back up the return on investment calculations and savings
More informationSolving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
More informationThe PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC2013 - Denver
1 The PHI solution Fujitsu Industry Ready Intel XEON-PHI based solution SC2013 - Denver Industrial Application Challenges Most of existing scientific and technical applications Are written for legacy execution
More informationHADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director brett.weninger@adurant.com Dave Smelker, Managing Principal dave.smelker@adurant.com
More informationWhite Paper. Enhancing Storage Performance and Investment Protection Through RAID Controller Spanning
White Paper Enhancing Storage Performance and Investment Protection Through RAID Controller Spanning May 2005 Introduction It is no surprise that the rapid growth of data within enterprise networks is
More informationNews and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren
News and trends in Data Warehouse Automation, Big Data and BI Johan Hendrickx & Dirk Vermeiren Extreme Agility from Source to Analysis DWH Appliances & DWH Automation Typical Architecture 3 What Business
More informationHPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014
HPC Cluster Decisions and ANSYS Configuration Best Practices Diana Collier Lead Systems Support Specialist Houston UGM May 2014 1 Agenda Introduction Lead Systems Support Specialist Cluster Decisions Job
More informationHigh Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates
High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of
More informationSeeking Opportunities for Hardware Acceleration in Big Data Analytics
Seeking Opportunities for Hardware Acceleration in Big Data Analytics Paul Chow High-Performance Reconfigurable Computing Group Department of Electrical and Computer Engineering University of Toronto Who
More informationBuilding a Linux Cluster
Building a Linux Cluster CUG Conference May 21-25, 2001 by Cary Whitney Clwhitney@lbl.gov Outline What is PDSF and a little about its history. Growth problems and solutions. Storage Network Hardware Administration
More informationMaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products
MaxDeploy Ready Hyper- Converged Virtualization Solution With SanDisk Fusion iomemory products MaxDeploy Ready products are configured and tested for support with Maxta software- defined storage and with
More informationUnified Computing Systems
Unified Computing Systems Cisco Unified Computing Systems simplify your data center architecture; reduce the number of devices to purchase, deploy, and maintain; and improve speed and agility. Cisco Unified
More informationCloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
More informationPower Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis
White Paper Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis White Paper March 2014 2014 Cisco and/or its affiliates. All rights reserved. This document
More informationShare and aggregate GPUs in your cluster. F. Silla Technical University of Valencia Spain
Share and aggregate s in your cluster F. Silla Technical University of Valencia Spain ... more technically... Remote virtualization F. Silla Technical University of Valencia Spain Accelerating applications
More informationCan Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation
Can Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation Forward-Looking Statements During our meeting today we may make forward-looking
More informationConfiguration Maximums VMware vsphere 4.0
Topic Configuration s VMware vsphere 4.0 When you select and configure your virtual and physical equipment, you must stay at or below the maximums supported by vsphere 4.0. The limits presented in the
More informationwww.thinkparq.com www.beegfs.com
www.thinkparq.com www.beegfs.com KEY ASPECTS Maximum Flexibility Maximum Scalability BeeGFS supports a wide range of Linux distributions such as RHEL/Fedora, SLES/OpenSuse or Debian/Ubuntu as well as a
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
More informationDell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820
Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820 This white paper discusses the SQL server workload consolidation capabilities of Dell PowerEdge R820 using Virtualization.
More informationOverview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet Version 1.1 June 2010 Authors: Mark Nowell, Cisco Vijay Vusirikala, Infinera Robert Hays, Intel 1. This work represents
More informationBuilding a Private Cloud with Eucalyptus
Building a Private Cloud with Eucalyptus 5th IEEE International Conference on e-science Oxford December 9th 2009 Christian Baun, Marcel Kunze KIT The cooperation of Forschungszentrum Karlsruhe GmbH und
More informationDiablo and VMware TM powering SQL Server TM in Virtual SAN TM. A Diablo Technologies Whitepaper. May 2015
A Diablo Technologies Whitepaper Diablo and VMware TM powering SQL Server TM in Virtual SAN TM May 2015 Ricky Trigalo, Director for Virtualization Solutions Architecture, Diablo Technologies Daniel Beveridge,
More informationComparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
More informationNew Cluster-Ready FAS3200 Models
New Cluster-Ready FAS3200 Models Steven Miller Senior Technical Director and Platform Architect NetApp recently introduced two new models in the FAS3200 series: the FAS3220 and the FAS3250. Our design
More informationCisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops
Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops Greater Efficiency and Performance from the Industry Leaders Citrix XenDesktop with Microsoft
More informationALPS Supercomputing System A Scalable Supercomputer with Flexible Services
ALPS Supercomputing System A Scalable Supercomputer with Flexible Services 1 Abstract Supercomputing is moving from the realm of abstract to mainstream with more and more applications and research being
More informationIOmark- VDI. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC- 150427- b Test Report Date: 27, April 2015. www.iomark.
IOmark- VDI HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC- 150427- b Test Copyright 2010-2014 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VM, VDI- IOmark, and IOmark
More informationColgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP
selects SAP HANA to improve the speed of business analytics with IBM and SAP Founded in 1806, is a global consumer products company which sells nearly $17 billion annually in personal care, home care,
More informationJames Serra Sr BI Architect JamesSerra3@gmail.com http://jamesserra.com/
James Serra Sr BI Architect JamesSerra3@gmail.com http://jamesserra.com/ Our Focus: Microsoft Pure-Play Data Warehousing & Business Intelligence Partner Our Customers: Our Reputation: "B.I. Voyage came
More informationThe virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.
White Paper Virtualized SAP: Optimize Performance with Cisco Data Center Virtual Machine Fabric Extender and Red Hat Enterprise Linux and Kernel-Based Virtual Machine What You Will Learn The virtualization
More informationIOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org
IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator
More informationTHE DEFINITIVE GUIDE FOR AWS CLOUD EC2 FAMILIES
THE DEFINITIVE GUIDE FOR AWS CLOUD EC2 FAMILIES Introduction Amazon Web Services (AWS), which was officially launched in 2006, offers you varying cloud services that are not only cost effective, but also
More informationHow to Deploy OpenStack on TH-2 Supercomputer Yusong Tan, Bao Li National Supercomputing Center in Guangzhou April 10, 2014
How to Deploy OpenStack on TH-2 Supercomputer Yusong Tan, Bao Li National Supercomputing Center in Guangzhou April 10, 2014 2014 年 云 计 算 效 率 与 能 耗 暨 第 一 届 国 际 云 计 算 咨 询 委 员 会 中 国 高 峰 论 坛 Contents Background
More informationUsing Synology SSD Technology to Enhance System Performance. Based on DSM 5.2
Using Synology SSD Technology to Enhance System Performance Based on DSM 5.2 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD Cache as Solution...
More informationMaximizing Server Storage Performance with PCI Express and Serial Attached SCSI. Article for InfoStor November 2003 Paul Griffith Adaptec, Inc.
Filename: SAS - PCI Express Bandwidth - Infostor v5.doc Maximizing Server Storage Performance with PCI Express and Serial Attached SCSI Article for InfoStor November 2003 Paul Griffith Adaptec, Inc. Server
More informationDell s SAP HANA Appliance
Dell s SAP HANA Appliance SAP HANA is the next generation of SAP in-memory computing technology. Dell and SAP have partnered to deliver an SAP HANA appliance that provides multipurpose, data source-agnostic,
More informationIBM Power Systems This is Power on a Smarter Planet
IBM Power Systems This is Power on a Smarter Planet Red Hat Enterprise Linux for IBM Power Systems! Filipe Miranda Global Lead for Linux on IBM System z and Power Systems!, #powerlinux, #bigdata, #IBMWatson,
More informationCrossing the Performance Chasm with OpenPOWER
Crossing the Performance Chasm with OpenPOWER Dr. Srini Chari Cabot Partners/IBM chari@cabotpartners.com #OpenPOWERSummit Join the conversation at #OpenPOWERSummit 1 Disclosure Copyright 215. Cabot Partners
More informationIOS110. Virtualization 5/27/2014 1
IOS110 Virtualization 5/27/2014 1 Agenda What is Virtualization? Types of Virtualization. Advantages and Disadvantages. Virtualization software Hyper V What is Virtualization? Virtualization Refers to
More informationThe Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays
The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays Executive Summary Microsoft SQL has evolved beyond serving simple workgroups to a platform delivering sophisticated
More informationImproved LS-DYNA Performance on Sun Servers
8 th International LS-DYNA Users Conference Computing / Code Tech (2) Improved LS-DYNA Performance on Sun Servers Youn-Seo Roh, Ph.D. And Henry H. Fong Sun Microsystems, Inc. Abstract Current Sun platforms
More informationLarge Scale Storage. Orlando Richards, Information Services orlando.richards@ed.ac.uk. LCFG Users Day, University of Edinburgh 18 th January 2013
Large Scale Storage Orlando Richards, Information Services orlando.richards@ed.ac.uk LCFG Users Day, University of Edinburgh 18 th January 2013 Overview My history of storage services What is (and is not)
More informationTechnical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment
Technical Paper Moving SAS Applications from a Physical to a Virtual VMware Environment Release Information Content Version: April 2015. Trademarks and Patents SAS Institute Inc., SAS Campus Drive, Cary,
More informationGlobal Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com
Global Headquarters: 5 Speen Street Framingham, MA 01701 USA P.508.872.8200 F.508.935.4015 www.idc.com W H I T E P A P E R O r a c l e V i r t u a l N e t w o r k i n g D e l i v e r i n g F a b r i c
More informationPRIMERGY server-based High Performance Computing solutions
PRIMERGY server-based High Performance Computing solutions PreSales - May 2010 - HPC Revenue OS & Processor Type Increasing standardization with shift in HPC to x86 with 70% in 2008.. HPC revenue by operating
More informationC460 M4 Flexible Compute for SAP HANA Landscapes. Judy Lee julee2@cisco.com Released: April, 2015
C460 M4 Flexible Compute for SAP HANA Landscapes Judy Lee julee2@cisco.com Released: April, 2015 UCS Compute Portfolio Performance Optimized for Bare Metal, Virtualized, and Cloud Applications Cloud Scale
More informationPower Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure
White Paper Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure White Paper March 2014 2014 Cisco and/or its affiliates. All rights reserved. This
More informationMichael Kagan. michael@mellanox.com
Virtualization in Data Center The Network Perspective Michael Kagan CTO, Mellanox Technologies michael@mellanox.com Outline Data Center Transition Servers S as a Service Network as a Service IO as a Service
More informationAppro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
More informationOracle Solaris: Aktueller Stand und Ausblick
Oracle Solaris: Aktueller Stand und Ausblick Detlef Drewanz Principal Sales Consultant, EMEA Server Presales The following is intended to outline our general product direction. It
More informationCan High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
More informationDoubling the I/O Performance of VMware vsphere 4.1
White Paper Doubling the I/O Performance of VMware vsphere 4.1 with Broadcom 10GbE iscsi HBA Technology This document describes the doubling of the I/O performance of vsphere 4.1 by using Broadcom 10GbE
More informationMotherboard- based Servers versus ATCA- based Servers
Motherboard- based Servers versus ATCA- based Servers Summary: A comparison of costs, features and applicability for telecom application hosting After many years of struggling for market acceptance, it
More informationCORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER
CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER Tender Notice No. 3/2014-15 dated 29.12.2014 (IIT/CE/ENQ/COM/HPC/2014-15/569) Tender Submission Deadline Last date for submission of sealed bids is extended
More informationLS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
More informationZadara Storage Cloud A whitepaper. @ZadaraStorage
Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers
More informationHPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect a.jackson@epcc.ed.ac.uk
HPC and Big Data EPCC The University of Edinburgh Adrian Jackson Technical Architect a.jackson@epcc.ed.ac.uk EPCC Facilities Technology Transfer European Projects HPC Research Visitor Programmes Training
More information9/26/2011. What is Virtualization? What are the different types of virtualization.
CSE 501 Monday, September 26, 2011 Kevin Cleary kpcleary@buffalo.edu What is Virtualization? What are the different types of virtualization. Practical Uses Popular virtualization products Demo Question,
More informationAchieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
More informationScaling from Workstation to Cluster for Compute-Intensive Applications
Cluster Transition Guide: Scaling from Workstation to Cluster for Compute-Intensive Applications IN THIS GUIDE: The Why: Proven Performance Gains On Cluster Vs. Workstation The What: Recommended Reference
More informationIntel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms
Intel Cloud Builders Guide Intel Xeon Processor-based Servers RES Virtual Desktop Extender Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms Client Aware Cloud with RES Virtual
More information