Customer Success Story University of Cologne

Similar documents
Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Introduction. Need for ever-increasing storage scalability. Arista and Panasas provide a unique Cloud Storage solution

Recommended hardware system configurations for ANSYS users

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Scale-out NAS Unifies the Technical Enterprise

Scala Storage Scale-Out Clustered Storage White Paper

Clusters: Mainstream Technology for CAE

Scaling from Workstation to Cluster for Compute-Intensive Applications

RAID for the 21st Century. A White Paper Prepared for Panasas October 2007

Oracle Database Scalability in VMware ESX VMware ESX 3.5

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

Cluster Implementation and Management; Scheduling

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Improved LS-DYNA Performance on Sun Servers

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Department of Computer Sciences University of Salzburg. HPC In The Cloud? Seminar aus Informatik SS 2011/2012. July 16, 2012

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

How To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory)

LS DYNA Performance Benchmarks and Profiling. January 2009

Make the Most of Big Data to Drive Innovation Through Reseach

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

SUN ORACLE EXADATA STORAGE SERVER

Enabling High performance Big Data platform with RDMA

A High-Performance Storage and Ultra-High-Speed File Transfer Solution

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration

Red Hat Enterprise Linux 6. Stanislav Polášek ELOS Technologies

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

Avid ISIS

Sun Constellation System: The Open Petascale Computing Architecture

SGI. High Throughput Computing (HTC) Wrapper Program for Bioinformatics on SGI ICE and SGI UV Systems. January, Abstract. Haruna Cofer*, PhD

Hadoop Cluster Applications

On-Demand Supercomputing Multiplies the Possibilities

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing


Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

Easier - Faster - Better

Accelerating Network Attached Storage with iscsi

- An Essential Building Block for Stable and Reliable Compute Clusters

IBM Communications Server for Linux - Network Optimization for On Demand business

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Life Sciences Opening the pipe to faster research, discovery, computation and resource sharing

PARALLELS CLOUD STORAGE

EMC ISILON AND ELEMENTAL SERVER

The Ultimate in Scale-Out Storage for HPC and Big Data

Cisco Prime Home 5.0 Minimum System Requirements (Standalone and High Availability)

Accelerating and Simplifying Apache

Delivers high performance, reliability, and security. Is certified by the leading hardware and software vendors

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Altix Usage and Application Programming. Welcome and Introduction

The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.

Esri ArcGIS Server 10 for VMware Infrastructure

PRIMERGY server-based High Performance Computing solutions

Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000

RevoScaleR Speed and Scalability

SUN ORACLE DATABASE MACHINE

Simple Introduction to Clusters

An On-line Backup Function for a Clustered NAS System (X-NAS)

Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage. White Paper

High Performance Computing (HPC)

EMC XtremSF: Delivering Next Generation Performance for Oracle Database

On Demand Satellite Image Processing

White Paper. Version 1.2 May 2015 RAID Incorporated

Meeting budget constraints. Integration and cooperation between network operations and other IT domains. Providing network performance

SMB Direct for SQL Server and Private Cloud

Scalable Multi-Node Event Logging System for Ba Bar

RLX Technologies Server Blades

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking

Milestone Solution Partner IT Infrastructure MTP Certification Report Scality RING Software-Defined Storage

Stanford HPC Conference. Panasas Storage System Integration into a Cluster

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Configuration Maximums VMware vsphere 4.0

Improving Scalability for Citrix Presentation Server

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Four Reasons To Start Working With NFSv4.1 Now

STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

Lab Validation Report

Oracle Exadata: The World s Fastest Database Machine Exadata Database Machine Architecture

Cray DVS: Data Virtualization Service

FOR SERVERS 2.2: FEATURE matrix

SOLID STATE DRIVES AND PARALLEL STORAGE

Qualcomm Achieves Significant Cost Savings and Improved Performance with Red Hat Enterprise Virtualization

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

How to Choose your Red Hat Enterprise Linux Filesystem

SUN ORACLE DATABASE MACHINE

ioscale: The Holy Grail for Hyperscale

Detailed Product Description

Getting the Most Out of VMware Mirage with Hitachi Unified Storage and Hitachi NAS Platform WHITE PAPER

Part V Applications. What is cloud computing? SaaS has been around for awhile. Cloud Computing: General concepts

BMC Recovery Manager for Databases: Benchmark Study Performed at Sun Laboratories

HP ProLiant BL460c takes #1 performance on Siebel CRM Release 8.0 Benchmark Industry Applications running Linux, Oracle

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

MAKING THE BUSINESS CASE

The Assessment of Benchmarks Executed on Bare-Metal and Using Para-Virtualisation

Transcription:

Customer Success Story University of Cologne University of Cologne Case Study: Panasas Storage Cluster Case Study July 2006

Abstract In 2004, the Center for Applied Informatics at the University of Cologne in Germany, sought to bring high-performance computing (HPC) in the form of Linux cluster computing to one of the oldest and most prestigious institutions of higher education. Over the years as the demand on their HPC resources continued to grow, their existing storage systems could not keep pace. This case study details the challenges faced by the Center at the University of Cologne and how Panasas Storage successfully met the Center s requirements for a scalable, ultra high-performance storage infrastructure. Introduction The Center for Applied Informatics at the University of Cologne in Germany is a centralized resource for the study of applied computer science and general-purpose computing services for more than 50,000 students and faculty. In addition to serving the University, this testing facility also supports regional academic research projects that require high-performance computing. The facility was established by merging the resources of the University s regional computing center with those of the University s center for parallel processing. In 2004, the Center sought to bring high-performance computing (HPC) in the form of Linux cluster computing to one of the oldest and most prestigious institutions of higher education. Researchers would be able to use a NPACI Rocks Linux cluster for climate modeling simulations and supporting other applications in chemistry and geophysics. Over time, researchers at other universities and regional technical laboratories would come to rely on the Center to provide the computing capabilities they needed to study complex problems ranging from molecular processes to stellar formation. The applications being undertaken at the Center required substantial computing power and generated large amounts of data. As the demand on HPC resources continued to grow, the existing storage and associated NFS file systems that supported the HPC environment could not keep pace. A comprehensive evaluation was launched to select a new high-performance storage solution and led to the selection of Panasas Storage. This case study details the challenges faced by the Center for Applied Informatics at the University of Cologne, and how the Panasas Storage system has successfully met the Center s requirements for a scalable, ultra high-performance storage infrastructure. Linux Cluster Architecture To support HPC users at the University of Cologne, the Center procured two new parallel compute systems including a distributed memory system known as Clio and a shared memory system called Altix1. 2

Clio is a 129-node Linux NPACI Rocks cluster that has been in operation since late 2004. It serves as an entrance computer and it is comprised of the following: l 2 x 2.2 GHz AMD Opteron processors l 4GB RAM l 2 x 36 GB SCSI discs l 6 x Cisco 3750 switches l 2GB Ethernet connections l 1 x 10 GB Infiniband interface l 1 x Sun Fire Z20z system l Red Hat Linux Enterprise 3.0 The cluster also supports a shared memory model for parallel programming with an MPI library. For research that is not conducive to efficient parallel processing, the Center secured an SGI Altix system an SMP-based computer with 24 Itanium-2, 1.5GHz processors and 48GB of memory. The Center also uses a variety of compute services, including Sun Microsystems SunFire and Sun Ultra Enterprise servers for user applications that require high disk storage and fast I/O to achieve results. System Characteristics The NPACI Rocks Linux cluster is used for meteorology research (climate modeling) that requires substantial computing power and generates very large data sets. Additionally, applications in chemistry and geophysics make use of the high-performance cluster to perform complex data analysis and simulations. Research conducted at regional scientific laboratories relies on the Center s HPC resources to test complex theories and conduct compute-intensive research that can only be accomplished with supercomputing capabilities. In each of these projects, the network storage infrastructure must be very fast. It must be capable of approximately one gigabyte per second of throughput with near-linear performance scalability as more storage is added. Putting the Industry s Best to the Test The Center realized its reliance on legacy NFS servers was creating a significant performance bottleneck. The management of the system was also become a burden to the IT staff. The decision was made to upgrade the storage infrastructure. The Center needed a new storage system that could easily integrate into the existing environment, and scale with ease to support multiple clusters in the future. This would dramatically reduce time to deployment and enable the Center to maximize previous IT investments. Because the Center projected the HPC environment would increase from just over a hundred nodes to thousands of nodes within two years, the new storage solution would have to easily scale to thousands of terabytes without significant reductions in performance. The new storage system also needed to deliver the highest levels of availability as the demand for cluster time was expected to grow. Finally, because the Center had a relatively small IT staff, system reliability, ease-of-use and low management overhead were considered must-have features. 3

Panasas Makes the Grade at University of Cologne The Center tested solutions from several leading storage vendors using agreed-upon parameters in performance and I/O. The Panasas Storage solution was purchased after only a few weeks of evaluation. While benchmarking continues today, the Center has thus far achieved more than a 30X performance improvement over the previous storage system. Benchmarks show a transfer rate of up to 1.3 GB/second when running on 128 dual-processor nodes accessing 20TB of disk space. The Panasas Storage solution was deployed in less than four hours while competitive solutions took weeks to implement. Each of the products offered a different methodology for incorporating NFS, CIFS or GPFS as part of their overall solution. Panasas PanFS, with its fully-parallel DirectFLOW protocol, was the only system to provide direct access between the Linux cluster nodes and storage system to eliminate the performance bottleneck caused by legacy storage systems. Unlike other systems, the Panasas solution required no modifications to the Linux OS kernel. Graphic of Deployment 4

Summary The Center for Applied Informatics at the University of Cologne has successfully deployed the Panasas Storage system, creating a storage infrastructure with unlimited scalability, higher performance and higher availability. The Center uses Panasas Storage with DirectFLOW client access software to provide very high bandwidth to all nodes in a NPACI Rocks Linux cluster without the performance bottlenecks associated with legacy file systems. The Panasas storage solution enables the Linux cluster to avoid idle time and continue processing jobs without waiting for I/O processes, improving the speed at which research projects can be completed. As the Linux cluster grows from hundreds to thousands of nodes, Panasas Storage is more than capable of keeping pace by scaling performance in near-linear fashion as capacity is increased. About Panasas Panasas, Inc., the leader in high-performance scale-out NAS storage solutions, enables enterprise customers to rapidly solve complex computing problems, speed innovation and bring new products to market faster. All Panasas solutions leverage the patented PanFS storage operating system to deliver exceptional performance, scalability and manageability. Phone: 1-888-PANASAS 2010 Panasas Incorporated. All rights reserved. Panasas is a trademark of Panasas, Inc. in the United States and other countries. 5 PW-10-20900