Clusters: Mainstream Technology for CAE



Similar documents
LS DYNA Performance Benchmarks and Profiling. January 2009

LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

PRIMERGY server-based High Performance Computing solutions

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

Microsoft Compute Clusters in High Performance Technical Computing. Björn Tromsdorf, HPC Product Manager, Microsoft Corporation

Recommended hardware system configurations for ANSYS users

Improved LS-DYNA Performance on Sun Servers

ALPS Supercomputing System A Scalable Supercomputer with Flexible Services

Building Clusters for Gromacs and other HPC applications

Sun Constellation System: The Open Petascale Computing Architecture

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

HP reference configuration for entry-level SAS Grid Manager solutions

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

IBM System Cluster 1350 ANSYS Microsoft Windows Compute Cluster Server

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Michael Kagan.

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Finite Elements Infinite Possibilities. Virtual Simulation and High-Performance Computing

High Performance Computing in CST STUDIO SUITE

An Enterprise Backup Solution for GOES Operations Ground Equipment (OGE) and Spacecraft Support Ground System (SSGS)

Easier - Faster - Better

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

SQL Server Business Intelligence on HP ProLiant DL785 Server

Performance Guide. 275 Technology Drive ANSYS, Inc. is Canonsburg, PA (T) (F)

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009

Fast Setup and Integration of ABAQUS on HPC Linux Cluster and the Study of Its Scalability

Cloud Computing through Virtualization and HPC technologies

HP ProLiant servers. Build an adaptive infrastructure that can grow in step with your business

SUN ORACLE EXADATA STORAGE SERVER

ECLIPSE Performance Benchmarks and Profiling. January 2009

Enabling Technologies for Distributed Computing

SAN TECHNICAL - DETAILS/ SPECIFICATIONS

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.

Cluster Computing at HRI

Managing Data Center Power and Cooling

SGI HPC Systems Help Fuel Manufacturing Rebirth

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

HPC Update: Engagement Model

High-Performance Computing Clusters

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

High Performance Oracle RAC Clusters A study of SSD SAN storage A Datapipe White Paper

FLOW-3D Performance Benchmark and Profiling. September 2012

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

Remote Visualization and Collaborative Design for CAE Applications

Dell High-Performance Computing Clusters and Reservoir Simulation Research at UT Austin.

Platfora Big Data Analytics

IBM BladeCenter H with Cisco VFrame Software A Comparison with HP Virtual Connect

Dell s SAP HANA Appliance

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

Enabling Technologies for Distributed and Cloud Computing

Maximize Performance and Scalability of RADIOSS* Structural Analysis Software on Intel Xeon Processor E7 v2 Family-Based Platforms

Recent Advances in HPC for Structural Mechanics Simulations

David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems

HP ProLiant Storage Server family. Radically simple storage

Integrated Grid Solutions. and Greenplum

HP ProLiant BL460c achieves #1 performance spot on Siebel CRM Release 8.0 Benchmark Industry Applications running Microsoft, Oracle

SERVER CLUSTERING TECHNOLOGY & CONCEPT

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking

An Oracle White Paper July Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Seradex White Paper. Focus on these points for optimizing the performance of a Seradex ERP SQL database:

Unified Computing Systems

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud

Panasas: High Performance Storage for the Engineering Workflow

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

QuickSpecs. What's New Support for two InfiniBand 4X QDR 36P Managed Switches

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Multicore Parallel Computing with OpenMP

SUN ORACLE DATABASE MACHINE

OpenPower: IBM s Strategy for Best of Breed 64-bit Linux

SPEED your path to virtualization.

Terminal Server Software and Hardware Requirements. Terminal Server. Software and Hardware Requirements. Datacolor Match Pigment Datacolor Tools

Evaluation Report: HP Blade Server and HP MSA 16GFC Storage Evaluation

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

Block based, file-based, combination. Component based, solution based

An Oracle White Paper June High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database

Sun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008

Certification: HP ATA Servers & Storage

UCS M-Series Modular Servers

Transcription:

Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP

Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux enabled the use industry-standard technologies Many more users and new applications Cluster growth rate is over 50% per year! (volume is half of HPC) Now a critical resource in meeting today s CAE challenges Increasingly complex CAE analysis demands more larger models; more jobs to run; longer runs Market is responding, adding enterprise RAS features to clusters Treating CLUSTERS like PRODUCTS, not custom deployments Integration with large SMP systems allows one to optimize resource deployment Some jobs just can t be distributed 2

Why cluster? Budget: Price-performance (+10 GFLOPs system today < $4K) Scale beyond practical SMP limits Faster time to market and profit, improved insights Resource consolidation Centralized management, optimize utilization Clusters aren t just for compute engines Can apply same principles to file systems and visualization Can help deal with exponential growth in volume of simulation data 3

Application Experience User Application (Courtesy of NTUST) A large-scale FE model (nonlinear continuum mechanics) Computing time of 80 days was necessary with 1-CPU in year 2000 14 processors of AMD Athlon 1600+ with Myrinet 67 hours 96 processor cores of HP Opteron 270 at NTUST cluster < 12 hours A home-made application ported in less than a day. NTUST: National Taiwan University of Science and Technology 4

SMP vs. Cluster (farm) Example 7000 MSC.Nastran: XLTDF Comparison 6000 Total elapsed time 5000 4000 3000 2000 1000 0 1 2 4 Number of processes Integrity rx5670 4 way SMP ProLiant DL145 G2-2 node cluster Integrity rx2620-2 node cluster 5

CAE Application Sub-Segments CAE Domain: Pre/Post Structures Impact Fluids Parallelized Serial (SMP*) SMP (MPI*) MPI MPI Job Scalability 32 64 GB 1 4 (8*) cores 2 16 (32*) cores 4 128 (256*) cores Typical Solution Workstation or SMP server Integrity SMP or Farm X64 Cluster X64 Cluster CPU cycles Auto CPU cycles Aero All jobs 30% 30% 60% 20% 10% 50% (*emerging capability) 6

HPC Cluster Implementation challenges System and workload management Scalable performance Scalable data management Interconnect/Network Complexity Application availability and scalability Power and cooling Acquisition and deployment 7

Latest Advancements in Clustering Multi-core delivering continued price-performance improvements Improvements in clustering software and tools More applications are being developed and tuned to leverage cluster/dmp solutions Principles of compute clusters being applied to storage and visualization InfiniBand now established in HPC Solutions now coming to market that address power and cooling concerns 8

HP Unified Cluster Portfolio 9

Powerful Solver Technology Applications: ISVs standardizing on HP-MPI Molpro University of Cardiff AMLS One of the top reasons that we went with HP-MPI is that we've had a great working relationship with HP. It was a win-win for ANSYS, HP and our customers - in terms of cost, interconnects, support and performance compared to other message passing interfaces for Linux and Unix. In addition, I've always had great turnaround from HP in response to hardware and software issues. Lisa Fordanich, Senior Systems Specialist, ANSYS www.ansys.com/services/ss-interconnects.htm HP-MPI is an absolute godsend, notes Keith Glassford, director of the Materials Science division at San Diego, CA-based Accelrys Software Inc. It allows us to focus our energy and resources on doing what we re good at, which is developing scientific and engineering software to solve customer problems. 10

CAE Reference Architecture Client Workstations Front End HA job scheduler pre/post SMP RGS direct attached Disk Array (or use SFS) compute SMPs Remote Workstations compute compute compute clusters clusters compute compute compute clusters clusters InfiniBand switched fabric interconnect LAN Scalable File Share meta data object data visualization cluster 11

A Cluster Alternative to Direct Attached Storage: HP Scalable File Share (SFS) Applying principles of clusters to file systems and storage enables the sharing of data sets without performance penalty MSC.Nastran is Fast on HP SFS: Replace extra-disk fat-nodes with flexible storage Traditional approach: Special nodes in the cluster w/ multiple local JBOD disks Expensive and hard to manage New approach Use fast centralized, virtualized HP SFS filesystem Similar performance Lower cost Shared rather than dedicated storage Easier to use Any node in the cluster can run Nastran Higher reliability: RAID 6 instead of RAID 0 12

MSC.Nastran Benchmark XXCMD Standard MSC benchmark XXCMD: solution of the natural frequencies of an automotive body Performs a medium amount of I/O compared to industry real-life customer datasets (4 TB of I/O with blocksize of 256 KB) Multiple jobs running simultaneously: no shared data Customers typically use direct attached storage for each host 1 controller and 5 drives per job are recommended for good throughput SFS performance 1 Object Storage Server node and 4 enclosures (with array of SATA drives) for every 4 hosts achieved excellent performance No degradation for up to 16 hosts, and small degradation from 16 to 32 hosts Significant (~6 times) advantage vs. small SCSI configuration time (sec) 140000 120000 100000 80000 60000 40000 20000 0 MSC.Nastran benchmark XXCMD - performs medium I/O (small is better) 1 2 4 8 16 32 # hosts SFS 2 jobs per host MSA 2 jobs per host SCSI 2 jobs per host 13

Key Considerations in Designing a Solution What processor and interconnect for the mix of jobs Centralized resource or single purpose systems Can applications co-exist? Economics of consolidation Environmentals: power, cooling, weight, space Roll your own system or acquire a total solution Production scalability requirements Performance Availability and Reliability Manageability (provisioning, booting, monitoring, upgrades) Budget, of course and TCO 14

For more information see www.hp.com/go/hptc Cluster Platform Express: www.hp.com/go/cp-express alanna.dwyer@hp.com 15

Implementations of CAE Reference Architecture: AMD Opteron example Fastest Faster Fast HP xw9300 Workstation Opteron Workstation for Pre/Post XW9300 2 Dual Core Opteron 2.6 GHz CPUs 2 internal 146 GB drives 32 GB memory DVD ProLiant DL585 Server with Disk Array Opteron Server for Structural Analysis DL585 22U Rack with Factory integration 4 Dual Core Opteron CPUs 2 internal 146 GB drives 32 GB memory MSA30 Dual Bus CP 4000 Cluster Opteron Cluster for CFD and Impact Analysis HP Cluster Platform 4000 compute cluster 42U Rack Sidewinder option DL385 head node for cluster administration DL145G2 with two Dual Core Opteron CPUs, each with 1 internal drive and 4 GB memory (1GB/core) DL585 front end node with 64GB for grid generation and domain decompositioin XC Software Operating Environment support 16