Leistungsanalyse von Rechnersystemen
|
|
- Clara Booth
- 8 years ago
- Views:
Transcription
1 Center for Information Services and High Performance Computing (ZIH) Leistungsanalyse von Rechnersystemen Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) Organization Lecture: Every Wednesday in INF E07 from 13:00 to 14:30 Exercise: Every Thursday in INF E010 from 9:20 to10:50 First Exercise: October 19th, guided tour through all new machine rooms at ZIH All slides will be in English Ten minute summary of last lecture at the beginning of each lecture. Also given in English 1
2 Class Material on the Web Slides will be put on the web prior or shortly after each class Bildungsportal Sachsen Login required but identical to ZIH or INF login ZIH web pages Class Outline 14 lectures with 12 corresponding exercises Class structure Introduction and motivation Performance requirements and common evaluation mistakes Performance metrics and evaluation techniques Workload types, selection, and characterization Commonly used benchmarks Benchmarks specialized on I/O Monitoring techniques Capacity planning for future systems Performance data presentation Comparing system using sample data Regression models Experimental design Performance simulation and prediction Introduction to queueing theory 2
3 Literature Raj Jain: The Art of Computer Systems Performance Analysis John Wiley & Sons, Inc., 1991 (ISBN: ) Rainer Klar, Peter Dauphin, Fran Hartleb, Richard Hofmann, Bernd Mohr, Andreas Quick, Markus Siegle Messung und Modellierung paralleler und verteilter Rechensysteme B.G. Teubner Verlag, Stuttgart, 1995 (ISBN: ) Dongarra, Gentzsch, Eds.: Computer Benchmarks, Advances in Parallel Computing 8, North Holland, 1993 (ISBN: x) Motivation Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) 3
4 Innovations that changed our daily life steam engine, motor railway, car, airplanes fertilizer telephone computer energy transportation food communication data processing Speed of data processing Human Workstation, PC Supercomputer 10-2 FLOPS 10 8 FLOPS FLOPS Ratio: factor
5 5
6 6
7 HPC A key technology? USA defines strategic mission of HPC Software, methods and human beings Main motivation from military applications Integration of know-how in the country Attraction of experts from all over the world Japan: Creator of the Earth-Simulator Petaflop special purpose machine for MD simulations Petaflop project is in preparation EU is still discussing an initiative 7
8 Accelerated Strategic Computing Initiative (ASCI) Strategic Initiative in the U.S.A. ASCI Red (Sandia): Intel-System with 1TFLOP (sustained) ASCI Blue (LANL;LLNL): IBM und SGI, 3 TFLOP each (sustained) ASCI White (LLNL): IBM Power 3 (10 TFLOPS) ASCI Q (LANL): COMPAQ-Rechner (20 TFLOPS) Red Storm (Sandia) Cray XT3, Opteron (40 TFLOPS) ASCI Purple (LLNL): IBM Power 4 (100 TFLOPS) ASCI BlueGene (LLNL): IBM PowerPC (180/360 TFLOPS) What kind of know-how is required for HPC? Algorithms and methods Performance Programming (Paradigms and details of implementations) Operation of supercomputers (network, infrastructure, service, support) 8
9 Challenges Languages Fortran95, C++, Java Parallelization: MPI, OpenMP Network ATM, IPv6, Gigabit Scheduling Distributed Components, mobile agents System architecture Processors, memory hierarchy What is the best programming models for clustered SMPs with a deep memory hierarchy? Software a key technology Software is a key factor for progress in our country Is Germany a location for software development? WWW is everywhere (E-Commerce, Google, EBay, ) Contribution of HPC: Optimizing Servers Optimizing Access to data bases Optimizing applications 9
10 Center of Information Services and HPC A short introduction Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) HPC in Germany 10
11 Center for Information Services and HPC (ZIH) Central Scientific Unit at TU Dresden Merged institution: TUD Computing Center (URZ) and Center for High Performance Computing (ZHR) Competence Center for Parallel Computing and Software Tools Strong commitment to support real users Development of algorithms and methods: Cooperation with users from all departments Structure Management Director: Deptuy. Directors: Prof. Dr. W. E. Nagel Dr. P. Fischer Dr. M. Müller Unit ZSD Central Systems and Services Dr. S. Maletti Unit IAK Interdiciplinary Application Development and Coordination Dr. M. Müller Unit NK Network and Communication W. Wünsch Unit IMC Innovative Methods of Computing PD Dr. A. Deutsch Unit PSW Programming and Software Tools Dr. H. Mix 11
12 Responsibilities of ZIH Providing infrastructure and qualified service for TU Dresden and Saxony Research topics Architecture and performance analysis of High Performance Computers Programming methods and techniques for HPC systems Software tools to support programming and optimization Modeling algorithms of biological processes Mathematical models, algorithms, and efficient implementations Role of mediator between vendors, developers, and users Pick up and preparation of new concepts, methods, and techniques Teaching and Education Procurement: Overall Infrastructure / Future Directions HPC-Server Main memory : 4 TB PC-Farm 8 GB/s 4 GB/s 4 GB/s HPC-SAN Capacity: > 50 TB PC-SAN Capacity: > 50 TB HPC-Component SGI Altix 4700 >2000 of the latest Itanium -Cores 6 TByte main memory 1,8 GB/s PetaByte Tape Silo Capacity: 1 PB PC-Farm System from Linux Networx AMD opteron CPUs >700 boards with >2500 cores Infiniband networks between the nodes 12
13 Timeline Machine Room Upgrade Jul Aug Sep Oct Nov Dec Jan Feb Mar Apr May Jun Jul Aug Sep Installation Stage 1a (Test operation) Installation Stage 1b Installation Stage 2 Center for Information Services and High Performance Computing (ZIH) Stage1a Test Systems Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) 13
14 HRSK Stage 1a HPC-Server merkur.hrsk.tu-dresden.de SGI Altix 3700 Bx2 192x 1.5GHz/4MB L3 Cache Itanium2 CPUs 1152 GFlops/s Peak Performance 768 GB Shared Memory (4 GB/CPU), NUMA 1 TB lokal discs + 34 TB SAN SuSE SLES 9 inkl. SGI ProPack 4 Intel Compiler and Tools: Alinea DDT Debugger Batchsystem LSF 1x DDN RAID System S2A9500: 2x S2A9500 Couplet (5GB Cache, 8x FC4 Ports) 292x 146GB 10k RPM FC Disks (4 hot spare) 34 TB net capacity PC-Farm Stage 1a phobos.hrsk.tu-dresden.de 64 dual CPU nodes 128 AMD Opteron DP GHz (Single-Core) CPUs 563,2 GFlops/s peak performance 256 GB main memory ( 4GB per node) SUSE operating system Infiniband 4x Interconnect 80 GB local disc per node 21,2 TB shared disc space: 2x DDN RAID System S2A x 146GB 10k RPM FC discs (4 Hot-Spare) 14
15 Center for Information Services and High Performance Computing (ZIH) Stage1b,2 Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) HRSK Stufe 1b/2 start of installation 31.Juli 2006 Petabyte-Bandarchiv SUN STK SL Slots 30 LTO-3 tape drives 2500 LTO-3 tapes 15
16 HRSK HPC-System SGI Altix x 42U Racks 1024 x Sockets mit Itanium2 Montecito Dual- Core CPUs (1.6 GHz/9MB L3 Cache) 13 TFlops/s Peak Performance 6,5 TB Shared Memory HRSK PC-Farm Linux Networx PC-Farm (final configuration) 26 x 42U Racks + water cooled AMD Opteron X85 Dual Core Chip mit 2,6 GHz 384x Single CPU Nodes 232x Dual CPU Nodes 112x Quad CPU Nodes 2 GB main memory (ECC) per Core für Stufe 2 12 TFlops/s Peak Performance 16
17 HRSK Stage 2 HPC-SAN and PC-SAN SGI InfiniteStorage 6700 (DDN S2A9500) HPC-SAN: 68 TB PC-SAN: 51 HRSK HPC-Komponente Hauptspeicher 6,5 TB PC-Farm 8 GB/s 4 GB/s 4 GB/s HPC-SAN Festplattenkapazität: 68 TB PC-SAN Festplattenkapazität: 51 TB 1,8 GB/s PetaByte- Bandarchiv Kapazität: 1 PB 17
18 Trefftz and Willers Building New Extension 18
19 Location of Computer Rooms Anbau Treffz-Bau 19
20 SGI Altix 4700 at ZIH PC-Farm at ZIH 20
21 Configuration of overall system: SAN Overview Beschreibung der Lösung von SGI HPC-SAN Gesamtkapazität: 68 TB durchgängig 4 Gb/s FC 4 x DDN S2A 9500 je 17 TB 584 Festplatten 146 GB CXFS/DMF auf Altix 350 (24 Itanium) TP 9300 (MDS Storage Subsystem) 14 x 73 GB für Metadaten Zugang von PC-Farm: NFS-Server auf 12 x Altix 350 mit je 2 CPUs oder Opteron (für RDMA-Zugriff) 21
22 Beschreibung der Lösung von SGI HPC-Komponente mehr als 500 dual-core Itanium-2 (Montecito) 1,6 GHz, 18 MB L3 (pro core 9 MB) 12,8 GFlops Peak 4 8 GB RAM (DDR2) S = 6 TB verbunden über SGI NumaLink 4 Bandbreite: 3,2 GB/s pro Knoten und Richtung Fat-Tree-Topologie Grafik-Pipes + Grafik-Compositor RASC Blade mit zwei FPGAs (RASClib) Beschreibung der Lösung von SGI PC-Farm mehr als 700 Boards Prozessoren: AMD Opteron Verbindungsnetzwerk: IB X4 Compute-Knoten verbunden über drei Switche (288 ports) Anbindung an HPC-SAN über 12 NFS-Server (CXFS-Clients) 22
23 Beschreibung der Lösung von SGI PC-SAN Lustre FS 2 x DDN S2A 9500 Kapazität: 50,9 TB 440 Festplatten 146 GB Tape Silo - Details CXFS/DMF-Server on Altix 350 (24 CPUs, 48 GB) Data Migration Facility (Licence for 1 bzw. 2 PB) 2 x FC-Switches (24 ports) StorageTek SL 8500 (SUN) ACSLS-Lizenz for 2500 Slots 23
24 Performance of Computers at ZIH Some Activities Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) 24
25 Vampir: Technical Components Trace 1 Trace 2 Trace 3 Trace N Tools Worker 1 Worker 2 Worker m Server Master 1. Trace generator 2. Classical Vampir viewer and analyzer 3. Vampir client viewer 4. Parallel server engine 5. Conversion and analysis tools Vampir: Timeline 25
26 Vampir: Scalability sppm ASCI Benchmark 3D Gas Dynamic Data to be analyzed 16 Processes 200 MByte Volume Number of Workers Load Time 47,33 22,48 10,80 5,43 3,01 3,16 Timeline 0,10 0,09 0,06 0,08 0,09 0,09 Summary Profile 1,59 0,87 0,47 0,30 0,28 0,25 Process Profile 1,32 0,70 0,38 0,26 0,17 0,17 Com. Matrix 0,06 0,07 0,08 0,09 0,09 0,09 Stack Tree 2,57 1,39 0,70 0,44 0,25 0,25 Vampir: A Large Test Case IRS ASCI Benchmark Implicit Radiation Solver Data to be analyzed: 64 Processes in 8 Streams Approx Events 40 GByte Data Volume Analysis Platform: Jump.fz-juelich.de 41 IBM p690 nodes 32 processors per node 128 GByte per node Visualization Platform: Remote Laptop 26
27 BenchIT: Key Components 1 BenchIT measurement core Measurement kernels Exact timer Running kernels with variable problem sizes Generating result files BenchIT: Key Components 2 BenchIT measurement core Command line interface 27
28 BenchIT: Key Components 3 BenchIT measurement core Command line interface GUI BenchIT: Key Components 4 BenchIT measurement core Command line interface GUI Website 28
29 Parbench: Influence of the operating system Full Load Test: 72 sequential jobs 9 jobs eightfold parallelized 144 CPUs Parallelized kernel sequences more then 250 s CPU time Influence by OS Parbench: Influence of the operating system Under Load Test: 68 sequential jobs 9 jobs eightfold parallelized 144 CPUs 4 CPUs free for OS Parallelized kernel sequences have fewer CPU time 29
30 Benchmarks: Scalability of /fastfs file system.8.6 bandwidth[gb/s] P a Pt2Pt latency between all possible pairs "64-2/result-all2all-latency.log" u 4:5:
31 Pt2Pt bandwidth between all possible pairs "64/result-all2all-bandwidth.log" u 3:4: Thank you! Hope to see you next time Zellescher Weg 12 Willers-Bau A113 Tel Matthias S. Mueller (matthias.mueller@tu-dresden.de) 31
Altix Usage and Application Programming. Welcome and Introduction
Zentrum für Informationsdienste und Hochleistungsrechnen Altix Usage and Application Programming Welcome and Introduction Zellescher Weg 12 Tel. +49 351-463 - 35450 Dresden, November 30th 2005 Wolfgang
More informationAppro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
More informationAgenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC
HPC Architecture End to End Alexandre Chauvin Agenda HPC Software Stack Visualization National Scientific Center 2 Agenda HPC Software Stack Alexandre Chauvin Typical HPC Software Stack Externes LAN Typical
More informationBSC - Barcelona Supercomputer Center
Objectives Research in Supercomputing and Computer Architecture Collaborate in R&D e-science projects with prestigious scientific teams Manage BSC supercomputers to accelerate relevant contributions to
More informationKriterien für ein PetaFlop System
Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working
More informationMEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit. Prof. Dieter Kranzlmüller (LRZ)
MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit Prof. Dieter Kranzlmüller (LRZ) LRZ HPC-Systems at the End of the UNIX-Era (Years 2000-2002) German national supercomputer Hitachi SR800
More informationPerformance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
More informationSun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
More informationHPC Update: Engagement Model
HPC Update: Engagement Model MIKE VILDIBILL Director, Strategic Engagements Sun Microsystems mikev@sun.com Our Strategy Building a Comprehensive HPC Portfolio that Delivers Differentiated Customer Value
More informationParallel Programming Survey
Christian Terboven 02.09.2014 / Aachen, Germany Stand: 26.08.2014 Version 2.3 IT Center der RWTH Aachen University Agenda Overview: Processor Microarchitecture Shared-Memory
More informationGPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
More informationCluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
More informationPerformance Tools for System Monitoring
Center for Information Services and High Performance Computing (ZIH) 01069 Dresden Performance Tools for System Monitoring 1st CHANGES Workshop, Jülich Zellescher Weg 12 Tel. +49 351-463 35450 September
More informationOpenMP Programming on ScaleMP
OpenMP Programming on ScaleMP Dirk Schmidl schmidl@rz.rwth-aachen.de Rechen- und Kommunikationszentrum (RZ) MPI vs. OpenMP MPI distributed address space explicit message passing typically code redesign
More informationCray XT3 Supercomputer Scalable by Design CRAY XT3 DATASHEET
CRAY XT3 DATASHEET Cray XT3 Supercomputer Scalable by Design The Cray XT3 system offers a new level of scalable computing where: a single powerful computing system handles the most complex problems every
More informationBuilding Clusters for Gromacs and other HPC applications
Building Clusters for Gromacs and other HPC applications Erik Lindahl lindahl@cbr.su.se CBR Outline: Clusters Clusters vs. small networks of machines Why do YOU need a cluster? Computer hardware Network
More informationTHE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC
THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC The Right Data, in the Right Place, at the Right Time José Martins Storage Practice Sun Microsystems 1 Agenda Sun s strategy and commitment to the HPC or technical
More informationClusters: Mainstream Technology for CAE
Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux
More informationJUROPA Linux Cluster An Overview. 19 May 2014 Ulrich Detert
Mitglied der Helmholtz-Gemeinschaft JUROPA Linux Cluster An Overview 19 May 2014 Ulrich Detert JuRoPA JuRoPA Jülich Research on Petaflop Architectures Bull, Sun, ParTec, Intel, Mellanox, Novell, FZJ JUROPA
More informationBuilding a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
More informationJuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering
JuRoPA Jülich Research on Petaflop Architecture One Year on Hugo R. Falter, COO Lee J Porter, Engineering HPC Advisoy Counsil, Workshop 2010, Lugano 1 Outline The work of ParTec on JuRoPA (HF) Overview
More informationSupercomputing 2004 - Status und Trends (Conference Report) Peter Wegner
(Conference Report) Peter Wegner SC2004 conference Top500 List BG/L Moors Law, problems of recent architectures Solutions Interconnects Software Lattice QCD machines DESY @SC2004 QCDOC Conclusions Technical
More informationPRIMERGY server-based High Performance Computing solutions
PRIMERGY server-based High Performance Computing solutions PreSales - May 2010 - HPC Revenue OS & Processor Type Increasing standardization with shift in HPC to x86 with 70% in 2008.. HPC revenue by operating
More informationCloud Computing through Virtualization and HPC technologies
Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC
More informationTrends in High-Performance Computing for Power Grid Applications
Trends in High-Performance Computing for Power Grid Applications Franz Franchetti ECE, Carnegie Mellon University www.spiral.net Co-Founder, SpiralGen www.spiralgen.com This talk presents my personal views
More informationPCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)
PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters from One Stop Systems (OSS) PCIe Over Cable PCIe provides greater performance 8 7 6 5 GBytes/s 4
More informationThematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science
Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science Call for Expression of Interest (EOI) for the Supply, Installation
More informationHow To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
More informationA Scalable Ethernet Clos-Switch
A Scalable Ethernet Clos-Switch Norbert Eicker John von Neumann-Institute for Computing Research Centre Jülich Technisches Seminar Desy Zeuthen 9.5.2006 Outline Motivation Clos-Switches Ethernet Crossbar
More informationCOMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)
COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1) Mary Thomas Department of Computer Science Computational Science Research Center (CSRC) San Diego State University
More informationCORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER
CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER Tender Notice No. 3/2014-15 dated 29.12.2014 (IIT/CE/ENQ/COM/HPC/2014-15/569) Tender Submission Deadline Last date for submission of sealed bids is extended
More informationPerformance Characteristics of Large SMP Machines
Performance Characteristics of Large SMP Machines Dirk Schmidl, Dieter an Mey, Matthias S. Müller schmidl@rz.rwth-aachen.de Rechen- und Kommunikationszentrum (RZ) Agenda Investigated Hardware Kernel Benchmark
More informationSRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center
SRNWP Workshop HP Solutions and Activities in Climate & Weather Research Michael Riedmann European Performance Center Agenda A bit of marketing: HP Solutions for HPC A few words about recent Met deals
More informationOracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
More informationMain Memory Data Warehouses
Main Memory Data Warehouses Robert Wrembel Poznan University of Technology Institute of Computing Science Robert.Wrembel@cs.put.poznan.pl www.cs.put.poznan.pl/rwrembel Lecture outline Teradata Data Warehouse
More informationLS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
More informationCooling and thermal efficiently in
Cooling and thermal efficiently in the datacentre George Brown HPC Systems Engineer Viglen Overview Viglen Overview Products and Technologies Looking forward Company Profile IT hardware manufacture, reseller
More informationIntroduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
More informationMPI / ClusterTools Update and Plans
HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski
More informationALPS Supercomputing System A Scalable Supercomputer with Flexible Services
ALPS Supercomputing System A Scalable Supercomputer with Flexible Services 1 Abstract Supercomputing is moving from the realm of abstract to mainstream with more and more applications and research being
More informationCray Gemini Interconnect. Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak
Cray Gemini Interconnect Technical University of Munich Parallel Programming Class of SS14 Denys Sobchyshak Outline 1. Introduction 2. Overview 3. Architecture 4. Gemini Blocks 5. FMA & BTA 6. Fault tolerance
More informationSUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION
SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION AFFORDABLE, RELIABLE, AND GREAT PRICES FOR EDUCATION Optimized Sun systems run Oracle and other leading operating and virtualization platforms with greater
More informationSun Microsystems Special Promotions for Education and Research January 9, 2007
Sun Microsystems Special Promotions for Education and Research Solve big problems on a small budget with Sun-Education s trusted partner for cutting-edge technology solutions. Sun solutions help your campus
More informationIntel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband
Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband A P P R O I N T E R N A T I O N A L I N C Steve Lyness Vice President, HPC Solutions Engineering slyness@appro.com Company Overview
More informationA Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures
11 th International LS-DYNA Users Conference Computing Technology A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures Yih-Yih Lin Hewlett-Packard Company Abstract In this paper, the
More informationDeploying and managing a Visualization Farm @ Onera
Deploying and managing a Visualization Farm @ Onera Onera Scientific Day - October, 3 2012 Network and computing department (DRI), Onera P.F. Berte pierre-frederic.berte@onera.fr Plan Onera global HPC
More informationbenchmarking Amazon EC2 for high-performance scientific computing
Edward Walker benchmarking Amazon EC2 for high-performance scientific computing Edward Walker is a Research Scientist with the Texas Advanced Computing Center at the University of Texas at Austin. He received
More informationLecture 1: the anatomy of a supercomputer
Where a calculator on the ENIAC is equipped with 18,000 vacuum tubes and weighs 30 tons, computers of the future may have only 1,000 vacuum tubes and perhaps weigh 1½ tons. Popular Mechanics, March 1949
More informationRemoving Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
More informationHigh Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/2015. 2015 CAE Associates
High Performance Computing (HPC) CAEA elearning Series Jonathan G. Dudley, Ph.D. 06/09/2015 2015 CAE Associates Agenda Introduction HPC Background Why HPC SMP vs. DMP Licensing HPC Terminology Types of
More informationINDIAN INSTITUTE OF TECHNOLOGY KANPUR Department of Mechanical Engineering
INDIAN INSTITUTE OF TECHNOLOGY KANPUR Department of Mechanical Engineering Enquiry No: Enq/IITK/ME/JB/02 Enquiry Date: 14/12/15 Last Date of Submission: 21/12/15 Formal quotations are invited for HPC cluster.
More informationCMS Tier-3 cluster at NISER. Dr. Tania Moulik
CMS Tier-3 cluster at NISER Dr. Tania Moulik What and why? Grid computing is a term referring to the combination of computer resources from multiple administrative domains to reach common goal. Grids tend
More informationHeadline in Arial Bold 30pt. The Need For Speed. Rick Reid Principal Engineer SGI
Headline in Arial Bold 30pt The Need For Speed Rick Reid Principal Engineer SGI Commodity Systems Linux Red Hat SUSE SE-Linux X86-64 Intel Xeon AMD Scalable Programming Model MPI Global Data Access NFS
More informationThe CNMS Computer Cluster
The CNMS Computer Cluster This page describes the CNMS Computational Cluster, how to access it, and how to use it. Introduction (2014) The latest block of the CNMS Cluster (2010) Previous blocks of the
More informationCopyright 2013, Oracle and/or its affiliates. All rights reserved.
1 Oracle SPARC Server for Enterprise Computing Dr. Heiner Bauch Senior Account Architect 19. April 2013 2 The following is intended to outline our general product direction. It is intended for information
More informationHigh Performance Computing in CST STUDIO SUITE
High Performance Computing in CST STUDIO SUITE Felix Wolfheimer GPU Computing Performance Speedup 18 16 14 12 10 8 6 4 2 0 Promo offer for EUC participants: 25% discount for K40 cards Speedup of Solver
More informationHigh Performance Computing. Course Notes 2007-2008. HPC Fundamentals
High Performance Computing Course Notes 2007-2008 2008 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs
More informationPanasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory
Customer Success Story Los Alamos National Laboratory Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory June 2010 Highlights First Petaflop Supercomputer
More informationFLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
More informationEnabling Technologies for Distributed and Cloud Computing
Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Multi-core CPUs and Multithreading
More informationECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009
ECLIPSE Best Practices Performance, Productivity, Efficiency March 29 ECLIPSE Performance, Productivity, Efficiency The following research was performed under the HPC Advisory Council activities HPC Advisory
More informationSun in HPC. Update for IDC HPC User Forum Tucson, AZ, Sept 2008
Sun in HPC Update for IDC HPC User Forum Tucson, AZ, Sept 2008 Bjorn Andersson Director, HPC Marketing Makia Minich Lead Architect, Sun HPC Software, Linux Edition Sun Microsystems Core Focus Areas for
More informationDELL. Virtual Desktop Infrastructure Study END-TO-END COMPUTING. Dell Enterprise Solutions Engineering
DELL Virtual Desktop Infrastructure Study END-TO-END COMPUTING Dell Enterprise Solutions Engineering 1 THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL
More informationImproved LS-DYNA Performance on Sun Servers
8 th International LS-DYNA Users Conference Computing / Code Tech (2) Improved LS-DYNA Performance on Sun Servers Youn-Seo Roh, Ph.D. And Henry H. Fong Sun Microsystems, Inc. Abstract Current Sun platforms
More informationMississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC
Mississippi State University High Performance Computing Collaboratory Brief Overview Trey Breckenridge Director, HPC Mississippi State University Public university (Land Grant) founded in 1878 Traditional
More informationAT&T Global Network Client for Windows Product Support Matrix January 29, 2015
AT&T Global Network Client for Windows Product Support Matrix January 29, 2015 Product Support Matrix Following is the Product Support Matrix for the AT&T Global Network Client. See the AT&T Global Network
More informationHow To Compare Amazon Ec2 To A Supercomputer For Scientific Applications
Amazon Cloud Performance Compared David Adams Amazon EC2 performance comparison How does EC2 compare to traditional supercomputer for scientific applications? "Performance Analysis of High Performance
More informationNew Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
More informationECLIPSE Performance Benchmarks and Profiling. January 2009
ECLIPSE Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox, Schlumberger HPC Advisory Council Cluster
More informationPARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN
1 PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN Introduction What is cluster computing? Classification of Cluster Computing Technologies: Beowulf cluster Construction
More informationComparing Dynamic Disk Pools (DDP) with RAID-6 using IOR
Comparing Dynamic Disk Pools (DDP) with RAID-6 using IOR December, 2012 Peter McGonigal petermc@sgi.com Abstract Dynamic Disk Pools (DDP) offer an exciting new approach to traditional RAID sets by substantially
More informationCurrent Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
More informationHETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
More information64-Bit versus 32-Bit CPUs in Scientific Computing
64-Bit versus 32-Bit CPUs in Scientific Computing Axel Kohlmeyer Lehrstuhl für Theoretische Chemie Ruhr-Universität Bochum March 2004 1/25 Outline 64-Bit and 32-Bit CPU Examples
More informationSR-IOV In High Performance Computing
SR-IOV In High Performance Computing Hoot Thompson & Dan Duffy NASA Goddard Space Flight Center Greenbelt, MD 20771 hoot@ptpnow.com daniel.q.duffy@nasa.gov www.nccs.nasa.gov Focus on the research side
More informationScaling Objectivity Database Performance with Panasas Scale-Out NAS Storage
White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage
More informationMSC - Scientific Computing Facility (SCF) Supercomputer Status
MSC - Scientific Computing Facility (SCF) Supercomputer Status RFP/Contract Status The formal SCF RFP was released on March 22, 2002 and closed June 10, 2002. The bid was competitive (more than one bid
More informationHPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect a.jackson@epcc.ed.ac.uk
HPC and Big Data EPCC The University of Edinburgh Adrian Jackson Technical Architect a.jackson@epcc.ed.ac.uk EPCC Facilities Technology Transfer European Projects HPC Research Visitor Programmes Training
More informationIntel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance
Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance Hybrid Storage Performance Gains for IOPS and Bandwidth Utilizing Colfax Servers and Enmotus FuzeDrive Software NVMe Hybrid
More informationHadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
More informationSupercomputing Resources in BSC, RES and PRACE
www.bsc.es Supercomputing Resources in BSC, RES and PRACE Sergi Girona, BSC-CNS Barcelona, 23 Septiembre 2015 ICTS 2014, un paso adelante para la RES Past RES members and resources BSC-CNS (MareNostrum)
More informationSAN TECHNICAL - DETAILS/ SPECIFICATIONS
SAN TECHNICAL - DETAILS/ SPECIFICATIONS Technical Details / Specifications for 25 -TB Usable capacity SAN Solution Item 1) SAN STORAGE HARDWARE : One No. S.N. Features Description Technical Compliance
More informationPerformance of the JMA NWP models on the PC cluster TSUBAME.
Performance of the JMA NWP models on the PC cluster TSUBAME. K.Takenouchi 1), S.Yokoi 1), T.Hara 1) *, T.Aoki 2), C.Muroi 1), K.Aranami 1), K.Iwamura 1), Y.Aikawa 1) 1) Japan Meteorological Agency (JMA)
More informationEnabling Technologies for Distributed Computing
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies
More informationJob Scheduling on a Large UV 1000. Chad Vizino SGI User Group Conference May 2011. 2011 Pittsburgh Supercomputing Center
Job Scheduling on a Large UV 1000 Chad Vizino SGI User Group Conference May 2011 Overview About PSC s UV 1000 Simon UV Distinctives UV Operational issues Conclusion PSC s UV 1000 - Blacklight Blacklight
More informationComputational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar
Computational infrastructure for NGS data analysis José Carbonell Caballero Pablo Escobar Computational infrastructure for NGS Cluster definition: A computer cluster is a group of linked computers, working
More informationSGI High Performance Computing
SGI High Performance Computing Accelerate time to discovery, innovation, and profitability 2014 SGI SGI Company Proprietary 1 Typical Use Cases for SGI HPC Products Large scale-out, distributed memory
More informationRecommended hardware system configurations for ANSYS users
Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range
More informationCluster Computing at HRI
Cluster Computing at HRI J.S.Bagla Harish-Chandra Research Institute, Chhatnag Road, Jhunsi, Allahabad 211019. E-mail: jasjeet@mri.ernet.in 1 Introduction and some local history High performance computing
More informationPower Efficiency Metrics for the Top500. Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab
Power Efficiency Metrics for the Top500 Shoaib Kamil and John Shalf CRD/NERSC Lawrence Berkeley National Lab Power for Single Processors HPC Concurrency on the Rise Total # of Processors in Top15 350000
More informationMicrosoft Compute Clusters in High Performance Technical Computing. Björn Tromsdorf, HPC Product Manager, Microsoft Corporation
Microsoft Compute Clusters in High Performance Technical Computing Björn Tromsdorf, HPC Product Manager, Microsoft Corporation Flexible and efficient job scheduling via Windows CCS has allowed more of
More informationPedraforca: ARM + GPU prototype
www.bsc.es Pedraforca: ARM + GPU prototype Filippo Mantovani Workshop on exascale and PRACE prototypes Barcelona, 20 May 2014 Overview Goals: Test the performance, scalability, and energy efficiency of
More informationCommoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre
Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor
More informationMulti-Threading Performance on Commodity Multi-Core Processors
Multi-Threading Performance on Commodity Multi-Core Processors Jie Chen and William Watson III Scientific Computing Group Jefferson Lab 12000 Jefferson Ave. Newport News, VA 23606 Organization Introduction
More informationPADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute
PADS GPFS Filesystem: Crash Root Cause Analysis Computation Institute Argonne National Laboratory Table of Contents Purpose 1 Terminology 2 Infrastructure 4 Timeline of Events 5 Background 5 Corruption
More informationNetwork Bandwidth Measurements and Ratio Analysis with the HPC Challenge Benchmark Suite (HPCC)
Proceedings, EuroPVM/MPI 2005, Sep. 18-21, Sorrento, Italy, LNCS, Springer-Verlag, 2005. c Springer-Verlag, http://www.springer.de/comp/lncs/index.html Network Bandwidth Measurements and Ratio Analysis
More informationA Crash course to (The) Bighouse
A Crash course to (The) Bighouse Brock Palen brockp@umich.edu SVTI Users meeting Sep 20th Outline 1 Resources Configuration Hardware 2 Architecture ccnuma Altix 4700 Brick 3 Software Packaged Software
More informationInterconnect Efficiency of Tyan PSC T-630 with Microsoft Compute Cluster Server 2003
Interconnect Efficiency of Tyan PSC T-630 with Microsoft Compute Cluster Server 2003 Josef Pelikán Charles University in Prague, KSVI Department, Josef.Pelikan@mff.cuni.cz Abstract 1 Interconnect quality
More informationUpgrading Small Business Client and Server Infrastructure E-LEET Solutions. E-LEET Solutions is an information technology consulting firm
Thank you for considering E-LEET Solutions! E-LEET Solutions is an information technology consulting firm that specializes in low-cost high-performance computing solutions. This document was written as
More informationLS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
More information