BSC - Barcelona Supercomputer Center



Similar documents
MareNostrum: Building and running the system - Lisbon, August 29th, 2005

Supercomputing Status und Trends (Conference Report) Peter Wegner

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

Spanish Supercomputing Network


HPC Update: Engagement Model

OpenMP Programming on ScaleMP

David Vicente Head of User Support BSC

Supercomputing Resources in BSC, RES and PRACE

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)

RES is a distributed infrastructure of Spanish HPC systems. The objective is to provide a unique service to HPC users in Spain

Building a Top500-class Supercomputing Cluster at LNS-BUAP

Current Trend of Supercomputer Architecture

Building a Linux Cluster

Hadoop on the Gordon Data Intensive Cluster

A Scalable Ethernet Clos-Switch

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Ethernet in the World s Top500 Supercomputers

Cluster Computing in a College of Criminal Justice

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

Trends in High-Performance Computing for Power Grid Applications

Altix Usage and Application Programming. Welcome and Introduction

Building Clusters for Gromacs and other HPC applications

Grid on Blades. Basil Smith 7/2/ IBM Corporation

Sun Constellation System: The Open Petascale Computing Architecture

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

MSC - Scientific Computing Facility (SCF) Supercomputer Status

CORRIGENDUM TO TENDER FOR HIGH PERFORMANCE SERVER

Jeff Wolf Deputy Director HPC Innovation Center

PRIMERGY server-based High Performance Computing solutions

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Clusters: Mainstream Technology for CAE

Energy efficient computing on Embedded and Mobile devices. Nikola Rajovic, Nikola Puzovic, Lluis Vilanova, Carlos Villavieja, Alex Ramirez

A-CLASS The rack-level supercomputer platform with hot-water cooling

INDIAN INSTITUTE OF TECHNOLOGY KANPUR Department of Mechanical Engineering

How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications

Linux Cluster Computing An Administrator s Perspective

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

HPC-related R&D in 863 Program

Thematic Unit of Excellence on Computational Materials Science Solid State and Structural Chemistry Unit, Indian Institute of Science

Mississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC

Purchase of High Performance Computing (HPC) Central Compute Resources by Northwestern Researchers

InfiniBand Strengthens Leadership as the High-Speed Interconnect Of Choice

Performance Characteristics of Large SMP Machines

1 Bull, 2011 Bull Extreme Computing

Lecture 2 Parallel Programming Platforms

Cisco SmartPlay Select. Cisco Global Data Center Promotional Program

Das TOP500-Projekt der Universitäten Mannheim und Tennessee zur Evaluierung des Supercomputer Marktes. Hans-Werner Meuer Universität Mannheim

Monitoring Infrastructure for Superclusters: Experiences at MareNostrum

Parallel Computing. Introduction

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

ANALYSIS OF SUPERCOMPUTER DESIGN

Smart Campus Management with Cloud Services

Cosmological simulations on High Performance Computers

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Remote Visualization and Collaborative Design for CAE Applications

MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit. Prof. Dieter Kranzlmüller (LRZ)

HPC Growing Pains. Lessons learned from building a Top500 supercomputer

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

An Overview of High- Performance Computing and Challenges for the Future

Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer

Cluster Computing at HRI

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

A Smart Investment for Flexible, Modular and Scalable Blade Architecture Designed for High-Performance Computing.

CLUSTER APPROACH TO HIGH PERFORMANCE COMPUTING

Lecture 1: the anatomy of a supercomputer

Algorithms of Scientific Computing II

Evoluzione dell Infrastruttura di Calcolo e Data Analytics per la ricerca

CENTRO DE SUPERCOMPUTACIÓN GALICIA CESGA

Comparing the performance of the Landmark Nexus reservoir simulator on HP servers

JuRoPA. Jülich Research on Petaflop Architecture. One Year on. Hugo R. Falter, COO Lee J Porter, Engineering

The TOP500 Project: Looking Back over 15 Years of Supercomputing Experience

Performance Engineering of the Community Atmosphere Model

A Flexible Cluster Infrastructure for Systems Research and Software Development

Cray XT3 Supercomputer Scalable by Design CRAY XT3 DATASHEET

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

Martin County Administration. Information Technology Services. Proposal For. Storage Area Network Systems. Supporting WinTel Server Consolidation

Secure File Sharing SRFS on Ether with host-to-host IPSec connection over the Pacific Ocean link

Software and Performance Engineering of the Community Atmosphere Model

Shared Parallel File System

A GPU COMPUTING PLATFORM (SAGA) AND A CFD CODE ON GPU FOR AEROSPACE APPLICATIONS

InfiniBand Experiences at Forschungszentrum Karlsruhe. Forschungszentrum Karlsruhe

QuickSpecs. HP Integrity cx2620 Server. Overview

Welcome to the. Jülich Supercomputing Centre. D. Rohe and N. Attig Jülich Supercomputing Centre (JSC), Forschungszentrum Jülich

High-Performance Computing and Big Data Challenge

RLX Technologies Server Blades

Performance Monitoring of Parallel Scientific Applications

Performance of the JMA NWP models on the PC cluster TSUBAME.

Cluster Grid Interconects. Tony Kay Chief Architect Enterprise Grid and Networking

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC. Wenhao Wu Program Manager Windows HPC team

SRNWP Workshop. HP Solutions and Activities in Climate & Weather Research. Michael Riedmann European Performance Center

The CNMS Computer Cluster

PCIe Over Cable Provides Greater Performance for Less Cost for High Performance Computing (HPC) Clusters. from One Stop Systems (OSS)

10- High Performance Compu5ng

Performance of HPC Applications on the Amazon Web Services Cloud

High Performance Computing Infrastructure in JAPAN

High Performance Computing and the Computational Grid. Technology Trends: Microprocessor Capacity

Parallel Programming Survey

Introduction to High Performance Cluster Computing. Cluster Training for UCL Part 1

Transcription:

Objectives Research in Supercomputing and Computer Architecture Collaborate in R&D e-science projects with prestigious scientific teams Manage BSC supercomputers to accelerate relevant contributions to research areas where intensive computing is an enabling technology Structure Presidencia CONSEJO RECTOR Comisión Científico Asesora Comisión Ejecutiva Patronato Dirección Comisión de Acceso Director de Investigación en Tecnologías de la Información Director de e-ciencia Dirección de Gestión Herramientas de Rendimiento Programación Paralela Métodos Numéricos Grids y Clusters Biología Computacional Química Computacional Física Computacional Medicina basada en la Información Gestión del Superordenador Desarrollo de Negocio Gestión Administrativo-Financiera Plataformas e-business Arquitectura de Ordenadores

IT research and development projects Continuation of CEPBA (European Center for Parallelism in Barcelona) research lines Deep Computing Performance Tools Parallel programing Grid Code Optimization Computer Architecture Vector processors Network processors e-science projects R&D collaborations Computational Biology Computational Chemistry Computational Physics Information based Medicine

Management projects Supercomputer Management System Administration Users support Business Development External Relations Promotion Technology Transfer Education Administration Accounting and Finances Human Resources MareNostrum / Iberus PowerPC 970 FX processors (dual processors) 4GB ECC 333 DDR memory per node 3 networks Myrinet Gigabit 10/00 Ethernet Diskless network support Linux cluster

MareNostrum: Overall system description 27 Compute Racks (RC01-RC27) 162 BC chassis w/opm and gigabit ether switch 2268 JS20+ nodes w/myrinet daughter card 4 Myrinet Racks (RM01-RM04) 10 clos256+256 myrinet switches 2 Myrinet spines 1280s 7 Storage Server Racks (RS01-RS07) 40 p615 storage servers 6/rack 20 FastT 100 3/rack 20 EXP100 3/rack 1 Operations Rack (RH01) 7316-TF3 display 2 p615 mgmt nodes 2 HMC model 7315-CR2 3 Remote Async Nodes 3 Cisco 3550 1 BC chassis (BCIO) 1 Gigabit Network Racks 1 Force10 E600 for Gb network 4 Cisco 3550 48-port for 10/100 network Processor: PowerPC 970FX

Blades, blade center and blade center racks JS20 Processor Blade 2-way 2.2 GHz Power PC 970 SMP 4GB memory (512KB L2 cache) Local IDE drive (40 GB) 2x1Gb Ethernet on board Myrinet daughter card 6 chassis in a rack (42U) 168 processors 336GB memory 14 blades per chassis (7U) 28 processors 56GB memory Gigabit ethernet switch MareNostrum: Floor plan with cabling 27 racks + 1 BC chassis 4564 processors 9TB memory Blade centers Myrinet racks Storage servers Operations rack Gigabit switch 10/100 switches

Blade centers Myrinet racks Storage servers Operations rack Gigabit switch 10/100 switches 256 blades connected to 1 clos 256+256 Myrinet 1280 blades connected to 5 clos 256+256 Myrinet and 1 spine 1280 2282 blades connected to 10 clos 256+256 Myrinet and 2 spine 1280 20 x 7TB storage nodes Management rack, Force 10 Gigabit, 10/100 Cisco switches

Blade center racks 6 s per rack 27 racks + 1 Cabling per rack 84 fiber cables to myrinet switch 6 Gb to Force10 E600 6 10/100 cat5 to Cisco Myrinet racks 10 switches Interconnect up to 256 Blades Connect to Spine (64 ports) 2 Spine 1280 Interconnect up to 10 switches Monitoring using 10/100 connection

Myrinet racks Spine 1280 Spine 1280 64 to Spine Spine 1280 256 Blades or Storage servers 320 from Clos Gb Subsystem: Force 10 E600 Interconnection of s Used for system boot of every blade center 212 internal network cables 170 for blades 42 for p615 76 connection available to external connection

Storage nodes Total of 20 storage nodes, 20 x 7 TBytes Each storage node 2xP615 FastT100 EXP100 Cabling per node 2 Myrinet 2 Gb to Force10 E600 2 10/100 cat5 to Cisco 1 Serial P615 P615 FastT100 EXP100 P615 P615 FastT100 EXP100 P615 P615 FastT100 EXP100 Management rack Contents BCIO Display 2 HMC 2 x p615 3 x Cisco 3 x 16-port Remote Async Nodes

Scientific Applications M B E 10 Y 12 M T O E R S 10 Y 10 = Recent Computations by NSF Grand Challenge Research = Next Step Projections by NSF Grand Challenge Research Teams = Long Range Projections from Recent Applications Workshop 10 14 10 8 NSF in 2004 (Projected) 2000 NSF Leading Edge 10 8 ASCI in 2004 1995 NSF Capability Turbulent Convection in Stars QCD Computational Cosmology 100 year climate model in hours Atomic/Diatomic Interaction Molecular Dynamics for Biological Molecules 10 10 10 12 10 14 10 16 10 18 10 20 MACHINE REQUIREMENT IN FLOPS Source: Bob Voigt, NSF Top500 - Performance 1 Pflop/s 100 Tflop/s 10 Tflop/s 1 Tflop/s 100 Gflop/s 10 Gflop/s 1 Gflop/s jun-93 nov-93 jun-94 nov-94 100 Mflop/s 1.17 TF/s 59.7 GF/s 0.4 GF/s Fujitsu 'NWT' NAL jun-95 nov-95 jun-96 nov-96 SUM N=1 N=500 jun-97 nov-97 jun-98 nov-98 Intel ASCI Red Sandia jun-99 nov-99 jun-00 nov-00 IBM ASCI White LLNL jun-01 nov-01 jun-02 220 TF/s 35.8 TF/s NEC Earth Simulator 134 GF/s!"#$%&'(&

Top 500 list, November 2004 (upcoming) Rank Site Country/year Computer/processors Manufacturer Rmax Rpeak 1 NASA Ames United States/2004 Project Columbia 16x504proc SGI Altix 3000 1.5GHz Itanium2 w/ Infiniband / 8064 SGI 42707 48384 2 IBM Thomas Watson Research Center United States/2004 Blue Gene/L DD2 Prototype cluster 0.7 GHz PowerPC 440 w/custom network / 16384 IBM 36010 45875 3 Earth Simulator Japan/2002 Earth Simulator/5120 NEC 35860 40960 4 BSC Spain/2004 MareNostrum JS20+ PowerPC 970 2.2GHz w. Myrinet/ 3564 IBM 20530 31363 5 Lawrence Livermore National Laboratory United States 2004 Thunder Intel Itanium 2 Tiger4 1.4 GHz Quadrics / 4096 California Digital Corporation 19940 22938 MareNostrum: number comparison Project Columbia BlueGene /L Earth Simulator MareNostrum Number of racks >64 640 40 Number of processors 10240 128k 5120 4500 System peak performance 65 TF 360 TF 41 TF 40 TF Total system Memory 16 TB 10 TB 9 TB Power < 1.6 MW 5.1 MW 600 KW Floor space 400 m 2 3250 m 2 120 m 2 Bisectional bandwidth 0.7 TB/s 7.8 TB/s 0.6 TB/s Number of cables 5000 83200 2600

Thank you! Prof. Mateo Valero BSC Director