How to Deploy OpenStack on TH-2 Supercomputer Yusong Tan, Bao Li National Supercomputing Center in Guangzhou April 10, 2014

Similar documents
AMD SEAMICRO OPENSTACK BLUEPRINTS CLOUD- IN- A- BOX OCTOBER 2013

An Introduction to OpenStack and its use of KVM. Daniel P. Berrangé

Ubuntu OpenStack on VMware vsphere: A reference architecture for deploying OpenStack while limiting changes to existing infrastructure

Release Notes for Fuel and Fuel Web Version 3.0.1

Cloud on TEIN Part I: OpenStack Cloud Deployment. Vasinee Siripoonya Electronic Government Agency of Thailand Kasidit Chanchio Thammasat University

SUSE Cloud 2.0. Pete Chadwick. Douglas Jarvis. Senior Product Manager Product Marketing Manager

Hadoop on OpenStack Cloud. Dmitry Mescheryakov Software

Using SUSE Cloud to Orchestrate Multiple Hypervisors and Storage at ADP

การใช งานและต ดต งระบบ OpenStack ซอฟต แวร สาหร บบร หารจ ดการ Cloud Computing เบ องต น

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman

Ubuntu OpenStack Fundamentals Training

Bright Cluster Manager

Multi Provider Cloud. Srinivasa Acharya, Engineering Manager, Hewlett-Packard

Running an OpenStack Cloud for several years and living to tell the tale. Alexandre Maumené Gaëtan Trellu Tokyo Summit, November 2015

Introducing ScienceCloud

FIA Athens 2014 ~OKEANOS: A LARGE EUROPEAN PUBLIC CLOUD BASED ON SYNNEFO. VANGELIS KOUKIS, TECHNICAL LEAD, ~OKEANOS

Savanna Hadoop on. OpenStack. Savanna Technical Lead

OpenStack: we drink our own Champagne. Teun Docter Software developer

OpenStack Assessment : Profiling & Tracing

Establishing Scientific Computing Clouds on Limited Resources using OpenStack

Achieving a High-Performance Virtual Network Infrastructure with PLUMgrid IO Visor & Mellanox ConnectX -3 Pro

Cloud Computing through Virtualization and HPC technologies

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012

FPGA Accelerator Virtualization in an OpenPOWER cloud. Fei Chen, Yonghua Lin IBM China Research Lab

Openstack. Cloud computing with Openstack. Saverio Proto

VSM Introduction & Major Features

One-click Hadoop Cluster Deployment on OpenPOWER Systems Pradeep K Surisetty IBM. #OpenPOWERSummit

Change the Game with HP Helion

How OpenStack is implemented at GMO Public Cloud service

HGST Virident Solutions 2.0

TUT5605: Deploying an elastic Hadoop cluster Alejandro Bonilla

Mirantis

What is the real cost of Commercial Cloud provisioning? Thursday, 20 June 13 Lukasz Kreczko - DICE 1

OpenStack The State of the Stack

How To Install Openstack On Ubuntu (Amd64)

Mobile Cloud Computing T Open Source IaaS

Déployer son propre cloud avec OpenStack. GULL François Deppierraz

OpenStack Introduction. November 4, 2015

Release Notes for Contrail Release 1.20

SUSE Cloud Installation: Best Practices Using a SMT, Xen and Ceph Storage Environment

rackspace.com/cloud/private

Installing Hadoop over Ceph, Using High Performance Networking

RED HAT INFRASTRUCTURE AS A SERVICE OVERVIEW AND ROADMAP. Andrew Cathrow Red Hat, Inc. Wednesday, June 12, 2013

SUSE Cloud 5 Private Cloud based on OpenStack

SURFsara HPC Cloud Workshop

IronPOD Piston OpenStack Cloud System Commodity Cloud IaaS Platforms for Enterprises & Service

How To Build A Cloud Server For A Large Company

Enabling Technologies for Distributed Computing

Big Fast Data Hadoop acceleration with Flash. June 2013

Hadoop: Embracing future hardware

What s new in Hyper-V 2012 R2

SYNNEFO: A COMPLETE CLOUD PLATFORM OVER GOOGLE GANETI WITH OPENSTACK APIs VANGELIS KOUKIS, TECH LEAD, SYNNEFO

version 7.0 Planning Guide

ONS Methodology Working Paper Series No 1. ONS Innovation Laboratories

vrealize Operations Management Pack for OpenStack

State of the Art Cloud Infrastructure

Building All-Flash Software Defined Storages for Datacenters. Ji Hyuck Yun Storage Tech. Lab SK Telecom

Building Storage as a Service with OpenStack. Greg Elkinbard Senior Technical Director

SUSE OpenStack Cloud 4 Private Cloud Platform based on OpenStack. Gábor Nyers Sales gnyers@suse.com

cloud functionality: advantages and Disadvantages

HP OpenStack & Automation

Huawei and Open Source. Industry development department Shi Hao

Analysis of VDI Storage Performance During Bootstorm

CERN Cloud Architecture

Cisco s Massively Scalable Data Center. Network Fabric for Warehouse Scale Computer

STeP-IN SUMMIT June 18 21, 2013 at Bangalore, INDIA. Performance Testing of an IAAS Cloud Software (A CloudStack Use Case)

NephOS A Licensed End-to-end IaaS Cloud Software Stack for Enterprise or OEM On-premise Use.

Ceph Optimization on All Flash Storage

StorPool Distributed Storage Software Technical Overview

Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales

POSIX and Object Distributed Storage Systems

Planning an OpenStack PoC

Zadara Storage Cloud A

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

Operating an OpenStack Cloud

Wojciech Furmankiewicz Senior Solution Architect Red Hat CEE

An Intro to OpenStack. Ian Lawson Senior Solution Architect, Red Hat

Enabling Technologies for Distributed and Cloud Computing

IBM Spectrum Protect in the Cloud

Red Hat Enterprise Linux OpenStack Platform Update February 17, 2016

SR-IOV In High Performance Computing

Boost Database Performance with the Cisco UCS Storage Accelerator

CloudCIX Bootcamp. The essential IaaS getting started guide.

Accelerate Cloud Initiatives with Cisco UCS and Ubuntu OpenStack

SUSE Cloud. Deployment Guide. January 26, 2015

Enabling High performance Big Data platform with RDMA

7 Ways OpenStack Enables Automation & Agility for KVM Environments

Achieving Performance Isolation with Lightweight Co-Kernels

Deploying Baremetal Instances with OpenStack

Preparation Guide. How to prepare your environment for an OnApp Cloud v3.0 (beta) deployment.

Transcription:

How to Deploy OpenStack on TH-2 Supercomputer Yusong Tan, Bao Li National Supercomputing Center in Guangzhou April 10, 2014 2014 年 云 计 算 效 率 与 能 耗 暨 第 一 届 国 际 云 计 算 咨 询 委 员 会 中 国 高 峰 论 坛

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 2

Background TH-2 Supercomputer Sponsored by 863 High Tech. Program of Ministry of Science and Tech. of China, Government of Guangdong province and Government of Guangzhou city Built by National University of Defense Technology 3

Background No. 1 of Top 500 list in June & November 2013 Rpeak 54.9PFlops,Rmax 33.9PFlops 4

Background TH-2 Supercomputer Installed in National Supercomputer Center in Guangzhou (NSCC-GZ) Open platform for research and education To provide HPC service Public information infrastructure To accelerate the industry and economy More than HPC! 5

Background TH-2 Supercomputer Neo-heterogeneous architecture Xeon CPU & Xeon Phi @ X86 ISA Items Configuration Nodes 16000 Processors Interconnect Memory Storage Cabinets Cooling Performance 32000 Intel Xeon CPUs + 48000 Xeon Phis + 4096 FT CPUs Proprietary high-speed interconnection network TH-NI 1PB in total Global shared parallel storage system, 12.4PB 162=125+13+24 compute/communication/storage Cabinets Closed Air cooling system Peak performance is 54.9PFlops Max performance is 33.9PFlops 6

Background TH-2 Supercomputer How to be used as a public information infrastructure Different environments Redhat, CentOS, Windows, Ubuntu, SUSE Different scales From several nodes to thousands of nodes Virtualization based IaaS cloud computing! 7

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 8

Deployment Hardware 6400 nodes on TH-2 50 cabinets 128 nodes in each cabinet Each node: Intel Xeon Ivy Bridge (12 cores) * 2 64 GB RAM 1TB local disk GE Nic * 2 TH-NI High-speed Network (160 Gbps) Intel Xeon Phi * 3 9

Background Software Havana Server 12.04 LTS v0.67 v2.7.11 10

Deployment Architecture 256 controller nodes 2 cabinets 124 api nodes: nova-api, nova-cells, glance-*, cinder-api, cinder-scheduler, keystone 4 LVS+Keepalived 96 network nodes 32 Ganglia+Nagios server nodes 3840 compute nodes 30 cells, each for one cabinet 2 cell servers in each cabinet 11

Deployment Network Topology 1 GE for management 1 GE for VMs communication TH-NI for Ceph Storage Take advantage of the speed and bandwidth of TH-NI 12

Deployment Storage Ceph RBD as all-in-one storage RBD for Glance image storage RBD for volume backend RBD for instance storage 13

Deployment Puppet Different manifests for different roles Each node reads its role from a predefined configuration file 14

Deployment Ubuntu Server 12.04.2 LTS based diskless system Using TH-NI to pull images to accelerate booting With native TH-NI RDMA driver support Partition hard disks when first boot SSH without password Decide IP and hostname based on TH-NI s nid dynamically Configure the role of each node using Puppet when startup 15

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 16

Optimization More nodes Load balance for *-apis Separated Keystone API service for Neutron 17

Optimization More workers Multi-processed Multi-processed neutron-server api_workers Multi-processed nova-api osapi_compute_workers metadata_workers ec2_workers Multi-processed nova-conductor workers Multi-processed glance-api workers 18

Optimization More workers Apache hosting Keystone puppet supported by its design and implementation 19

Optimization More powerful Nova Eliminate api rate limit api_rate_limit=false Use large db pool size sql_dbpool_enable sql_min_pool_size sql_max_pool_size sql_max_overflow 20

Optimization More powerful Database Use tmpfs as the storage Nova, Glance Use memcached as the direct storage Keystone 21

Optimization More powerful Rabbitmq Set high memory watermark rabbitmqctl set_vm_memory_high_watermark 0.6 Set high socket limit ulimit -n 102400 Rabbitmq Cluster Clustering rabbitmq by running more rabbitmq-servers to distribute workload 22

Optimization More powerful MySQL Use larger maximum connections max_connections=102400 Galera Mysql cluster 23

Optimization More powerful KVM Huge pages enabled Using VhostNet provides better latency and greater throughput Using virtio_blk Get higher performance of storage devices Kernel SamePage merging enabled(ksm) Because we want to oversell Kernel I/O scheduler to "Deadline" 24

Optimization Services HA/LB LVS + keepalived for all API servers Increased rpc_response_timeout Increased quantum_url_timeout 25

Optimization HA for MySQL Using corosync and pacemaker to provide HA DB can be accessed via multiple VIPs Each component connects to a seperate HA VIP according to workload 26

Optimization HA for instances 27

Optimization Ceph Inline data support for small file IO Typical file read/write traffic Client consult MDS for file metadata Client communicate with OSD for file data With inline data support, for small files Client consult MDS for both 28

Optimization Ceph Punch hole support Enable sparse file support Improves space efficiency in virtualization situation 29

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 30

Preliminary Evaluation Maximum VMs started simultaneously cirros test image, m1.tiny ~5300 scheduler cost most time Concurrent request handled by api server per second ~1100 list instances operations ( active instances num. =1000) Bottlenecks GE, use TH-NI instead nova-api (if TH-NI is used), use more api servers Time spent on query of database only takes less 40% 31

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 32

Contributions to Community For Havana 15 commits, 3976 LoC For Icehouse 32 commits, 1078 LoC are merged (till 2014.4.1) 33

Contributions to Community Quota per-project-user-quotas-support The administrators can set quotas on a per-project-user basis, rather than just a per-project basis. editable-default-quotas-support The administrators can edit the default quota for nova or cinder through horizon or CLI. 34

Contributions to Community And More Add Multi-level User Management to Keystone via Project Nesting (in development) Multiple Level User Quota Management (in development) More bugfixes 35

Contributions to Community Ceph Two blueprints presented at the first Ceph Developer Summit and pushed into upstream Inline data support for small file IO Punch hole support 23 commits are merged to v0.77 The 3 rd contributor by organizations 36

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 37

Cases E-Gov for Guangzhou City Websites of Guangzhou government Requirements:Web/Application/Database Servers Efforts: Load Balance High Availability 38

Cases E-Gov for Guangzhou City E-Gov Data Management System Requirements:JBoss/WebLogic/Oracle Servers Efforts Oracle @ Ceph RBD VM Optimize for Oracle & Java 39

Contents Background Deployment Optimization Preliminary Evaluation Contributions to Community Cases Next Steps 40

Next Steps For NSCC-GZ More performance tests Performance of cell Neutron Automatic Operation for large scale system How to operate thousands of nodes Upgrade to Icehouse IaaS Services More E-Government applications in Guangzhou IaaS resource lease XaaS Service Database, Hadoop, HA and more.. 41

Next Steps For Community Automatic operation tools for large scale system Monitor Metering & Billing Elastic resource management QoS-based VM cluster scale-in/out Ceph Read ahead optimization SSD based writeback cache support 42

This is our Team! 43

Q&A 44