CHAMELEON: A LARGE-SCALE, RECONFIGURABLE EXPERIMENTAL ENVIRONMENT FOR CLOUD RESEARCH



Similar documents
Presented By Joe Mambretti, Director, International Center for Advanced Internet Research, Northwestern University

Enabling Large-Scale Testing of IaaS Cloud Platforms on the Grid 5000 Testbed

The path to the cloud training

Characterizing Task Usage Shapes in Google s Compute Clusters

The path to the cloud training

Big Data and Cloud Computing for GHRSST

Accelerate POC to Production: OpenStack with FlexPod

High Availability Databases based on Oracle 10g RAC on Linux

FREE AND OPEN SOURCE SOFTWARE FOR CLOUD COMPUTING SERENA SPINOSO FULVIO VALENZA

IronPOD Piston OpenStack Cloud System Commodity Cloud IaaS Platforms for Enterprises & Service

Leveraging OpenStack Private Clouds

High Performance Computing OpenStack Options. September 22, 2015

AMD SEAMICRO OPENSTACK BLUEPRINTS CLOUD- IN- A- BOX OCTOBER 2013

Network Virtualization

HPC Cloud Computing Guide PENGUIN ( ) HPC

Network for Sustainable Ultrascale Computing (NESUS)

Private cloud computing advances

Next Generation Clouds, The Chameleon Cloud Testbed, and Software Defined Networking (SDN)

2) Xen Hypervisor 3) UEC

ANI Network Testbed Update

Network-Aware Scheduling of MapReduce Framework on Distributed Clusters over High Speed Networks

Workflow Allocations and Scheduling on IaaS Platforms, from Theory to Practice

The Massachusetts Open Cloud (MOC)

Cisco Data Center Optimization Services

Datacenters and Cloud Computing. Jia Rao Assistant Professor in CS

S06: Open-Source Stack for Cloud Computing

OpenStack Networking: Where to Next?

Cisco Intelligent Automation for Cloud

GENI Network Virtualization Concepts

DevOps. Josh Preston Solutions Architect Stardate

Status and Evolution of ATLAS Workload Management System PanDA

Designing and Experimenting with Data Center Architectures. Aditya Akella UW-Madison

Flauncher and DVMS Deploying and Scheduling Thousands of Virtual Machines on Hundreds of Nodes Distributed Geographically

The Virtualization Practice

OpenStack Introduction. November 4, 2015

GPU Accelerated Signal Processing in OpenStack. John Paul Walters. Computer Scien5st, USC Informa5on Sciences Ins5tute

Chapter 7: Distributed Systems: Warehouse-Scale Computing. Fall 2011 Jussi Kangasharju

Big Workflow: More than Just Intelligent Workload Management for Big Data

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions

A Complete Open Cloud Storage, Virt, IaaS, PaaS. Dave Neary Open Source and Standards, Red Hat

Cluster, Grid, Cloud Concepts

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

CUDA in the Cloud Enabling HPC Workloads in OpenStack With special thanks to Andrew Younge (Indiana Univ.) and Massimo Bernaschi (IAC-CNR)

Infrastructure as a Service (IaaS)

THE JOURNEY TO COMPOSABLE INFRASTRUCTURE CISCO TAKES AN EVOLUTIONARY APPROACH WITH UCS M-SERIES & C3260 COMBINED WITH UCS MANAGEMENT FRAMEWORK

CERN Cloud Storage Evaluation Geoffray Adde, Dirk Duellmann, Maitane Zotes CERN IT

High Performance OpenStack Cloud. Eli Karpilovski Cloud Advisory Council Chairman

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

PISTON CLOUDOS WITH OPENSTACK: TURN-KEY WEB-SCALE INFRASTRUCTURE SOFTWARE. Easy. CloudOS Compendium TECHNICAL WHITEPAPER

Private Cloud Management

The National Consortium for Data Science (NCDS)

Future Cities Bristol is Open

1. Implementation of a testbed for testing Energy Efficiency by server consolidation using Vmware

Server Consolidation with SQL Server 2008

How To Build A Cloud Server For A Large Company

CloudStack and Big Data. Sebastien May 22nd 2013 LinuxTag, Berlin

How To Use Openstack At Cern

9/26/2011. What is Virtualization? What are the different types of virtualization.

Oracle Infrastructure Systems Management with Enterprise Manager and Ops Center CON4954

Building Storage as a Service with OpenStack. Greg Elkinbard Senior Technical Director

KVM, OpenStack, and the Open Cloud

Bright Cluster Manager

Experiments with Complex Scientific Applications on Hybrid Cloud Infrastructures

High-Performance, Low-Cost Computational Chemistry: Servers in a Stick, Box, and Cloud. Nathan Vance Polik Group Hope College February 19, 2015

Nimbus: Cloud Computing with Science

ServerCentral Cloud Services Reliable. Adaptable. Robust.

New Jersey Big Data Alliance

Before we can talk about virtualization security, we need to delineate the differences between the

Networking Virtualization Using FPGAs

STeP-IN SUMMIT June 18 21, 2013 at Bangalore, INDIA. Performance Testing of an IAAS Cloud Software (A CloudStack Use Case)

Shareable Private Space on a Public Cloud

Transcription:

CHAMELEON: A LARGE-SCALE, RECONFIGURABLE EXPERIMENTAL ENVIRONMENT FOR CLOUD RESEARCH Principal Investigator: Kate Keahey Co-PIs: J. Mambretti, D.K. Panda, P. Rad, W. Smith, D. Stanzione MARCH 1 1, 2015 1

CHAMELEON: A POWERFUL AND FLEXIBLE EXPERIMENTAL INSTRUMENT Large-scale instrument Targeting Big Data, Big Compute, Big Instrument research ~650 nodes (~14,500 cores), 5 PB disk over two sites, 2 sites connected with 100G network Reconfigurable instrument Bare metal reconfiguration, operated as single instrument, graduated approach for ease-of-use Connected instrument Workload and Trace Archive Partnerships with production clouds: CERN, OSDC, Rackspace, Google, and others Partnerships with users Complementary instrument Complementing GENI, Grid 5000, and other testbeds Sustainable instrument Industry connections

CHAMELEON HARDWARE SCUs connect to core and fully connected to each other Switch Standard Cloud Unit 42 compute 4 storage x2 Switch Standard Cloud Unit 42 compute 4 storage x10 Core Services Front End and Data Mover Nodes Chameleon Core Network 100Gbps uplink public network (each site) Core Services 3 PB Central File Systems, Front End and Data Movers To UTSA, GENI, Future Partners 504 x86 Compute Servers 48 Dist. Storage Servers 102 Heterogeneous Servers 16 Mgt and Storage Nodes Chicago Austin Heterogeneous Cloud Units Alternate Processors and Networks

CAPABILITIES AND SUPPORTED RESEARCH Development of new models, algorithms, platforms, auto-scaling HA, etc., innovative application and educational uses Persistent, reliable, shared clouds Repeatable experiments in new models, algorithms, platforms, auto-scaling, high-availability, cloud federation, etc. Isolated partition, pre-configured images reconfiguration Virtualization technology (e.g., SR-IOV, accelerators), systems, networking, infrastructure-level resource management, etc. Isolated partition, full bare metal reconfiguration

SOFTWARE: CORE CAPABILITIES Persistent Clouds OpenStack User-Deployed Clouds Pre-configured Image Catalog Bare metal images Provisioning, Network, Scheduling and Orchestration Linux Operating System Framework (LosF), (TACC) KaDeploy, KaVLAN, OAR2, (Grid 5000) Ironic, Neuron, OnMetal (OpenStack, Rackspace) Orchestration: Nimbus, Interactive Experiment Management

EXPERIMENT WORKFLOW User interface: log in, manage profile Find Resources Machine-parsable description (JSON) Versioning (hardware upgrades, etc.) Verification (maintenance, failures, etc.) Reserve Resources (browsing vs matching) Reconfigure testbed Shape experimental conditions Monitoring and metrics Including fine-grain and energy monitoring Integration with workload generators, simulation, etc.

OUTREACH AND ENGAGEMENT Early User Program Committed users, driving and testing new capabilities, enhanced level of support Chameleon Workshop Annual workshop to inform, share experimental techniques solutions and platforms, discuss upcoming requirements, and showcase research Advisory Bodies Research Steering Committee: advise on capabilities needed to investigate upcoming research challenges Industry Advisory Board: provide synergy between industry and academia

PROJECT SCHEDULE Fall 2014: FutureGrid resources at UC and TACC available as OpenStack clouds Spring 2015: Initial bare metal reconfiguration capabilities available on FutureGrid UC&TACC resources for Early Users Summer 2015: New hardware: large-scale homogenous partitions available to Early Users Fall 2015: Large-scale homogenous partitions and bare metal reconfiguration generally available 2015/2016: Refinements to experiment management capabilities Fall 2016: Heterogeneous hardware available

TEAM Kate Keahey Chameleon PI Science Director Paul Rad Industry Liason Joe Mambretti Programmable networks Warren Smith Director of Operations DK Panda High-performance networks Dan Stanzione Facilities Director

PARTING THOUGHTS Large-scale, responsive experimental testbed Targeting critical research problems at scale Evolve with the community input Reconfigurable environment Support use cases from bare metal to production clouds Support for repeatable and reproducible experiments One-stop shopping for experimental needs Trace and Workload Archive, user contributions, requirement discussions Engage the community Network of partnerships and connections with scientific production testbeds and industry Partnerships with existing experimental testbeds Outreach activities Come visit us at www.chameleoncloud.org!