Executive Summary Introduction Nutanix Complete Cluster A Scalable Architecture Project Colossus The Setup...

Similar documents
DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Nutanix Complete Cluster Reference Architecture for Virtual Desktop Infrastructure

Nutanix Tech Note. Configuration Best Practices for Nutanix Storage with VMware vsphere

Virtualizing Microsoft SQL Server on Dell XC Series Web-scale Converged Appliances Powered by Nutanix Software. Dell XC Series Tech Note

PARALLELS CLOUD STORAGE

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array

Getting the Most Out of VMware Mirage with Hitachi Unified Storage and Hitachi NAS Platform WHITE PAPER

Benchmarking Hadoop & HBase on Violin

MaxDeploy Hyper- Converged Reference Architecture Solution Brief

SSDs: Practical Ways to Accelerate Virtual Servers

SQL Server Virtualization

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Deep Dive on SimpliVity s OmniStack A Technical Whitepaper

Use of Hadoop File System for Nuclear Physics Analyses in STAR

SSDs: Practical Ways to Accelerate Virtual Servers

The next step in Software-Defined Storage with Virtual SAN

Microsoft Private Cloud Fast Track

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

The Data Placement Challenge

MS EXCHANGE SERVER ACCELERATION IN VMWARE ENVIRONMENTS WITH SANRAD VXL

Performance in a Gluster System. Versions 3.1.x

How To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory)

Diablo and VMware TM powering SQL Server TM in Virtual SAN TM. A Diablo Technologies Whitepaper. May 2015

Delivering SDS simplicity and extreme performance

Comparison of Windows IaaS Environments

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

Integrating SSDs into Virtual Servers

Maxta Storage Platform Enterprise Storage Re-defined

Understanding Enterprise NAS

THE SUMMARY. ARKSERIES - pg. 3. ULTRASERIES - pg. 5. EXTREMESERIES - pg. 9

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Understanding Data Locality in VMware Virtual SAN

FlashSoft Software from SanDisk : Accelerating Virtual Infrastructures

Hadoop: Embracing future hardware

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

WHITE PAPER 1

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE

Reducing Storage TCO With Private Cloud Storage

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

HGST Virident Solutions 2.0

Scientific Computing Data Management Visions

Tableau Server 7.0 scalability

Hyperscale Use Cases for Scaling Out with Flash. David Olszewski

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

StarWind Virtual SAN for Microsoft SOFS

A Close Look at PCI Express SSDs. Shirish Jamthe Director of System Engineering Virident Systems, Inc. August 2011

Scaling from Datacenter to Client

High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software

Optimizing Web Infrastructure on Intel Architecture

HP Z Turbo Drive PCIe SSD

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

Business-centric Storage FUJITSU Hyperscale Storage System ETERNUS CD10000

Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

WHITE PAPER Guide to 50% Faster VMs No Hardware Required

Amazon EC2 Product Details Page 1 of 5

IOmark- VDI. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC b Test Report Date: 27, April

Benchmarking Cassandra on Violin

On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform

Condusiv s V-locity 4 Boosts Virtual Machine Performance Over 50% Without Additional Hardware

PARALLELS CLOUD SERVER

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

VMware Virtual SAN Design and Sizing Guide TECHNICAL MARKETING DOCUMENTATION V 1.0/MARCH 2014

Cloudmark Slays Spam with Fusion iomemory Solutions

Lab Validation Report

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

THESUMMARY. ARKSERIES - pg. 3. ULTRASERIES - pg. 5. EXTREMESERIES - pg. 9

VMware Virtual SAN Hardware Guidance. TECHNICAL MARKETING DOCUMENTATION v 1.0

DSS. Diskpool and cloud storage benchmarks used in IT-DSS. Data & Storage Services. Geoffray ADDE

All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at

Fusion iomemory iodrive PCIe Application Accelerator Performance Testing

VMware Virtual SAN 6.0 Performance

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment

Database Server Configuration Best Practices for Aras Innovator 10

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Generational Performance Comparison: Microsoft Azure s A- Series and D-Series. A Buyer's Lens Report by Anne Qingyang Liu

SECURE, ENTERPRISE FILE SYNC AND SHARE WITH EMC SYNCPLICITY UTILIZING EMC ISILON, EMC ATMOS, AND EMC VNX

Lab Evaluation of NetApp Hybrid Array with Flash Pool Technology

Big Fast Data Hadoop acceleration with Flash. June 2013

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture

A Holistic Model of the Energy-Efficiency of Hypervisors

TekSouth Fights US Air Force Data Center Sprawl with iomemory

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

VDI Without Compromise with SimpliVity OmniStack and Citrix XenDesktop

EMC SCALEIO OPERATION OVERVIEW

WHITE PAPER Guide to 50% Faster VMs No Hardware Required

Intel RAID RS25 Series Performance

Boost Database Performance with the Cisco UCS Storage Accelerator

White Paper. Educational. Measuring Storage Performance

Full and Para Virtualization

VMware Virtual SAN 6.2 with Virtual Desktop Infrastructure Workload

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

SQL Server Consolidation Using Cisco Unified Computing System and Microsoft Hyper-V

Down to earth report: A hyper-converged infrastructure for the rest of us from HP.

Data Center Solutions

Transcription:

Project Colossus

Table of Contents Executive Summary... 3 Introduction... 4 Nutanix Complete Cluster A Scalable Architecture... 5 Project Colossus... 7 The Setup... 7 Benchmark Results... 8 Conclusion... 12

EXECUTIVE SUMMARY Executive Summary Predictable scalability is a big challenge for virtualized datacenters today. Organizations are looking to design their next-generation datacenters with virtualization at the core of the architecture. But most of them are unsure of how their needs are going to evolve as they virtualize more and more applications. As part of the design exercise, they need to make critical sizing decisions upfront. They need to accurately predict the size of server and storage infrastructure they will need over three to five years and procure a right size system. This guesswork often results in scenarios where organizations have to either live with a system that s under-designed or do a forklift upgrade and buy a bigger box. Nutanix Complete Cluster solves this problem with its building Block approach for the virtualized datacenter. It can seamlessly scale from one Block to a large numbers of Blocks in an elastic manner. As more Blocks are added to the cluster, Nutanix SOCS (Scale-Out Converged Storage) continues to grow as a shared storage pool that s accessible to any VM running anywhere on the cluster. This takes the guesswork out of the datacenter design and allows organizations to grow as their needs grow, while providing a predictable increase in performance and cost. Project Colossus is a fifty-node benchmark cluster that demonstrated linear scalability of Nutanix Complete Cluster from four nodes (one Complete Block) to fifty nodes (12.5 Complete Blocks). In this benchmark, Colossus delivered 375,000 random write IOPs and 28 Gigabytes per second of sequential bandwidth. 3

INTRODUCTION Introduction Capacity planning is a one of the biggest challenges for most data center architects. A typical data center design tries to address a three to five year horizon, if not more. For this, the architects have to know (or guess) in advance what they are designing for. Some of the most common questions an architect tries to address are: How much storage capacity is needed How many virtual machines will they need to run How much performance (IOPs, bandwidth) will the storage system need to deliver Apart from these technical requirements, they also often have to estimate how much budget is available to build and support this infrastructure over the planning horizon. Most of this unnatural upfront guesswork is a result of the rigid architecture of traditional storage systems. Traditional vendors typically sell their systems in small, medium and large configurations but they often require forklift upgrades to move from one configuration to the next one. Most small systems offer one storage controller, medium configurations offer two to four controllers and large systems might typically offer four to eight storage controllers (a few might go higher). As the number of controllers increases, performance often doesn t increase linearly while costs may grow exponentially. This makes it very hard for architects to right size their systems without making huge investments upfront. This problem is further accentuated due to the rapid growth in number of virtual machines. In such a dynamic environment, the ability to plan for the future is reduced and capacity management becomes an even bigger challenge. What is required is a system that can grow incrementally as new VMs are created. The system should provide predictable performance growth as it scales and should not require a forklift upgrade to get higher performance and capacity. At the same time, the system should provide predictability in investments needed to grow the datacenter over the planning horizon. 4

NUTANIX COMPLETE CLUSTER A SCALABLE ARCHITECTURE Nutanix Complete Cluster A Scalable Architecture Nutanix Complete Cluster is designed to be a highly scalable system with a small unit of growth (one node at a time). The architecture has its roots in the distributed computing model that has been adopted by companies such Google, Facebook, Yahoo and Amazon to build their data centers that are highly scalable. The following key aspects of its architecture make Nutanix Complete Cluster highly scalable: No Single Master In the Nutanix architecture, no one node plays the role of a master. Most common distributed systems have a single master that acts as a central point of control for key system operations. This master ends up becoming a bottleneck as the system scales. In traditional storage arrays, storage controllers play the role of such a central authority and are the limiting factor in scaling the system. Even if a storage array has multiple controllers, such controllers have to synchronize their caches, making it hard to scale them efficiently. Medusa (Distributed Metadata Layer) Nutanix Complete Cluster stores its metadata (information on what is stored where) in a distributed metadata layer named Medusa. As more nodes are added to the cluster, the metadata layer also continues to scale. Medusa stores the metadata on PCIe-based iomemory from Fusion-io, making the metadata accesses faster without limiting the metadata to RAM, as is the case in traditional solutions. This allows the cluster to keep rich metadata for advanced functionality without the metadata layer becoming a bottleneck. Curator (Distributed Data Maintenance Service) In large-scale storage systems, data maintenance is often centralized within a few controllers and it can easily become a system bottleneck. As more data is modified and deleted, garbage collection adds significant overhead to the system. Also, data consistency checks start consuming more and more resources from the controllers as data sizes grow. 5

NUTANIX COMPLETE CLUSTER A SCALABLE ARCHITECTURE Nutanix has developed a distributed data maintenance framework called Curator for addressing these limitations. Curator is a MapReduce-based distributed service that enables Nutanix Complete Cluster to scale linearly. It executes background data management operations in a massively parallel manner. Such operations include data-tiering, garbage collection, consistency checks, data redistribution in case of failures, etc. HOTCache (Distributed High Performance Cache) Nutanix HOTCache is a distributed cache that provides high-performance to virtual machines. It s a software layer that is backed by high performance Fusionio iomemory. As VMs write data, the data is immediately written to HOTCache first and then flushed to the backend storage layer in the background. HOTCache is persistent and also highly available data is persistently stored on iomemory and also replicated across the cluster to ensure that there is no data loss even if nodes fail. 6

Project Colossus To demonstrate true scalability, Nutanix created a benchmark called Project Colossus. Project Colossus involved running performance tests on a single Block (a four-node cluster) and scaling it to a large number of nodes (fifty) to validate performance at large scale. The Setup Single Block Hardware The baseline setup involved a standard Nutanix Complete Block (four VMware hosts) with the following aggregate hardware resources: 8 Intel Xeon Processors (48 Cores) 1.28 TB of Fusion-io iomemory 1.20 TB of SATA SSD 20 TB of SATA hard disks (7200 RPM) 192 GB RAM 4X10 GbE and 8X1GbE network ports Colossus Hardware Colossus was a cluster with fifty nodes (ESX hosts) and had the following aggregate hardware resources: 100 Intel Xeon Processors (600 Cores) 16 TB of Fusion-io iomemory 15 TB of SATA SSD 250 TB of SATA hard disks (7200 RPM) 2.4 TB RAM 50X10 GbE and 100X1GbE network ports The nodes were connected to each other over a 10 GbE network in both cases. Software Setup Each node of the Colossus cluster was running ESXi 4.1 and had a Nutanix Controller VM running on it. Thus, Project Colossus was running fifty controllers in a single cluster, making it one of the most scalable enterprise-class storage systems in the industry. 7

Performance Test Setup On each node, six virtual machines were created, representing a highperformance server virtualization environment. Each virtual machine was connected to one Nutanix virtual disk. Thus, Colossus had a total of 300 virtual machines with 300 virtual disks. Inside each virtual machine, a test application was running the test workloads (random/sequential). A centralized test manager orchestrated these test applications running on 300 virtual machines. To test a variety of workloads, the system was benchmarked for random write performance and sequential read performance. Benchmark Results Sequential Read Performance Four Nodes (One Block) The chart below shows performance results delivered by a four-node cluster. 2500 Bandwidth (MB/s) 2000 1500 1000 500 Bandwidth (MB/s) 0 Time (sec) 20 40 60 80 100 120 140 160 180 200 220 240 260 280 300 This chart shows that the four-node cluster delivered an average of 2 Gigabytes per second of sequential read throughput. The test virtual machines completed their tasks in ~6 minutes. 8

Sequential Read Performance Fifty Nodes (12.5 Blocks) To test how sequential read throughput increased with the cluster size, the test was run on Colossus, the fifty-node cluster. 29000 28500 Aggregate Bandwidth vs Time (secs) 28000 27500 Bandwidth (MB/s) 27000 26500 26000 Time (sec) 20 40 60 80 100 120 140 160 180 200 220 240 260 280 This chart shows that the cluster delivered an average of about 28 Gigabytes per second of sequential read throughput. Thus, as the cluster size increased 12.5 times, the performance increased almost 14 times, demonstrating linear scalability (adjusting for statistical variations). 9

Random Write Performance Four Nodes (One Block) Random write workloads are very common in virtualized environments. When a large number of virtual machines are writing data, shared storage often ends up witnessing a random data stream even if the VMs are writing data sequentially due to an IO blender effect. Aggregate IOPS vs Time (sec) 35000 30000 25000 20000 15000 IOPS 10000 5000 0 The chart above shows random write performance delivered by a 4-node cluster. As can be seen here, the average IOPs delivered are about 25,000, with 30,000 IOPs being reached frequently during the test. 10

Random Write Performance Fifty Nodes (12.5 Blocks) The chart below shows random write performance delivered by Colossus: 410,000 400,000 390,000 380,000 370,000 360,000 350,000 340,000 330,000 320,000 Time (sec) Aggregate IOPS vs Time (secs) 20 40 60 80 100 120 140 160 180 200 220 240 260 280 IOPS Here Colossus delivered 375,000 IOPs on an average, touching almost 400,000 IOPs a number of times during the tests. This is almost a 12X-14X increase over a single-block performance, demonstrating linear scalability as more nodes are added to the cluster. 11

CONCLUSION Conclusion Nutanix Complete Cluster solves the biggest capacity planning dilemma faced by data center architects today overpaying for an oversized system from the beginning or hoping that they don t hit a performance wall before the system is ready for refresh. It allows them to start with a small footprint of a single Block and grow linearly as their needs grow. It can deliver high performance at small scale as well as large scale, making it fit for a broad variety of use cases. By using Nutanix Complete Cluster, organizations can enjoy the benefits of high performance and low and predictable cost at the same time. 12