SOLUTION BRIEF AUGUST 2015. All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

Similar documents
Solution Brief July All-Flash Server-Side Storage for Oracle Real Application Clusters (RAC) on Oracle Linux

HGST Virident Solutions 2.0

Building a Flash Fabric

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

SMB Direct for SQL Server and Private Cloud

Flash Performance for Oracle RAC with PCIe Shared Storage A Revolutionary Oracle RAC Architecture

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Building a Scalable Storage with InfiniBand

How To Scale Myroster With Flash Memory From Hgst On A Flash Flash Flash Memory On A Slave Server

Mellanox Academy Online Training (E-learning)

Virtual Compute Appliance Frequently Asked Questions

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

Configuring VMware vsphere 5.1 with Oracle ZFS Storage Appliance and Oracle Fabric Interconnect

Michael Kagan.

Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct

SX1012: High Performance Small Scale Top-of-Rack Switch

Running Highly Available, High Performance Databases in a SAN-Free Environment

Connecting the Clouds

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

ECLIPSE Performance Benchmarks and Profiling. January 2009

Cisco for SAP HANA Scale-Out Solution on Cisco UCS with NetApp Storage

Software-defined Storage at the Speed of Flash

Performance Analysis: Scale-Out File Server Cluster with Windows Server 2012 R2 Date: December 2014 Author: Mike Leone, ESG Lab Analyst

Stovepipes to Clouds. Rick Reid Principal Engineer SGI Federal by SGI Federal. Published by The Aerospace Corporation with permission.

LS DYNA Performance Benchmarks and Profiling. January 2009

Best Practices for Optimizing Storage for Oracle Automatic Storage Management with Oracle FS1 Series Storage ORACLE WHITE PAPER JANUARY 2015

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Optimizing SQL Server AlwaysOn Implementations with OCZ s ZD-XL SQL Accelerator

Accelerating Microsoft Exchange Servers with I/O Caching

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Storage, Cloud, Web 2.0, Big Data Driving Growth

SMB Advanced Networking for Fault Tolerance and Performance. Jose Barreto Principal Program Managers Microsoft Corporation

Boost Database Performance with the Cisco UCS Storage Accelerator

Connecting Flash in Cloud Storage

High Performance MySQL Cluster Cloud Reference Architecture using 16 Gbps Fibre Channel and Solid State Storage Technology

Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems

Oracle Exadata Database Machine for SAP Systems - Innovation Provided by SAP and Oracle for Joint Customers

White paper FUJITSU Storage ETERNUS DX series

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Frequently Asked Questions December ZeusRAM Solid-State Drive. Frequently Asked Questions

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

EMC VFCACHE ACCELERATES ORACLE

High Availability Databases based on Oracle 10g RAC on Linux

3G Converged-NICs A Platform for Server I/O to Converged Networks

Mellanox Accelerated Storage Solutions

Long-Haul System Family. Highest Levels of RDMA Scalability, Simplified Distance Networks Manageability, Maximum System Productivity

Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

State of the Art Cloud Infrastructure

SUN HARDWARE FROM ORACLE: PRICING FOR EDUCATION

Intel RAID SSD Cache Controller RCS25ZB040

How To Write An Article On An Hp Appsystem For Spera Hana

Violin Memory Arrays With IBM System Storage SAN Volume Control

Oracle Acceleration with the SanDisk ION Accelerator Solution

Frequently Asked Questions March s620 SATA SSD Enterprise-Class Solid-State Device. Frequently Asked Questions

Microsoft Windows Server Hyper-V in a Flash

EMC XtremSF: Delivering Next Generation Performance for Oracle Database

High Performance Oracle RAC Clusters A study of SSD SAN storage A Datapipe White Paper

Simplifying Big Data Deployments in Cloud Environments with Mellanox Interconnects and QualiSystems Orchestration Solutions

Optimizing Large Arrays with StoneFly Storage Concentrators

End-to-end Data integrity Protection in Storage Systems

Configuration Maximums

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

Scalable NAS for Oracle: Gateway to the (NFS) future

InfiniBand vs Fibre Channel Throughput. Figure 1. InfiniBand vs 2Gb/s Fibre Channel Single Port Storage Throughput to Disk Media

MS Exchange Server Acceleration

Microsoft Windows Server in a Flash

StarWind Virtual SAN for Microsoft SOFS

Replacing SAN with High Performance Windows Share over a Converged Network

FLOW-3D Performance Benchmark and Profiling. September 2012

HP reference configuration for entry-level SAS Grid Manager solutions

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

SX1024: The Ideal Multi-Purpose Top-of-Rack Switch

Installing Hadoop over Ceph, Using High Performance Networking

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture

Dell s SAP HANA Appliance

InfiniBand Switch System Family. Highest Levels of Scalability, Simplified Network Manageability, Maximum System Productivity

Implementing FlashSystem 840 with SAN Volume Controller IBM Redbooks Solution Guide

All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at

Lab Validation Report

Configuration Maximums VMware Infrastructure 3

OceanStor 18500/18800/18800F Data Sheet

Open-E Data Storage Software and Intel Modular Server a certified virtualization solution

Hyper-V over SMB: Remote File Storage Support in Windows Server 2012 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

Zadara Storage Cloud A

Enabling High performance Big Data platform with RDMA

Hadoop on the Gordon Data Intensive Cluster

ORACLE BIG DATA APPLIANCE X3-2

Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops

Using EonStor FC-host Storage Systems in VMware Infrastructure 3 and vsphere 4

Transcription:

AUGUT 2015 All-Flash erver-ide torage for Oracle Real Application Clusters (RAC) on Oracle Linux

Introduction Traditional AN storage systems cannot keep up with growing application performance needs. The FlashMAX server-side flash platform together with Oracle Automatic torage Manager (AM) offers a new cost-effective approach to implementing highperformance scalable flash storage for Oracle RAC databases. Then Fibre Channel witches Now olution Architecture Highlights FlashMAX II PCIe cards inside servers used as primary storage with shared access across the cluster Oracle AM performs volume management and data mirroring 0.5TB to 72TB of flash in a standard x86 server Consistently high performance with linear scalability Distributed architecture with no single point of failure At a fraction of the AN cost d PCIe Flash Pool Managed by Oracle AM Example of clustered database performance reported by Oracle CALIBRATE_IO. Test Configuration. 3-node RAC cluster Two FlashMAX II 2.2 TB cards per node CPU: Dual Xeon E5-2690 8KB database page size Calibrate_IO results: MAX_IOP = 1076921 LATENCY = 0 MAX_MBP = 12249 Example of LOB2 performance improvements in an actual customer environment. 2-node Oracle RAC cluster ervers AN ervers with PCIe Flash One FlashMAX II 1.1 TB card per node Validated by Oracle Robustness of the novel architecture has been validated with Oracle Linux Test (OLT) suite on Oracle Linux 6.5 with Unbreakable Enterprise Kernel (UEK) Release 3. The validated status ensures that customers get wellcoordinated support from both and Oracle. Compared to existing FC AN Validated configuration details: http://linux.oracle.com/ pls/apex/f?p=102:2:::no::p2_vc_id:649 OLT suite description: https://oss.oracle.com/projects/ olt/dist/documentation/olt_testcoverage.pdf 2

QL Executions per econd 400 350 300 250 200 150 100 50 0 29 4 FC AN 116 FlashMAX II with olutions oftware Consistently High Performance 67 185 8 Having all data on flash ensures that the performance stays consistently high regardless of the hot data size. Placing flash inside the servers with direct PCIe connectivity ensures the lowest latencies and the highest bandwidth. The non-blocking InfiniBand interconnect with Remote Direct Memory Access (RDMA) technology keeps the microsecond latencies while achieving up to 1 million IOP and up to 10GB/s of bandwidth per node. 47 264 16 Concurrent essions 45 350 32 Linear caling of torage Performance and Capacity tart with the Exact ize You Need Today Add erver to a Cluster When Needed Linear calability Add New Clusters as Needed When adding more nodes to a cluster, each new node adds both storage capacity and performance. eparate clusters use their own separate storage inside each cluster without interference between the clusters. The architecture is fully distributed. There is no centralized controller limiting the performance. With the nonblocking InfiniBand interconnect the storage performance can scale linearly up to 64 nodes per cluster with unlimited number of clusters. d Access to PCIe Flash with d access to storage from all nodes of a cluster is a mandatory requirement with Oracle RAC. software provides shared access to all FlashMAX PCIe flash drives across the cluster. With each FlashMAX II card in a cluster is accessible from all nodes in the cluster as if it was a local block storage device. comes with additional tools that provide integration with Oracle AM and simplify configuration and maintenance tasks. Network Configuration Options provides the following interconnect options: InfiniBand 56 Gb/s or 40 Gb/s with RDMA for the best performance 10 GbE or 40 GbE for broad interoperability 2-node clusters can use direct back-to-back links without using additional switches. If there are 3 or more nodes in a cluster then using switches is required. High availability of Virident interconnect is achieved by having two links to each node. In configurations with switches these two links should be connected to two separate switches. When using InfiniBand, provides embedded dual-path capability. When using 10/40 GbE, standard network bonding mechanisms are available. In either case, there is no need for a separate multipath driver. 3

eparate or torage Nodes Typically, installing PCIe flash cards inside each database server is the most effective configuration. However, in some scenarios it makes sense to have additional storage nodes, to have additional database nodes, or to have separate database or storage nodes. All of these options are supported. ici/na torage (<1GB) for OCR and Voting Files App ervers and Clients Ethernet and Oracle Private Networks Back-to-Back InfiniBand or RAC Node 1 10/40 GbE RAC Node 2 Optional ici/na torage (<1GB) for OCR and Voting Files App ervers and Clients Ethernet RAC Node 1 RAC Node 2 RAC Node 3 and Oracle Private Networks InfiniBand or 10/40 GbE witches torage Node 1 torage Node 2 App ervers and Clients Optional ici/na torage (<1GB) for OCR and Voting Files Ethernet RAC Node 1 RAC Node 2 RAC Node 3 and Oracle Private Networks InfiniBand or 10/40 GbE witches 4

High Availability The distributed architecture has no single point of failure. Oracle AM performs volume management including mirroring and striping data across cards and nodes. To ensure mirroring of the data, Normal Redundancy (2-way mirroring) or High Redundancy (3-way mirroring) must be selected when creating an AM disk group. Each block of data has mirrored copies on two or three nodes depending on redundancy level selected for the AM disk group. Device access mapping with 2 nodes and 1 card per node. Device access mapping with 3 nodes and 1 card per node. Node 1 LUN 3 Oracle AM Normal Redundancy (2-way Mirroring) or High Redundancy (3-way Mirroring) Node 2 LUN 3 Node 3 LUN 3 Node 1 Node 2 Card 1 Card 2 Card 3 Oracle AM Normal Redundancy (2-way Mirroring) Card 1 Card 2 To ensure data safety in case of a whole node failure, all cards residing in the same node must be configured as belonging to the same Failure Group when creating an AM disk group. Each node must have a Failure Group corresponding to it. AM stores mirrored copies of the data in separate Failure Groups, which means on separate nodes. Additionally, FlashMAX II provides the following reliability and data protection capabilities on an individual PCIe flash card level: Embedded flash-aware RAID Global wear leveling Extended ECC End-to-end data protection with CRC Failure group configuration with 2 nodes and 2 cards per node. Node 1 Oracle AM Disk Group with Normal Redundancy (2-way Mirror) A B Failure Group 1 A Node 2 B Failure Group 2 Capacitors protecting write data Card 1A Card 2A Card 1B Card 2B 5

Blocks with user data A1 A2 A3 A4 A5 A6 A7 AP B1 B2 Parity blocks allow reconstructing data when a block with user data fails B3 B4 B5 B6 B7 pare blocks replace bad blocks BP Virident oftware FlashMAX II PCIe Ds Interconnect Operating ystems InfiniBand Adapters Oracle Versions Version 1.2x 550 GB tandard 1100 GB tandard 1100 GB Performance 2200 GB tandard InfiniBand Oracle Linux 6 with UEK R2 Oracle Linux 6 with UEK R3 Oracle Linux 6 with RH kernel Redhat Enterprise Linux 6 Mellanox ConnectX-2 Mellanox ConnectX-3 11.2 12.1 Version 2.0.X 550 GB tandard 1100 GB tandard 1100 GB Performance 2200 GB tandard 4800 GB Capacity InfiniBand 10 GbE / 40 GbE Oracle Linux 6 with UEK R3 Oracle Linux 6 with RH kernel Redhat Enterprise Linux 5/6 Mellanox ConnectX-2 Mellanox ConnectX-3 Mellanox Connect-IB 11.2 12.1 2015, Inc., 3403 Yerba Buena Road, an Jose, CA 95135 UA. Produced in the United tates. All rights reserved. Please visit the upport section of our website, www.hgst.com/support, for additional information on product specifications. Photographs may show design models. FlashMAX is a registered trademark of, Inc. and its affiliates in the United tates and/or other countries. Other trademarks are the property of their respective owners. trademarks are intended and authorized for use only in countries and jurisdictions in which has obtained the rights to use, market and advertise the brand. Contact for additional information. shall not be liable to third parties for unauthorized use of this document or unauthorized use of its trademarks. References in this publication to s products, programs, or services do not imply that intends to make these available in all countries in which it operates. Product specifications provided are sample specifications and do not constitute a warranty. Information is true as of the date of publication and is subject to change. Actual specifications for unique part numbers may vary. B02-OracleRAC-EN-02