Violin: A Framework for Extensible Block-level Storage
|
|
|
- Milton Craig
- 9 years ago
- Views:
Transcription
1 Violin: A Framework for Extensible Block-level Storage Michail Flouris Dept. of Computer Science, University of Toronto, Canada [email protected] Angelos Bilas ICS-FORTH & University of Crete, Greece [email protected]
2 Large-scale Storage Large-scale storage systems in a data center Driven by Need for massive amounts of scalable storage Consolidation potential for lower costs Challenges in scalable storage Scalability and performance with commodity components Reduction of management cost Wide-area storage sharing 13/4/2005 [email protected] 2
3 Reducing Costs With Emerging scalable, low-cost hardware Commodity clusters / grids (x86 with Linux / BSD) Commodity interconnects and standard protocols (PCI-X/Express/AS, SATA, SCSI, iscsi, GigE) Storage virtualization software that Offers diverse storage views for different applications Automates storage management functions Supports monitoring Exhibits scalability and low overhead We want to improve virtualization at the block-level 13/4/2005 [email protected] 3
4 Block-level Virtualization Virtualization has two meanings Notion 1: Indirection Mapping between physical and logical resources Facilitates resource management Notion 2: Sharing Hides system behind abstractions for sharing Our goal Provide block-level virtualization mechanisms to improve indirection and resource management Why block-level? Transparency, Performance, Flexibility 13/4/2005 4
5 Issue with existing virtualization Current software has configuration flexibility Use of a small set of predefined modules (RAID levels, volume management) Module combination in arbitrary manner But missing Functional Flexibility Ability to extend the system with new functionality Extensions implemented by modules loaded on-demand Not compromising configuration flexibility (New extension modules are combined with old ones) Add management, performance, reliability-related features (e.g. encryption, versioning, migration, compression) 13/4/2005 [email protected] 5
6 Why Functional Flexibility? Reduce implementation complexity Combine simple modules to build complex features Customizing system to user s/application s needs Adaptivity to existing or new applications Incremental system evolution Add new functionality before or after deployment Optimize or upgrade components Prototyping and evaluation of new ideas Not compromising configuration flexibility Creating extensions should be easy Developed by vendors, users, or storage administrators 13/4/2005 [email protected] 6
7 Our Goals Designing a system with automated storage management without extensions, does not work We intend to add desirable management, performance, reliability-related features, in an incremental evolution fashion Violin provides the mechanisms to achieve this Management automation Will evolve over time Will include an initial configuration phase and continuous monitoring and dynamic reconfiguration afterwards System will be able adapt to new applications 13/4/2005 [email protected] 7
8 Extensible Filesystems Related Work Ficus, FiST. Not at block-level, complementary approach. Extensible Network Protocols Click Router, X-kernel, Horus Similar layer stacking for network protocol extensions But: basic differences between network and storage Block-level virtualization software OSS Volume Managers: Linux MD, LVM, EVMS, GEOM Numerous Commercial Solutions Provide only configuration flexibility 13/4/2005 8
9 Outline Motivation Violin Design Implementation Evaluation Conclusions 13/4/2005 9
10 Violin An extensible block-level hierarchy over physical devices User Space Kernel Space Kernel components Violin context User extensions Filesystem Apps Filesystem Block I/O Buffer Cache Violin Modules Block I/O Apps Disk Device Drivers Raw I/O Apps Raw I/O Persistent Metadata LXT High level I/O API Violin kernel module Disks 13/4/
11 Violin Design Goals Easy to develop new extensions Easy to combine them in I/O hierarchy Low overhead Violin achieves this by providing 1. Convenient semantics and mappings 2. Simple control of the I/O request path 3. Persistent metadata support 13/4/
12 Virtualization Hierarchies Device graph maps I/O sources to I/O sinks I/O requests pass through devices (layers) in graph Nodes are virtual devices Edges are mappings Hierarchies: connected device sub-graphs, or independent I/O stacks Graph is DAG Directed acyclic graph Internal Devices (A M) E I F Sink Source Source Source H Hierarchy A A B C D Sink J G Sink Sink Hierarchy B Sink L K M Input Capacities: Output Capacities: 13/4/2005 [email protected] 12
13 Virtual Devices A virtual device is driven by an extension module Device/Layer is runtime instance of module Sees input, output address spaces and one or more output devices Maps arbitrarily blocks between devices Transforms data between input and output, vice versa Some modules need logical translation table (LXT) A type of logical device metadata Input Address Space LXT Output Address Space /4/2005 [email protected] 13
14 Control of I/O Requests Violin API allows layers simple control on requests passing though them Layer can initiate, forward, complete or terminate I/O requests using simple tags or calls Initiating I/O: Useful for multiplying I/O flows Layers initiate asynchronous I/O requests using callback handlers, executed on completion Forward I/O: Send I/O request to a lower layer Complete I/O: Useful for caching layers Terminate I/O: Error-handling 13/4/2005 [email protected] 14
15 Left Example: Tagged Request Control Request is forwarded through the hierarchy from source to sink (Layer order: F, E, D, B) and then back up Right Example: Request is forwarded until layer C, where it is tagged complete and returns upwards without reaching the sink Fwd Path Fwd to E Fwd to D Fwd to B C A Source Source 13/4/2005 [email protected] 15 < < < Sink B Return Path F E Sink D > Tag Request Complete Fwd To Sink Return Path Fwd Path Fwd to E Fwd to D Fwd to B < A < < < Sink C B F E Sink D
16 Persistent Metadata Storage layers need persistent state For superblocks, partition tables, block maps, etc. Violin offers persistent objects for layer metadata Persistent Objects are memory-mapped storage blocks, accessed as generic memory objects Automatically synchronized to stable storage periodically Automatically loaded / unloaded during startup / shutdown Layers need only allocate objects once Violin internal metadata are also persistent objects Device graph and hierarchy info stored at superblocks 13/4/2005 [email protected] 16
17 Metadata Consistency Three levels of metadata consistency (weaker to stronger) 1. Lazy-updates Synchronized overwriting older metadata periodically Similar to non-journaling filesystems 2. Shadow-updates Using two copies of all metadata (normal & shadow) Synchronization overwrites first normal metadata, then shadow Guarantees module metadata consistency 3. Atomic versioned-metadata consistency Module metadata and application data are versioned On failure the system rolls back to a consistent snapshot Violin currently supports levels 1 and 2 13/4/2005 [email protected] 17
18 Block Size and Memory Overhead Small block sizes can increase memory footprint of module metadata When metadata proportional to total number of blocks Many OSes have small block sizes Linux 2.4.x: 4KB block device size Modules need own independent block size to manage metadata in larger chunks Violin supports larger internal block size Size set by modules Independent from OS block size Reduces memory overhead effectively 13/4/
19 Outline Motivation Violin Design Implementation Evaluation Conclusions 13/4/
20 Violin Core Implementation Linux 2.4 loadable block device driver Registers extension modules Provides API & services to extension modules Violin Extension Modules Loadable Linux kernel modules that bind to Core Not device drivers themselves (much simpler) 13/4/
21 Example Modules RAID RAID Levels 0, 1 and 5 with recovery. Aggregation (plain or striped volumes) Volume Remapping (add, remove, move Volumes) Partitioning Managing Partitions (create, delete, resize partitions) Versioning (Online Snapshots) Online Migration Data Fingerprinting (MD5) Encryption Currently DES, 3DES and Blowfish algorithms 13/4/2005 [email protected] 21
22 Evaluation Ease of module development Configuration Flexibility Performance 13/4/
23 Evaluation: Ease Of Development Loose comparison of number of code lines Code lines reduced 2-6 times for similar functionality Little to reasonable effort for module development 13/4/
24 Evaluation: Configuration Flexibility Easily creating a hierarchy with complex functionality from implemented modules Violin allows arbitrary combinations of extension modules VERSIONING PARTITION BLOWFISH Disk RAID 0 PARTITION BLOWFISH Disk 13/4/2005 [email protected] 24
25 Evaluation: Performance Platform: Dual Athlon-MP PCs, 512MB RAM, GigE NIC, Western Digital 80GB IDE Disks RedHat Linux 9.0 (Kernel smp) Benchmarks IOmeter ( ) for raw block I/O Postmark for filesystem measurements Experiment Cases 1. Pass-through: System Disk vs. Violin Pass-through Layer 2. Vol. Manager: LVM vs. Violin Aggregate+Partition Layers 3. RAID-0: Linux MD vs. Violin RAID 4. RAID-1: Linux MD vs. Violin RAID 13/4/
26 Violin Pass-through vs. System Disk Violin Raw Disk - 100% Read Violin Raw Disk - 100% Write Violin Raw Disk - 70% Read, 30% Write Raw System Disk - 100% Read Raw System Disk - 100% Write Raw System Disk - 70% Read, 30% Write Sequential I/O 7 6 Random I/O MBytes / sec MBytes / sec B 1KB 2KB 4KB 8KB 16KB 32KB 64KB 512B 1KB 2KB 4KB 8KB 16KB 32KB 64KB Block size Block size IOmeter throughput for Violin Raw Disk vs. System Raw Disk for One Disk 13/4/2005 [email protected] 26
27 Violin vs. LVM (2 Striped Disks) Violin Aggr+Part - 100% Read Violin Aggr+Part - 100% Write Violin Aggr+Part - 70% Read, 30% Write LVM - 100% Read LVM - 100% Write LVM - 70% Read, 30% Write Random I/O MBytes / sec MBytes / sec Sequential I/O B 1KB 2KB 4KB 8KB 16KB 32KB 64KB 512B 1KB 2KB 4KB 8KB 16KB 32KB 64KB Block size Block size IOmeter throughput for Violin Aggregation+Partition vs. LVM for 2 Striped Disks (32K stripe) 13/4/2005 [email protected] 27
28 Violin vs. MD (RAID-1 Mirroring, 2 Disks) Violin RAID-1-100% Read Violin RAID-1-100% Write Violin RAID-1-70% Read, 30% Write Linux MD RAID-1-100% Read Linux MD RAID-1-100% Write Linux MD RAID-1-70% Read, 30% Write Sequential I/O 7 6 Random I/O MBytes / sec MBytes / sec B 1KB 2KB 4KB 8KB 16KB 32KB 64KB 512B 1KB 2KB 4KB 8KB 16KB 32KB 64KB Block size Block size IOmeter throughput for RAID-1: Violin vs. Linux MD for 2 Disks 13/4/2005 [email protected] 28
29 Postmark Results over Ext2 Filesystem Transactions per second Total run time (seconds) Transactions / sec Total Run Time (sec) 0 0 LVM Vl. Aggr+Part System Disk Violin Disk MD RAID-1 Violin RAID-1 MD RAID-0 Violin RAID-0 LVM Vl. Aggr+Part System Disk Violin Disk MD RAID-1 Violin RAID-1 MD RAID-0 Violin RAID-0 13/4/2005 [email protected] 29
30 Multiple Layer Performance Partition Partition + RAID-0 Partition + RAID-0 + Version. Partition + RAID-0 + Version. + Blowfish Seq. Read MBytes / sec MBytes / sec Seq. Write B 1KB 2KB 4KB 8KB 16KB 32KB 64KB 512B 1KB 2KB 4KB 8KB 16KB 32KB 64KB Block size Workload: 100% Sequential Read Block size Workload: 100% Sequential Write 13/4/2005 [email protected] 30
31 Storage Node A Violin Violin Limitations and Future Work iscsi Application Server Application FS Violin iscsi iscsi iscsi Violin Storage Node N Distributed Hierarchy Violin does not fully support distributed hierarchies No consistency for dynamically updated metadata Future work: Supporting distributed hierarchies in a cluster Automated hierarchy configuration and management, according to specified requirements and policies 13/4/2005 [email protected] 31
32 Conclusions Goal is to improve virtualization in storage cluster Propose Violin, an extensible I/O layer stack Violin s contributions are mechanisms for Convenient virtualization semantics and mappings Simple control of I/O requests from layers Persistent metadata These mechanisms Make it easy to write extensions Make it easy to combine them Exhibit low overhead (< 10% in our implementation) We believe that Violin s mechanisms are a step towards automated storage management 13/4/2005 [email protected] 32
33 Thank You. Questions? Violin: A Framework for Extensible Block-level Storage, Michail Flouris and Angelos Bilas [email protected], [email protected] 13/4/2005 [email protected] 33
Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1
Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System
Flexible Storage Allocation
Flexible Storage Allocation A. L. Narasimha Reddy Department of Electrical and Computer Engineering Texas A & M University Students: Sukwoo Kang (now at IBM Almaden) John Garrison Outline Big Picture Part
Filesystems Performance in GNU/Linux Multi-Disk Data Storage
JOURNAL OF APPLIED COMPUTER SCIENCE Vol. 22 No. 2 (2014), pp. 65-80 Filesystems Performance in GNU/Linux Multi-Disk Data Storage Mateusz Smoliński 1 1 Lodz University of Technology Faculty of Technical
File System & Device Drive. Overview of Mass Storage Structure. Moving head Disk Mechanism. HDD Pictures 11/13/2014. CS341: Operating System
CS341: Operating System Lect 36: 1 st Nov 2014 Dr. A. Sahu Dept of Comp. Sc. & Engg. Indian Institute of Technology Guwahati File System & Device Drive Mass Storage Disk Structure Disk Arm Scheduling RAID
<Insert Picture Here> Btrfs Filesystem
Btrfs Filesystem Chris Mason Btrfs Goals General purpose filesystem that scales to very large storage Feature focused, providing features other Linux filesystems cannot Administration
Maximizing Hadoop Performance and Storage Capacity with AltraHD TM
Maximizing Hadoop Performance and Storage Capacity with AltraHD TM Executive Summary The explosion of internet data, driven in large part by the growth of more and more powerful mobile devices, has created
AIX NFS Client Performance Improvements for Databases on NAS
AIX NFS Client Performance Improvements for Databases on NAS October 20, 2005 Sanjay Gulabani Sr. Performance Engineer Network Appliance, Inc. [email protected] Diane Flemming Advisory Software Engineer
Distributed RAID Architectures for Cluster I/O Computing. Kai Hwang
Distributed RAID Architectures for Cluster I/O Computing Kai Hwang Internet and Cluster Computing Lab. University of Southern California 1 Presentation Outline : Scalable Cluster I/O The RAID-x Architecture
Datacenter Operating Systems
Datacenter Operating Systems CSE451 Simon Peter With thanks to Timothy Roscoe (ETH Zurich) Autumn 2015 This Lecture What s a datacenter Why datacenters Types of datacenters Hyperscale datacenters Major
Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms
Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes
Backup Software Technology
Backup Software Technology Well over 100 Backup Applications In The Wild Three Backup Software Categories: 1. Legacy 2. Online 3. near-continuous Data Protection Legacy Backup Software Data Lost in Disaster
Client/Server Computing Distributed Processing, Client/Server, and Clusters
Client/Server Computing Distributed Processing, Client/Server, and Clusters Chapter 13 Client machines are generally single-user PCs or workstations that provide a highly userfriendly interface to the
VMware Virtual Machine File System: Technical Overview and Best Practices
VMware Virtual Machine File System: Technical Overview and Best Practices A VMware Technical White Paper Version 1.0. VMware Virtual Machine File System: Technical Overview and Best Practices Paper Number:
Solid State Storage in Massive Data Environments Erik Eyberg
Solid State Storage in Massive Data Environments Erik Eyberg Senior Analyst Texas Memory Systems, Inc. Agenda Taxonomy Performance Considerations Reliability Considerations Q&A Solid State Storage Taxonomy
Cray DVS: Data Virtualization Service
Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
The Panasas Parallel Storage Cluster. Acknowledgement: Some of the material presented is under copyright by Panasas Inc.
The Panasas Parallel Storage Cluster What Is It? What Is The Panasas ActiveScale Storage Cluster A complete hardware and software storage solution Implements An Asynchronous, Parallel, Object-based, POSIX
WHITE PAPER Optimizing Virtual Platform Disk Performance
WHITE PAPER Optimizing Virtual Platform Disk Performance Think Faster. Visit us at Condusiv.com Optimizing Virtual Platform Disk Performance 1 The intensified demand for IT network efficiency and lower
BlobSeer: Towards efficient data storage management on large-scale, distributed systems
: Towards efficient data storage management on large-scale, distributed systems Bogdan Nicolae University of Rennes 1, France KerData Team, INRIA Rennes Bretagne-Atlantique PhD Advisors: Gabriel Antoniu
IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org
IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator
Oracle Cloud Storage and File system
2012 Tieto Corporation Oracle Cloud Storage and File system Andrejs Karpovs Oracle Apps DBA Tieto, [email protected] Few notes about me I Am a DBA Work in Tieto Have 4 years exprerience working
Virtual Machine Monitors. Dr. Marc E. Fiuczynski Research Scholar Princeton University
Virtual Machine Monitors Dr. Marc E. Fiuczynski Research Scholar Princeton University Introduction Have been around since 1960 s on mainframes used for multitasking Good example VM/370 Have resurfaced
Storage Systems Autumn 2009. Chapter 6: Distributed Hash Tables and their Applications André Brinkmann
Storage Systems Autumn 2009 Chapter 6: Distributed Hash Tables and their Applications André Brinkmann Scaling RAID architectures Using traditional RAID architecture does not scale Adding news disk implies
Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.
Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance
COS 318: Operating Systems. Virtual Machine Monitors
COS 318: Operating Systems Virtual Machine Monitors Andy Bavier Computer Science Department Princeton University http://www.cs.princeton.edu/courses/archive/fall10/cos318/ Introduction Have been around
Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.
Oracle BI EE Implementation on Netezza Prepared by SureShot Strategies, Inc. The goal of this paper is to give an insight to Netezza architecture and implementation experience to strategize Oracle BI EE
Network Attached Storage. Jinfeng Yang Oct/19/2015
Network Attached Storage Jinfeng Yang Oct/19/2015 Outline Part A 1. What is the Network Attached Storage (NAS)? 2. What are the applications of NAS? 3. The benefits of NAS. 4. NAS s performance (Reliability
Ultimate Guide to Oracle Storage
Ultimate Guide to Oracle Storage Presented by George Trujillo [email protected] George Trujillo Twenty two years IT experience with 19 years Oracle experience. Advanced database solutions such
Technology Insight Series
Evaluating Storage Technologies for Virtual Server Environments Russ Fellows June, 2010 Technology Insight Series Evaluator Group Copyright 2010 Evaluator Group, Inc. All rights reserved Executive Summary
SMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
Simplifying Server Workload Migrations
Simplifying Server Workload Migrations This document describes the migration challenges that organizations face and how the Acronis AnyData Engine supports physical-to-physical (P2P), physical-to-virtual
Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle
Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server
June 2009. Blade.org 2009 ALL RIGHTS RESERVED
Contributions for this vendor neutral technology paper have been provided by Blade.org members including NetApp, BLADE Network Technologies, and Double-Take Software. June 2009 Blade.org 2009 ALL RIGHTS
High Availability Databases based on Oracle 10g RAC on Linux
High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN, June 2006 Luca Canali, CERN IT Outline Goals Architecture of an HA DB Service Deployment at the CERN Physics Database
High Availability with Windows Server 2012 Release Candidate
High Availability with Windows Server 2012 Release Candidate Windows Server 2012 Release Candidate (RC) delivers innovative new capabilities that enable you to build dynamic storage and availability solutions
DualFS: A New Journaling File System for Linux
2007 Linux Storage & Filesystem Workshop February 12-13, 13, 2007, San Jose, CA DualFS: A New Journaling File System for Linux Juan Piernas SDM Project Pacific Northwest National
How To Virtualize A Storage Area Network (San) With Virtualization
A New Method of SAN Storage Virtualization Table of Contents 1 - ABSTRACT 2 - THE NEED FOR STORAGE VIRTUALIZATION 3 - EXISTING STORAGE VIRTUALIZATION METHODS 4 - A NEW METHOD OF VIRTUALIZATION: Storage
SolidFire and NetApp All-Flash FAS Architectural Comparison
SolidFire and NetApp All-Flash FAS Architectural Comparison JUNE 2015 This document provides an overview of NetApp s All-Flash FAS architecture as it compares to SolidFire. Not intended to be exhaustive,
StorPool Distributed Storage Software Technical Overview
StorPool Distributed Storage Software Technical Overview StorPool 2015 Page 1 of 8 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard
Oracle Database 10g: Performance Tuning 12-1
Oracle Database 10g: Performance Tuning 12-1 Oracle Database 10g: Performance Tuning 12-2 I/O Architecture The Oracle database uses a logical storage container called a tablespace to store all permanent
Distributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
TELE 301 Lecture 7: Linux/Unix file
Overview Last Lecture Scripting This Lecture Linux/Unix file system Next Lecture System installation Sources Installation and Getting Started Guide Linux System Administrators Guide Chapter 6 in Principles
COS 318: Operating Systems. Virtual Machine Monitors
COS 318: Operating Systems Virtual Machine Monitors Kai Li and Andy Bavier Computer Science Department Princeton University http://www.cs.princeton.edu/courses/archive/fall13/cos318/ Introduction u Have
On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform
On- Prem MongoDB- as- a- Service Powered by the CumuLogic DBaaS Platform Page 1 of 16 Table of Contents Table of Contents... 2 Introduction... 3 NoSQL Databases... 3 CumuLogic NoSQL Database Service...
<Insert Picture Here> Oracle Cloud Storage. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska
Oracle Cloud Storage Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska Oracle Cloud Storage Automatic Storage Management (ASM) Oracle Cloud File System ASM Dynamic
Maginatics Cloud Storage Platform for Elastic NAS Workloads
Maginatics Cloud Storage Platform for Elastic NAS Workloads Optimized for Cloud Maginatics Cloud Storage Platform () is the first solution optimized for the cloud. It provides lower cost, easier administration,
Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation
Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V Jose Barreto Principal Program Manager Microsoft Corporation Agenda Hyper-V over SMB - Overview How to set it up Configuration Options
MODULE 3 VIRTUALIZED DATA CENTER COMPUTE
MODULE 3 VIRTUALIZED DATA CENTER COMPUTE Module 3: Virtualized Data Center Compute Upon completion of this module, you should be able to: Describe compute virtualization Discuss the compute virtualization
PARALLELS CLOUD STORAGE
PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...
Table of Contents Introduction and System Requirements 9 Installing VMware Server 35
Table of Contents Introduction and System Requirements 9 VMware Server: Product Overview 10 Features in VMware Server 11 Support for 64-bit Guest Operating Systems 11 Two-Way Virtual SMP (Experimental
Storage Management for the Oracle Database on Red Hat Enterprise Linux 6: Using ASM With or Without ASMLib
Storage Management for the Oracle Database on Red Hat Enterprise Linux 6: Using ASM With or Without ASMLib Sayan Saha, Sue Denham & Lars Herrmann 05/02/2011 On 22 March 2011, Oracle posted the following
Storage Architectures for Big Data in the Cloud
Storage Architectures for Big Data in the Cloud Sam Fineberg HP Storage CT Office/ May 2013 Overview Introduction What is big data? Big Data I/O Hadoop/HDFS SAN Distributed FS Cloud Summary Research Areas
Microsoft SMB File Sharing Best Practices Guide
Technical White Paper Microsoft SMB File Sharing Best Practices Guide Tintri VMstore, Microsoft SMB 3.0 Protocol, and VMware 6.x Author: Neil Glick Version 1.0 06/15/2016 @tintri www.tintri.com Contents
CS 377: Operating Systems. Outline. A review of what you ve learned, and how it applies to a real operating system. Lecture 25 - Linux Case Study
CS 377: Operating Systems Lecture 25 - Linux Case Study Guest Lecturer: Tim Wood Outline Linux History Design Principles System Overview Process Scheduling Memory Management File Systems A review of what
Avid ISIS 7000. www.avid.com
Avid ISIS 7000 www.avid.com Table of Contents Overview... 3 Avid ISIS Technology Overview... 6 ISIS Storage Blade... 6 ISIS Switch Blade... 7 ISIS System Director... 7 ISIS Client Software... 8 ISIS Redundant
Introduction to Gluster. Versions 3.0.x
Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster
Big data management with IBM General Parallel File System
Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers
Chapter 11 I/O Management and Disk Scheduling
Operating Systems: Internals and Design Principles, 6/E William Stallings Chapter 11 I/O Management and Disk Scheduling Dave Bremer Otago Polytechnic, NZ 2008, Prentice Hall I/O Devices Roadmap Organization
Enterprise Backup and Restore technology and solutions
Enterprise Backup and Restore technology and solutions LESSON VII Veselin Petrunov Backup and Restore team / Deep Technical Support HP Bulgaria Global Delivery Hub Global Operations Center November, 2013
WHITE PAPER. Permabit Albireo Data Optimization Software. Benefits of Albireo for Virtual Servers. January 2012. Permabit Technology Corporation
WHITE PAPER Permabit Albireo Data Optimization Software Benefits of Albireo for Virtual Servers January 2012 Permabit Technology Corporation Ten Canal Park Cambridge, MA 02141 USA Phone: 617.252.9600 FAX:
Support for Storage Volumes Greater than 2TB Using Standard Operating System Functionality
Support for Storage Volumes Greater than 2TB Using Standard Operating System Functionality Introduction A History of Hard Drive Capacity Starting in 1984, when IBM first introduced a 5MB hard drive in
RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University
RAMCloud and the Low- Latency Datacenter John Ousterhout Stanford University Most important driver for innovation in computer systems: Rise of the datacenter Phase 1: large scale Phase 2: low latency Introduction
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage
Moving Virtual Storage to the Cloud Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage Table of Contents Overview... 1 Understanding the Storage Problem... 1 What Makes
Software-defined Storage Architecture for Analytics Computing
Software-defined Storage Architecture for Analytics Computing Arati Joshi Performance Engineering Colin Eldridge File System Engineering Carlos Carrero Product Management June 2015 Reference Architecture
IOmark- VDI. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC- 150427- b Test Report Date: 27, April 2015. www.iomark.
IOmark- VDI HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC- 150427- b Test Copyright 2010-2014 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VM, VDI- IOmark, and IOmark
Moving Virtual Storage to the Cloud
Moving Virtual Storage to the Cloud White Paper Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage www.parallels.com Table of Contents Overview... 3 Understanding the Storage
ZooKeeper. Table of contents
by Table of contents 1 ZooKeeper: A Distributed Coordination Service for Distributed Applications... 2 1.1 Design Goals...2 1.2 Data model and the hierarchical namespace...3 1.3 Nodes and ephemeral nodes...
Maximizing SQL Server Virtualization Performance
Maximizing SQL Server Virtualization Performance Michael Otey Senior Technical Director Windows IT Pro SQL Server Pro 1 What this presentation covers Host configuration guidelines CPU, RAM, networking
MagFS: The Ideal File System for the Cloud
: The Ideal File System for the Cloud is the first true file system for the cloud. It provides lower cost, easier administration, and better scalability and performance than any alternative in-cloud file
Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation
Overview of I/O Performance and RAID in an RDBMS Environment By: Edward Whalen Performance Tuning Corporation Abstract This paper covers the fundamentals of I/O topics and an overview of RAID levels commonly
High Availability Solutions with MySQL
High Availability Solutions with MySQL best OpenSystems Day Fall 2008 Ralf Gebhardt Senior Systems Engineer MySQL Global Software Practice [email protected] 1 HA Requirements and Considerations HA
IOmark-VM. DotHill AssuredSAN Pro 5000. Test Report: VM- 130816-a Test Report Date: 16, August 2013. www.iomark.org
IOmark-VM DotHill AssuredSAN Pro 5000 Test Report: VM- 130816-a Test Report Date: 16, August 2013 Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark-VM, IOmark-VDI, VDI-IOmark, and IOmark
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007
Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms Cray User Group Meeting June 2007 Cray s Storage Strategy Background Broad range of HPC requirements
SALSA Flash-Optimized Software-Defined Storage
Flash-Optimized Software-Defined Storage Nikolas Ioannou, Ioannis Koltsidas, Roman Pletka, Sasa Tomic,Thomas Weigold IBM Research Zurich 1 New Market Category of Big Data Flash Multiple workloads don t
VDI Optimization Real World Learnings. Russ Fellows, Evaluator Group
Russ Fellows, Evaluator Group SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies and individual members may use this material
Violin Memory Arrays With IBM System Storage SAN Volume Control
Technical White Paper Report Best Practices Guide: Violin Memory Arrays With IBM System Storage SAN Volume Control Implementation Best Practices and Performance Considerations Version 1.0 Abstract This
Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated
Taking Linux File and Storage Systems into the Future Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated 1 Overview Going Bigger Going Faster Support for New Hardware Current Areas
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011
Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011 Executive Summary Large enterprise Hyper-V deployments with a large number
Application Brief: Using Titan for MS SQL
Application Brief: Using Titan for MS Abstract Businesses rely heavily on databases for day-today transactions and for business decision systems. In today s information age, databases form the critical
Understanding Enterprise NAS
Anjan Dave, Principal Storage Engineer LSI Corporation Author: Anjan Dave, Principal Storage Engineer, LSI Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA
Scala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
Stream Processing on GPUs Using Distributed Multimedia Middleware
Stream Processing on GPUs Using Distributed Multimedia Middleware Michael Repplinger 1,2, and Philipp Slusallek 1,2 1 Computer Graphics Lab, Saarland University, Saarbrücken, Germany 2 German Research
Big Fast Data Hadoop acceleration with Flash. June 2013
Big Fast Data Hadoop acceleration with Flash June 2013 Agenda The Big Data Problem What is Hadoop Hadoop and Flash The Nytro Solution Test Results The Big Data Problem Big Data Output Facebook Traditional
Managing Storage Space in a Flash and Disk Hybrid Storage System
Managing Storage Space in a Flash and Disk Hybrid Storage System Xiaojian Wu, and A. L. Narasimha Reddy Dept. of Electrical and Computer Engineering Texas A&M University IEEE International Symposium on
Sanbolic s SAN Storage Enhancing Software Portfolio
Software to Simplify and Share SAN Storage Sanbolic s SAN Storage Enhancing Software Portfolio Overview of Product Suites www.sanbolic.com Version 2.0 Page 2 of 10 Contents About Sanbolic... 3 Sanbolic
September 2009 Cloud Storage for Cloud Computing
September 2009 Cloud Storage for Cloud Computing This paper is a joint production of the Storage Networking Industry Association and the Open Grid Forum. Copyright 2009 Open Grid Forum, Copyright 2009
SAP HANA Operation Expert Summit BUILD - High Availability & Disaster Recovery
SAP HANA Operation Expert Summit BUILD - High Availability & Disaster Recovery Dr. Ralf Czekalla/SAP HANA Product Management May 09, 2014 Customer Disclaimer This presentation outlines our general product
Solution Guide Parallels Virtualization for Linux
Solution Guide Parallels Virtualization for Linux Overview Created in 1991, Linux was designed to be UNIX-compatible software that was composed entirely of open source or free software components. Linux
High Availability Solutions for the MariaDB and MySQL Database
High Availability Solutions for the MariaDB and MySQL Database 1 Introduction This paper introduces recommendations and some of the solutions used to create an availability or high availability environment
: HP HP0-771. Version : R6.1
Exam : HP HP0-771 Title : Designing & Implementing HP Enterprise Backup Solutions Version : R6.1 Prepking - King of Computer Certification Important Information, Please Read Carefully Other Prepking products
An Oracle White Paper July 2014. Oracle ACFS
An Oracle White Paper July 2014 Oracle ACFS 1 Executive Overview As storage requirements double every 18 months, Oracle customers continue to deal with complex storage management challenges in their data
Two Parts. Filesystem Interface. Filesystem design. Interface the user sees. Implementing the interface
File Management Two Parts Filesystem Interface Interface the user sees Organization of the files as seen by the user Operations defined on files Properties that can be read/modified Filesystem design Implementing
Solaris For The Modern Data Center. Taking Advantage of Solaris 11 Features
Solaris For The Modern Data Center Taking Advantage of Solaris 11 Features JANUARY 2013 Contents Introduction... 2 Patching and Maintenance... 2 IPS Packages... 2 Boot Environments... 2 Fast Reboot...
Oracle Database Scalability in VMware ESX VMware ESX 3.5
Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises
Next Generation Operating Systems
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015 The end of CPU scaling Future computing challenges Power efficiency Performance == parallelism Cisco Confidential 2 Paradox of the
Feature Comparison. Windows Server 2008 R2 Hyper-V and Windows Server 2012 Hyper-V
Comparison and Contents Introduction... 4 More Secure Multitenancy... 5 Flexible Infrastructure... 9 Scale, Performance, and Density... 13 High Availability... 18 Processor and Memory Support... 24 Network...
A Performance Analysis of the iscsi Protocol
A Performance Analysis of the iscsi Protocol Stephen Aiken [email protected] Dirk Grunwald [email protected] Jesse Willeke [email protected] Andrew R. Pleszkun [email protected]
Best Practices for Deploying Citrix XenDesktop on NexentaStor Open Storage
Best Practices for Deploying Citrix XenDesktop on NexentaStor Open Storage White Paper July, 2011 Deploying Citrix XenDesktop on NexentaStor Open Storage Table of Contents The Challenges of VDI Storage
