A Novel Data Placement Model for Highly-Available Storage Systems
|
|
|
- Erik Freeman
- 10 years ago
- Views:
Transcription
1 A Novel Data Placement Model for Highly-Available Storage Systems Rama, Microsoft Research joint work with John MacCormick, Nick Murphy, Kunal Talwar, Udi Wieder, Junfeng Yang, and Lidong Zhou
2 Introduction Kinesis: Framework for placing data and replicas in a datacenter storage system Three Design Principles: Structured organization Freedom of choice Scattered placement
3 Kinesis Design => Structured Organization Segmentation: Divide storage servers into k segments Each segment is an independent hash-based system Failure Isolation: Servers with shared components in the same segment r replicas per item, each on a different segment Reduces impact of correlated failures
4 Kinesis Design => Freedom of Choice Balanced Resource Usage: Multiple-choice paradigm Write: r out of k choices Read: 1 out of r choices
5 Kinesis Design => Scattered Data Placement Independent hash functions Each server stores different set of items Parallel Recovery Spread recovery load among multiple servers uniformly Recover faster
6 Motivation => Data Placement Models Hash-based Directory-based Data is stored on a server determined by a hash function Server identified by a local computation during reads Low overhead Any server can store any data A directory provides the server to fetch data form Expensive to maintain a globally-consistent directory in a large-scale system Limited control in placing data items and replicas Can place data items carefully to maintain load balance, avoid correlated failure, etc.
7 Kinesis => Advantages Enables hash-based storage systems to have advantages of directory-based systems Near-optimal load balance Tolerance to shared-component failures Freedom from problems induced by correlated replica placement No bottlenecks induced by directory maintenance Avoids slow data/replica placement algorithms
8 Evaluation Theory Simulations Experiments
9 Theory => The Balls and Bins Problem Load balancing is often modeled by the task of throwing balls (items) into bins (servers) Throw m balls into n bins: Pick a bin uniformly at random Insert the ball into the bin Single-Choice Paradigm: Max Load: (with high prob.)
10 Theory => Multiple-Choice Paradigm Throw m balls into n bins: Pick d bins uniformly at random (d 2) Insert the ball into the less-loaded bin Excess load is independent of m, number of balls! [BCSV00] Max Load: (with high prob.)
11 Theory vs. Practice Storage load vs. network load: [M91] Network load is not persistent unlike storage load Non-uniform sampling: [V99],[W07] Servers chosen based on consistent (or linear) hashing Replication: [MMRWYZ08] Choose r out of k servers instead of 1 out of d bins Heterogeneity: Variable-sized servers [W07] Variable-sized items [TW07]
12 Theory => Planned Expansions Adding new, empty servers into the system Creates sudden, large imbalances in load Empty servers vs. filled servers Eventual storage balance Do subsequent inserts/writes fill up new servers? Eventual network load balance Are new items distributed between old and new servers? New items are often more popular than old items
13 Theory => Planned Expansions Expanding a linear-hashing system: 0 2 b-1 2 b N = 2 N = 4 N = 6
14 Theory => Planned Expansions Before Expansion: Assume there are k n servers on k segments Storage is evenly balanced Expansion: Add α n new servers to each segment After expansion: R = (1 - α) n servers with twice the load as L = 2 α n servers (new servers + split servers)
15 Theory => Planned Expansions Relative Load Distribution: R L Ratio of expected number of replicas inserted in L servers to expected number of replicas inserted on all servers R L > 1 => eventual storage balance! R L < 1 => no eventual storage balance Theorem: R L > 1 if k 2 r [MMRWYZ08]
16 Evaluation Theory Simulations Experiments
17 Simulations => Overview Real-World Traces: Trace Num Files Total Size Type MSNBC 30, TB Read only Plan Million 173 GB Read/write Compare with Chain: one segment single-choice paradigm chained replica placement E.g. PAST, CFS, Boxwood, Petal 0 2 b 2 b-1
18 Storage: Max - Avg (GB) Storage: Max - Avg (GB) Simulations => Load Balance MSNBC Trace Plan-9 Trace Chain(3) Kinesis(7,3) Chain(3) Kinesis(7,3) Number of Objects Number of Operations (10 6 )
19 Percentage Overprovisioning Simulations => System Provisioning 60% 52.0% 45% 45.0% 35.0% 30% 29.0% 26.0% 23.0% 20.0% 15% 0% K (5) K (6) K (7) K (8) K (9) K (10) Chain
20 90 th pc. Queuing Delay (sec) Simulations => User Experienced Delays 10 8 Chain (3) Kinesis (7,3) Number of Servers
21 Simulations => Read Load vs. Write Load Read load: short term resource consumption Network bandwidth, computation, disk bandwidth Directly impacts user experience Write load: short and long term resource consumption Storage space Solution (Kinesis S+L): Choose short term over long term Storage balance is restored when transient resources are not bottlenecked
22 Queuing Delay: Max - Avg (ms) Simulations => Read Load vs. Write Load Chain Kinesis S Kinesis S+L Kinesis L Update to Read Ratio (%)
23 Request Rate (Gb/s) Storage Imbalance (%) Simulations => Read Load vs. Write Load Request Rate Kinesis S+L Kinesis S 15% 10% 0.5 5% Time (hours) 0%
24 Evaluation Theory Simulations Experiments
25 Kinesis => Prototype Implementation Storage system for variable-sized objects Read/Write interface Storage servers Linear hashing system Read, Append, Exists, Size (storage), Load (queued requests) Failure recovery for primary replicas Primary for an item is the highest numbered server with a replica
26 Kinesis => Prototype Implementation Front end Two-step read protocol 1) Query k candidate servers 2) Fetch from least loaded server with a replica Versioned updates with copy-on-write semantics RPC-based communication with servers Non implemented! Failure detector Failure-consistent updates
27 Kinesis => Experiments 15 node LAN test-bed 14 storage servers and 1 front end MSNBC trace of 6 hours duration 5000 files, 170 GB total size, and 200,000 reads Failure induced at 3 hours
28 Experiments => Kinesis vs. Chain Kinesis(7,3) Chain(3) Kinesis: 1.73 sec Average Latency Chain: 3 sec
29 Experiments => Failure Recovery Kinesis(7,3) Chain(3) Kinesis: 17 min (12 servers) Total Recovery Time Chain: 44 min (5 servers)
30 Related Work Prior work: Hashing: Archipelago, OceanStore, PAST, CFS Two-choice paradigm: common load balancing technique Parallel recovery: Chain Replication [RS04] Random distribution: Ceph [WBML06] Our contributions: Putting them together in the context of storage systems Extending theoretical results to the new design Demonstrating the power of these simple design principles
31 Summary and Conclusions Kinesis: A data/replica placement framework for LAN storage systems Structured organization, freedom-of-choice, and scattered distribution Simple and easy to implement, yet quite powerful! Reduces infrastructure cost, tolerates correlated failures, and quickly restores lost replicas Soon to appear in ACM Transactions on Storage, 2008 Download:
32
33 Average Latency (sec) Experiments => Kinesis vs. Chain 15 Chain (3) Kinesis (7,3) Time (hours)
34 Request Rate Server Load: Max - Avg Experiments => Kinesis vs. Chain Total Load Chain (3) Kinesis (7,3) Time (hours) 0
Storage Systems Autumn 2009. Chapter 6: Distributed Hash Tables and their Applications André Brinkmann
Storage Systems Autumn 2009 Chapter 6: Distributed Hash Tables and their Applications André Brinkmann Scaling RAID architectures Using traditional RAID architecture does not scale Adding news disk implies
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card
File System Management
Lecture 7: Storage Management File System Management Contents Non volatile memory Tape, HDD, SSD Files & File System Interface Directories & their Organization File System Implementation Disk Space Allocation
Load Balancing in Distributed Web Server Systems With Partial Document Replication
Load Balancing in Distributed Web Server Systems With Partial Document Replication Ling Zhuo, Cho-Li Wang and Francis C. M. Lau Department of Computer Science and Information Systems The University of
RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University
RAMCloud and the Low- Latency Datacenter John Ousterhout Stanford University Most important driver for innovation in computer systems: Rise of the datacenter Phase 1: large scale Phase 2: low latency Introduction
The Advantages and Disadvantages of Network Computing Nodes
Big Data & Scripting storage networks and distributed file systems 1, 2, in the remainder we use networks of computing nodes to enable computations on even larger datasets for a computation, each node
Berkeley Ninja Architecture
Berkeley Ninja Architecture ACID vs BASE 1.Strong Consistency 2. Availability not considered 3. Conservative 1. Weak consistency 2. Availability is a primary design element 3. Aggressive --> Traditional
Solbox Cloud Storage Acceleration
DATA SHEET Solbox Cloud Storage Acceleration Today s ongoing and rapidly-accelerating growth in data comes at the same time that organizations of all sizes are focused on cost deduction. Cloud storage
Scala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
MinCopysets: Derandomizing Replication In Cloud Storage
MinCopysets: Derandomizing Replication In Cloud Storage Asaf Cidon, Ryan Stutsman, Stephen Rumble, Sachin Katti, John Ousterhout and Mendel Rosenblum Stanford University [email protected], {stutsman,rumble,skatti,ouster,mendel}@cs.stanford.edu
1 Storage Devices Summary
Chapter 1 Storage Devices Summary Dependability is vital Suitable measures Latency how long to the first bit arrives Bandwidth/throughput how fast does stuff come through after the latency period Obvious
Massive Data Storage
Massive Data Storage Storage on the "Cloud" and the Google File System paper by: Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung presentation by: Joshua Michalczak COP 4810 - Topics in Computer Science
Distributed File Systems
Distributed File Systems Mauro Fruet University of Trento - Italy 2011/12/19 Mauro Fruet (UniTN) Distributed File Systems 2011/12/19 1 / 39 Outline 1 Distributed File Systems 2 The Google File System (GFS)
RAID. RAID 0 No redundancy ( AID?) Just stripe data over multiple disks But it does improve performance. Chapter 6 Storage and Other I/O Topics 29
RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f. one large disk) Parallelism improves performance Plus extra disk(s) for redundant data storage Provides fault tolerant
GraySort on Apache Spark by Databricks
GraySort on Apache Spark by Databricks Reynold Xin, Parviz Deyhim, Ali Ghodsi, Xiangrui Meng, Matei Zaharia Databricks Inc. Apache Spark Sorting in Spark Overview Sorting Within a Partition Range Partitioner
International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 ISSN 2229-5518
International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 349 Load Balancing Heterogeneous Request in DHT-based P2P Systems Mrs. Yogita A. Dalvi Dr. R. Shankar Mr. Atesh
POWER ALL GLOBAL FILE SYSTEM (PGFS)
POWER ALL GLOBAL FILE SYSTEM (PGFS) Defining next generation of global storage grid Power All Networks Ltd. Technical Whitepaper April 2008, version 1.01 Table of Content 1. Introduction.. 3 2. Paradigm
SCALABILITY AND AVAILABILITY
SCALABILITY AND AVAILABILITY Real Systems must be Scalable fast enough to handle the expected load and grow easily when the load grows Available available enough of the time Scalable Scale-up increase
Cluster Computing. ! Fault tolerance. ! Stateless. ! Throughput. ! Stateful. ! Response time. Architectures. Stateless vs. Stateful.
Architectures Cluster Computing Job Parallelism Request Parallelism 2 2010 VMware Inc. All rights reserved Replication Stateless vs. Stateful! Fault tolerance High availability despite failures If one
The Google File System
The Google File System By Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung (Presented at SOSP 2003) Introduction Google search engine. Applications process lots of data. Need good file system. Solution:
Lecture 5: GFS & HDFS! Claudia Hauff (Web Information Systems)! [email protected]
Big Data Processing, 2014/15 Lecture 5: GFS & HDFS!! Claudia Hauff (Web Information Systems)! [email protected] 1 Course content Introduction Data streams 1 & 2 The MapReduce paradigm Looking behind
Enabling Multi-pipeline Data Transfer in HDFS for Big Data Applications
Enabling Multi-pipeline Data Transfer in HDFS for Big Data Applications Liqiang (Eric) Wang, Hong Zhang University of Wyoming Hai Huang IBM T.J. Watson Research Center Background Hadoop: Apache Hadoop
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel Computation Parallel I/O (I) I/O basics Spring 2008 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network
Outline. Clouds of Clouds lessons learned from n years of research Miguel Correia
Dependability and Security with Clouds of Clouds lessons learned from n years of research Miguel Correia WORKSHOP ON DEPENDABILITY AND INTEROPERABILITY IN HETEROGENEOUS CLOUDS (DIHC13) August 27 th 2013,
CS 153 Design of Operating Systems Spring 2015
CS 153 Design of Operating Systems Spring 2015 Lecture 22: File system optimizations Physical Disk Structure Disk components Platters Surfaces Tracks Arm Track Sector Surface Sectors Cylinders Arm Heads
Load Balancing in Fault Tolerant Video Server
Load Balancing in Fault Tolerant Video Server # D. N. Sujatha*, Girish K*, Rashmi B*, Venugopal K. R*, L. M. Patnaik** *Department of Computer Science and Engineering University Visvesvaraya College of
Computing Load Aware and Long-View Load Balancing for Cluster Storage Systems
215 IEEE International Conference on Big Data (Big Data) Computing Load Aware and Long-View Load Balancing for Cluster Storage Systems Guoxin Liu and Haiying Shen and Haoyu Wang Department of Electrical
3. PGCluster. There are two formal PGCluster Web sites. http://pgfoundry.org/projects/pgcluster/ http://pgcluster.projects.postgresql.
3. PGCluster PGCluster is a multi-master replication system designed for PostgreSQL open source database. PostgreSQL has no standard or default replication system. There are various third-party software
G22.3250-001. Porcupine. Robert Grimm New York University
G22.3250-001 Porcupine Robert Grimm New York University Altogether Now: The Three Questions! What is the problem?! What is new or different?! What are the contributions and limitations? Porcupine from
Corporate PC Backup - Best Practices
A Druva Whitepaper Corporate PC Backup - Best Practices This whitepaper explains best practices for successfully implementing laptop backup for corporate workforce. White Paper WP /100 /009 Oct 10 Table
Self-Compressive Approach for Distributed System Monitoring
Self-Compressive Approach for Distributed System Monitoring Akshada T Bhondave Dr D.Y Patil COE Computer Department, Pune University, India Santoshkumar Biradar Assistant Prof. Dr D.Y Patil COE, Computer
RAID Storage System of Standalone NVR
ACTi Knowledge Base Category: Installation & Configuration Note Sub-category: Hardware; Recording & Storage Model: XNR-4200, GNR-2000 Firmware: Software: Author: Ando.Meritee Published: 2011/05/04 Reviewed:
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What
Load Re-Balancing for Distributed File. System with Replication Strategies in Cloud
Contemporary Engineering Sciences, Vol. 8, 2015, no. 10, 447-451 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ces.2015.5263 Load Re-Balancing for Distributed File System with Replication Strategies
The Pros and Cons of Erasure Coding & Replication vs. RAID in Next-Gen Storage Platforms. Abhijith Shenoy Engineer, Hedvig Inc.
The Pros and Cons of Erasure Coding & Replication vs. RAID in Next-Gen Storage Platforms Abhijith Shenoy Engineer, Hedvig Inc. @hedviginc The need for new architectures Business innovation Time-to-market
A Deduplication-based Data Archiving System
2012 International Conference on Image, Vision and Computing (ICIVC 2012) IPCSIT vol. 50 (2012) (2012) IACSIT Press, Singapore DOI: 10.7763/IPCSIT.2012.V50.20 A Deduplication-based Data Archiving System
Mike Canney Principal Network Analyst getpackets.com
Mike Canney Principal Network Analyst getpackets.com 1 My contact info contact Mike Canney, Principal Network Analyst, getpackets.com [email protected] 319.389.1137 2 Capture Strategies capture Capture
EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server
White Paper EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server Abstract This white paper addresses the challenges currently facing business executives to store and process the growing
Chapter 6. 6.1 Introduction. Storage and Other I/O Topics. p. 570( 頁 585) Fig. 6.1. I/O devices can be characterized by. I/O bus connections
Chapter 6 Storage and Other I/O Topics 6.1 Introduction I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections
Avamar. Technology Overview
1 Technology Overview 2 Growth of Remote Offices Remote offices are growing rapidly Over 4 million remote offices in the US No standard remote office size, data, staff, network, etc. The average branch
Energy Efficient MapReduce
Energy Efficient MapReduce Motivation: Energy consumption is an important aspect of datacenters efficiency, the total power consumption in the united states has doubled from 2000 to 2005, representing
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
Google File System. Web and scalability
Google File System Web and scalability The web: - How big is the Web right now? No one knows. - Number of pages that are crawled: o 100,000 pages in 1994 o 8 million pages in 2005 - Crawlable pages might
A SURVEY OF POPULAR CLUSTERING TECHNOLOGIES
A SURVEY OF POPULAR CLUSTERING TECHNOLOGIES By: Edward Whalen Performance Tuning Corporation INTRODUCTION There are a number of clustering products available on the market today, and clustering has become
Backup of NAS devices with Avamar
Backup of NAS devices with Avamar Extremely fast / no load Video describing NAS backup using Avamar based on this ppt: https://youtu.be/swg1ejldgmw The most fresh version of this document, you will find
How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda
How swift is your Swift? Ning Zhang, OpenStack Engineer at Zmanda Chander Kant, CEO at Zmanda 1 Outline Build a cost-efficient Swift cluster with expected performance Background & Problem Solution Experiments
Big Data With Hadoop
With Saurabh Singh [email protected] The Ohio State University February 11, 2016 Overview 1 2 3 Requirements Ecosystem Resilient Distributed Datasets (RDDs) Example Code vs Mapreduce 4 5 Source: [Tutorials
Cassandra A Decentralized, Structured Storage System
Cassandra A Decentralized, Structured Storage System Avinash Lakshman and Prashant Malik Facebook Published: April 2010, Volume 44, Issue 2 Communications of the ACM http://dl.acm.org/citation.cfm?id=1773922
Multimedia Caching Strategies for Heterogeneous Application and Server Environments
Multimedia Tools and Applications 4, 279 312 (1997) c 1997 Kluwer Academic Publishers. Manufactured in The Netherlands. Multimedia Caching Strategies for Heterogeneous Application and Server Environments
In Memory Accelerator for MongoDB
In Memory Accelerator for MongoDB Yakov Zhdanov, Director R&D GridGain Systems GridGain: In Memory Computing Leader 5 years in production 100s of customers & users Starts every 10 secs worldwide Over 15,000,000
Load Balancing. Load Balancing 1 / 24
Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait
Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms
Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
Mike Canney. Application Performance Analysis
Mike Canney Application Performance Analysis 1 Welcome to Sharkfest 12 contact Mike Canney, Principal Network Analyst, Tektivity, Inc. [email protected] 319-365-3336 www.getpackets.com 2 Agenda agenda
Snapshots in Hadoop Distributed File System
Snapshots in Hadoop Distributed File System Sameer Agarwal UC Berkeley Dhruba Borthakur Facebook Inc. Ion Stoica UC Berkeley Abstract The ability to take snapshots is an essential functionality of any
Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database
WHITE PAPER Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive
Designing a Cloud Storage System
Designing a Cloud Storage System End to End Cloud Storage When designing a cloud storage system, there is value in decoupling the system s archival capacity (its ability to persistently store large volumes
Review. Lecture 21: Reliable, High Performance Storage. Overview. Basic Disk & File System properties CSC 468 / CSC 2204 11/23/2006
S 468 / S 2204 Review Lecture 2: Reliable, High Performance Storage S 469HF Fall 2006 ngela emke rown We ve looked at fault tolerance via server replication ontinue operating with up to f failures Recovery
Cloud Computing with Microsoft Azure
Cloud Computing with Microsoft Azure Michael Stiefel www.reliablesoftware.com [email protected] http://www.reliablesoftware.com/dasblog/default.aspx Azure's Three Flavors Azure Operating
Simulating a File-Sharing P2P Network
Simulating a File-Sharing P2P Network Mario T. Schlosser, Tyson E. Condie, and Sepandar D. Kamvar Department of Computer Science Stanford University, Stanford, CA 94305, USA Abstract. Assessing the performance
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing
Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Hsin-Wen Wei 1,2, Che-Wei Hsu 2, Tin-Yu Wu 3, Wei-Tsong Lee 1 1 Department of Electrical Engineering, Tamkang University
Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage
Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage Technical white paper Table of contents Executive summary... 2 Introduction... 2 Test methodology... 3
Non-Stop for Apache HBase: Active-active region server clusters TECHNICAL BRIEF
Non-Stop for Apache HBase: -active region server clusters TECHNICAL BRIEF Technical Brief: -active region server clusters -active region server clusters HBase is a non-relational database that provides
Scale and Availability Considerations for Cluster File Systems. David Noy, Symantec Corporation
Scale and Availability Considerations for Cluster File Systems David Noy, Symantec Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted.
Web Load Stress Testing
Web Load Stress Testing Overview A Web load stress test is a diagnostic tool that helps predict how a website will respond to various traffic levels. This test can answer critical questions such as: How
Giving life to today s media distribution services
Giving life to today s media distribution services FIA - Future Internet Assembly Athens, 17 March 2014 Presenter: Nikolaos Efthymiopoulos Network architecture & Management Group Copyright University of
Distributed file system in cloud based on load rebalancing algorithm
Distributed file system in cloud based on load rebalancing algorithm B.Mamatha(M.Tech) Computer Science & Engineering [email protected] K Sandeep(M.Tech) Assistant Professor PRRM Engineering College
Key Components of WAN Optimization Controller Functionality
Key Components of WAN Optimization Controller Functionality Introduction and Goals One of the key challenges facing IT organizations relative to application and service delivery is ensuring that the applications
Oracle Database In-Memory The Next Big Thing
Oracle Database In-Memory The Next Big Thing Maria Colgan Master Product Manager #DBIM12c Why is Oracle do this Oracle Database In-Memory Goals Real Time Analytics Accelerate Mixed Workload OLTP No Changes
Reliability of Data Storage Systems
Zurich Research Laboratory Ilias Iliadis April 2, 25 Keynote NexComm 25 www.zurich.ibm.com 25 IBM Corporation Long-term Storage of Increasing Amount of Information An increasing amount of information is
Cloud DBMS: An Overview. Shan-Hung Wu, NetDB CS, NTHU Spring, 2015
Cloud DBMS: An Overview Shan-Hung Wu, NetDB CS, NTHU Spring, 2015 Outline Definition and requirements S through partitioning A through replication Problems of traditional DDBMS Usage analysis: operational
Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012
Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012 1 Market Trends Big Data Growing technology deployments are creating an exponential increase in the volume
WHITE PAPER. SQL Server License Reduction with PernixData FVP Software
WHITE PAPER SQL Server License Reduction with PernixData FVP Software 1 Beyond Database Acceleration Poor storage performance continues to be the largest pain point with enterprise Database Administrators
Principles and characteristics of distributed systems and environments
Principles and characteristics of distributed systems and environments Definition of a distributed system Distributed system is a collection of independent computers that appears to its users as a single
White Paper. PiLab Technology Performance Test, Deep Queries. pilab.pl
White Paper PiLab Technology Performance Test, Deep Queries pilab.pl Table of Content Forward... 3 Summary of Results... 3 Technical Introduction... 3 Hardware... 4 Disk read/write speed... 4 Network speed
EMC XtremSF: Delivering Next Generation Performance for Oracle Database
White Paper EMC XtremSF: Delivering Next Generation Performance for Oracle Database Abstract This white paper addresses the challenges currently facing business executives to store and process the growing
The Hadoop Distributed File System
The Hadoop Distributed File System Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia, Chansler}@Yahoo-Inc.com Presenter: Alex Hu HDFS
A block based storage model for remote online backups in a trust no one environment
A block based storage model for remote online backups in a trust no one environment http://www.duplicati.com/ Kenneth Skovhede (author, [email protected]) René Stach (editor, [email protected]) Abstract
Towards an Optimized Big Data Processing System
Towards an Optimized Big Data Processing System The Doctoral Symposium of the IEEE/ACM CCGrid 2013 Delft, The Netherlands Bogdan Ghiţ, Alexandru Iosup, and Dick Epema Parallel and Distributed Systems Group
Disk Storage & Dependability
Disk Storage & Dependability Computer Organization Architectures for Embedded Computing Wednesday 19 November 14 Many slides adapted from: Computer Organization and Design, Patterson & Hennessy 4th Edition,
Input / Ouput devices. I/O Chapter 8. Goals & Constraints. Measures of Performance. Anatomy of a Disk Drive. Introduction - 8.1
Introduction - 8.1 I/O Chapter 8 Disk Storage and Dependability 8.2 Buses and other connectors 8.4 I/O performance measures 8.6 Input / Ouput devices keyboard, mouse, printer, game controllers, hard drive,
Ceph Optimization on All Flash Storage
Ceph Optimization on All Flash Storage Somnath Roy Lead Developer, SanDisk Corporation Santa Clara, CA 1 Forward-Looking Statements During our meeting today we may make forward-looking statements. Any
Cloud Computing Is In Your Future
Cloud Computing Is In Your Future Michael Stiefel www.reliablesoftware.com [email protected] http://www.reliablesoftware.com/dasblog/default.aspx Cloud Computing is Utility Computing Illusion
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at distributing load b. QUESTION: What is the context? i. How
Data Deduplication and Tivoli Storage Manager
Data Deduplication and Tivoli Storage Manager Dave Cannon Tivoli Storage Manager rchitect Oxford University TSM Symposium September 2007 Disclaimer This presentation describes potential future enhancements
SOFT 437. Software Performance Analysis. Ch 5:Web Applications and Other Distributed Systems
SOFT 437 Software Performance Analysis Ch 5:Web Applications and Other Distributed Systems Outline Overview of Web applications, distributed object technologies, and the important considerations for SPE
Practical Data Integrity Protection in Network-Coded Cloud Storage
Practical Data Integrity Protection in Network-Coded Cloud Storage Henry C. H. Chen Department of Computer Science and Engineering The Chinese University of Hong Kong Outline Introduction FMSR in NCCloud
Best Practices for Virtualised SharePoint
Best Practices for Virtualised SharePoint Brendan Law [email protected] @FlamerNZ Flamer.co.nz/spag/ Nathan Mercer [email protected] @NathanM blogs.technet.com/nmercer/ Agenda Why Virtualise? Hardware
Graph Database Proof of Concept Report
Objectivity, Inc. Graph Database Proof of Concept Report Managing The Internet of Things Table of Contents Executive Summary 3 Background 3 Proof of Concept 4 Dataset 4 Process 4 Query Catalog 4 Environment
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT
MINIMIZING STORAGE COST IN CLOUD COMPUTING ENVIRONMENT 1 SARIKA K B, 2 S SUBASREE 1 Department of Computer Science, Nehru College of Engineering and Research Centre, Thrissur, Kerala 2 Professor and Head,
POSIX and Object Distributed Storage Systems
1 POSIX and Object Distributed Storage Systems Performance Comparison Studies With Real-Life Scenarios in an Experimental Data Taking Context Leveraging OpenStack Swift & Ceph by Michael Poat, Dr. Jerome
CS 6290 I/O and Storage. Milos Prvulovic
CS 6290 I/O and Storage Milos Prvulovic Storage Systems I/O performance (bandwidth, latency) Bandwidth improving, but not as fast as CPU Latency improving very slowly Consequently, by Amdahl s Law: fraction
