IO latency tracking. Daniel Lezcano (dlezcano) Linaro Power Management Team
|
|
|
- Adam Morton
- 10 years ago
- Views:
Transcription
1 IO latency tracking Daniel Lezcano (dlezcano) Linaro Power Management Team
2 CPUIDLE CPUidle is a framework divided into three parts: The generic framework The back-end driver: low level calls or hidden by firmware (ACPI, PSCI, MWAIT) The governor Select idle state Idle task Enter idle state CPUidle Generic framework Governor PM Back end driver
3 CPUIDLE : the generic framework Provides an API: To register a driver and a governor To ask for the governor suggestion To enter idle with the state abstraction No algorithm
4 CPUIDLE : the back end drivers It gives the abstraction for the complex PM code Very generic for some hardware based on firmware abstraction (eg. ACPI, PSCI,...) Very hardware dependent and complex if directly handled in the kernel (eg. ARM SoC specific drivers)
5 CPUIDLE : the governors Two governors used today: Ladder with periodic tick configuration: for server focused on performance Menu with nohz configuration: for most of the platforms trying to provide the best trade-off between performance and energy saving
6 CPUIDLE : the menu governor Tries to predict the next event on the system Does statistics on each CPU for the sleep duration Weighting the idle states regarding the pending IOs
7 CPUIDLE : the menu governor Source of wakeups: IRQ Timer, IOs, keyboard, IPI Mostly generated by the scheduler But most of the interrupts are coming from: Timers, IOs and rescheduling IPIs
8 CPUIDLE : the menu governor It takes all the source of interrupts without distinction: Is it up to the scheduler to predict the scheduler behavior (IPI reschedule)? What about if a task is moved to another CPU? Could stay to a very shallow state for seconds because the statistics are different Why not focus on event which are predictable? Timers and IOs and consider the rest as noise preventing us to be accurate
9 Why an IO could be predictable? The duration of an IO is inside a reasonable interval, with repeating pattern The tasks are blocked on IO, hence these latencies information can be tied with it The information can follow the task when this one is migrated
10 Experimentation Latency measurements for: SSD 6Gb/s HDD 6Gb/s with different block sizes From 4KB to 512KB
11 SSD behavior
12 SSD behavior
13 HDD behavior
14 HDD behavior
15 HDD behavior
16 HDD behavior
17 Observations Let's group the different latencies into buckets Experiment with different bucket size : 50us, 100us, 200us and 500us And observe the distribution : how many times a bucket is hit?
18 SSD 4KB Random RW
19 SSD 8KB Random RW
20 SSD 16KB Random RW
21 SSD 32KB Random RW
22 SSD 64KB Random RW
23 SSD 128KB Randow RW
24 SSD 256KB Random RW
25 SSD 512KB Randow RW
26 HDD 4KB Random RW
27 HDD 8KB Random RW
28 HDD 16KB Random RW
29 HDD 32KB Random RW
30 HDD 64KB Randow RW
31 HDD 128KB Randow RW
32 HDD 256KB Randow RW
33 HDD 512KB Random RW
34 Observations - Conclusions The smaller the bucket is, the more there are buckets Reduce probability to estimate the right bucket On slow media, there is an important deviation from the average latency The 200us bucket size shows a interesting trade-off between the number of buckets and the accuracy We can rely on these observations to build a model to predict the next IO
35 IO latency tracking infrastructure Each time a task is unblocked after an IO completion, measure the duration Add this latency to the IO latency tracking infrastructure Next time the task is blocked on a IO, ask the IO latency tracking infrastructure the guessed blocking time When going to idle, take the remaining time to be blocked on an IO as part of the equation to compute the sleep duration
36 IO latency tracking infrastructure Task Scheduler IO latency tracking io_schedule io_latency_begin io_latency_guess + io_latency_insert_node io_latency_end wakeup io_latency_delete_node + io_latency_add
37 IO latency tracking infrastructure Group the latencies into buckets, representing an interval (200us) Each bucket has an index Each index gives the bucket's interval index : 0 => [0, 199] index : 10 => [2000, 2799] Index : 5 => [1000, 1199]
38 IO latency tracking infrastructure io_latency_add(413us) find_bucket 413%200 = 2 bucket2 bucket3 bucket6 bucket12 update_history(bucket2,413us) bucket2.hits++ bucket2.suchits++ bucket2.avg =...
39 IO latency tracking infrastructure How do we guess the next blocking time? Buckets are sorted Position in the list shows the history (first happening more, last happening less) Buckets have the number of hits and the position in the list weight these numbers Compute a score with the position in the list and the number of hits with a decaying function
40 IO latency tracking infrastructure Score = nrhits / (pos + 1)² first bucket2.hits = 123 Score = 123 / (0+1)² = 123 bucket3.hits = 321 Score = 321 / (1+1)² = 80 bucket6.hits = 12 Score = 12 / (2+1)² = 1 last bucket12.hits = 32 Score = 12 / (3+1)² = 0 Bucket 2 is the next expected IO latency
41 IO latency tracking infrastructure What happens when there are several tasks blocked on an IO? A red-black tree per cpu The number of elements of this tree is the number of tasks blocked on a IO Each node is the guessed IO duration for the corresponding task Left most node is the smaller guessed IO duration
42 IO latency tracking infrastructure + CPUidle When entering idle we have now: The next timer event which is reliable (next_timer_event) The next IO completion (next_io_event) The next event is: next_event = min(next_timer_event, next_io_event)
43 IO latency tracking + CPUidle Choosing the idle state is straighforward: for (i = 0; i < drv->state_count; i++) { struct cpuidle_state *s = &drv->states[i]; struct cpuidle_state_usage *su = &dev->states_usage[i]; if (s->disabled su->disable) continue; if (s->target_residency > next_event) continue; if (s->exit_latency > latency_req) continue; } idle_state = i;
44 The big picture Task waiting for an IO Task waiting for a timer Idle task IO latency tracking Timer framework idle_for(min(t1, tn)) t1 tn+5 tn+4 tn+3 tn+2 tn+1 tn CPUIDLE Choose idle state Basic selection loop sleep Backend driver
45 Some results Tools used to measure: Some extra infos added to sysfs for each cpu Over predicted : the sleep duration was shorter Under predicted : the sleep duration was longer Idlestat : a tool computing the idle state statistics Iolatsimu: a tool implementing the prediction algorithm and randomly read/write a file
46 Some results Tests done under certain circumstances Process pinned on one cpu in order to prevent migration Tried on the media described at the beginning of the documentation For each block size measure the prediction regarding the idle state which has been choose
47 Some results IO latency framework + timer give an expected sleep time of 30us 1. Effective sleep time = Effective sleep time = Effective sleep time = 45 residency = 10 residency = 20 residency = 40 residency = 70 residency = 150 Shallow Deep 1. and 3. are failed predictions, 2. is a right prediction
48 SDD results - 4KB
49 SDD results 8KB
50 SSD results - 16KB
51 SSD results - 32KB
52 SDD results - 64KB
53 SDD results - 128KB
54 SDD results - 256KB
55 SDD results - 512KB
56 HDD results - 4KB
57 HDD results - 8KB
58 HDD results - 16KB
59 HDD results - 32KB
60 HDD results - 64KB
61 HDD results - 128KB
62 HDD results - 256KB
63 HDD results - 512KB
64 Conclusion A noticeable improvement in terms of prediction, more than 30% under some circumstances The resulting design makes the idle state decision within the scheduler Idling decision integrated Smarter decision could be made in the future
65 Next steps Improve the IO latency tracking framework code Investigate pointless wakeup resulting from an IPI when an IO completes Fix the IO completions measurement probes Make the latency per device Big kernel impact Improve the scheduler to take smarter decision regarding idle
66 THANKS!
Power Management in the Linux Kernel
Power Management in the Linux Kernel Tate Hornbeck, Peter Hokanson 7 April 2011 Intel Open Source Technology Center Venkatesh Pallipadi Senior Staff Software Engineer 2001 - Joined Intel Processor and
Comparing Power Saving Techniques for Multi cores ARM Platforms
Comparing Power Saving Techniques for Multi cores ARM Platforms Content Why to use CPU hotplug? Tests environment CPU hotplug constraints What we have / What we want How to
Agenda. Context. System Power Management Issues. Power Capping Overview. Power capping participants. Recommendations
Power Capping Linux Agenda Context System Power Management Issues Power Capping Overview Power capping participants Recommendations Introduction of Linux Power Capping Framework 2 Power Hungry World Worldwide,
Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat
Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat Why Computers Are Getting Slower The traditional approach better performance Why computers are
Update on big.little scheduling experiments. Morten Rasmussen Technology Researcher
Update on big.little scheduling experiments Morten Rasmussen Technology Researcher 1 Agenda Why is big.little different from SMP? Summary of previous experiments on emulated big.little. New results for
Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation
Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance Alex Ho, Product Manager Innodisk Corporation Outline Innodisk Introduction Industry Trend & Challenge
EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications
ECE6102 Dependable Distribute Systems, Fall2010 EWeb: Highly Scalable Client Transparent Fault Tolerant System for Cloud based Web Applications Deepal Jayasinghe, Hyojun Kim, Mohammad M. Hossain, Ali Payani
10.04.2008. Thomas Fahrig Senior Developer Hypervisor Team. Hypervisor Architecture Terminology Goals Basics Details
Thomas Fahrig Senior Developer Hypervisor Team Hypervisor Architecture Terminology Goals Basics Details Scheduling Interval External Interrupt Handling Reserves, Weights and Caps Context Switch Waiting
Understanding Linux on z/vm Steal Time
Understanding Linux on z/vm Steal Time June 2014 Rob van der Heij [email protected] Summary Ever since Linux distributions started to report steal time in various tools, it has been causing
Seeking Fast, Durable Data Management: A Database System and Persistent Storage Benchmark
Seeking Fast, Durable Data Management: A Database System and Persistent Storage Benchmark In-memory database systems (IMDSs) eliminate much of the performance latency associated with traditional on-disk
CFS-v: I/O Demand-driven VM Scheduler in KVM
CFS-v: Demand-driven VM Scheduler in KVM Hyotaek Shim and Sung-Min Lee (hyotaek.shim, [email protected] com) Software R&D Center, Samsung Electronics 2014. 10. 16 Problem in Server Consolidation 2/16
An Implementation Of Multiprocessor Linux
An Implementation Of Multiprocessor Linux This document describes the implementation of a simple SMP Linux kernel extension and how to use this to develop SMP Linux kernels for architectures other than
Jorix kernel: real-time scheduling
Jorix kernel: real-time scheduling Joris Huizer Kwie Min Wong May 16, 2007 1 Introduction As a specialized part of the kernel, we implemented two real-time scheduling algorithms: RM (rate monotonic) and
Real-time KVM from the ground up
Real-time KVM from the ground up KVM Forum 2015 Rik van Riel Red Hat Real-time KVM What is real time? Hardware pitfalls Realtime preempt Linux kernel patch set KVM & qemu pitfalls KVM configuration Scheduling
The Data Placement Challenge
The Data Placement Challenge Entire Dataset Applications Active Data Lowest $/IOP Highest throughput Lowest latency 10-20% Right Place Right Cost Right Time 100% 2 2 What s Driving the AST Discussion?
Completely Fair Scheduler and its tuning 1
Completely Fair Scheduler and its tuning 1 Jacek Kobus and Rafał Szklarski 1 Introduction The introduction of a new, the so called completely fair scheduler (CFS) to the Linux kernel 2.6.23 (October 2007)
Process Scheduling in Linux
The Gate of the AOSP #4 : Gerrit, Memory & Performance Process Scheduling in Linux 2013. 3. 29 Namhyung Kim Outline 1 Process scheduling 2 SMP scheduling 3 Group scheduling - www.kandroid.org 2/ 41 Process
Accelerating Server Storage Performance on Lenovo ThinkServer
Accelerating Server Storage Performance on Lenovo ThinkServer Lenovo Enterprise Product Group April 214 Copyright Lenovo 214 LENOVO PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER
Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6
Multiprocessor Scheduling and Scheduling in Linux Kernel 2.6 Winter Term 2008 / 2009 Jun.-Prof. Dr. André Brinkmann [email protected] Universität Paderborn PC² Agenda Multiprocessor and
Redpaper. Performance Metrics in TotalStorage Productivity Center Performance Reports. Introduction. Mary Lovelace
Redpaper Mary Lovelace Performance Metrics in TotalStorage Productivity Center Performance Reports Introduction This Redpaper contains the TotalStorage Productivity Center performance metrics that are
IST 301. Class Exercise: Simulating Business Processes
IST 301 Class Exercise: Simulating Business Processes Learning Objectives: To use simulation to analyze and design business processes. To implement scenario and sensitivity analysis As-Is Process The As-Is
Kernel Optimizations for KVM. Rik van Riel Senior Software Engineer, Red Hat June 25 2010
Kernel Optimizations for KVM Rik van Riel Senior Software Engineer, Red Hat June 25 2010 Kernel Optimizations for KVM What is virtualization performance? Benefits of developing both guest and host KVM
1. Computer System Structure and Components
1 Computer System Structure and Components Computer System Layers Various Computer Programs OS System Calls (eg, fork, execv, write, etc) KERNEL/Behavior or CPU Device Drivers Device Controllers Devices
A Close Look at PCI Express SSDs. Shirish Jamthe Director of System Engineering Virident Systems, Inc. August 2011
A Close Look at PCI Express SSDs Shirish Jamthe Director of System Engineering Virident Systems, Inc. August 2011 Macro Datacenter Trends Key driver: Information Processing Data Footprint (PB) CAGR: 100%
Operating Systems. Lecture 03. February 11, 2013
Operating Systems Lecture 03 February 11, 2013 Goals for Today Interrupts, traps and signals Hardware Protection System Calls Interrupts, Traps, and Signals The occurrence of an event is usually signaled
Linux Load Balancing
Linux Load Balancing Hyunmin Yoon 2 Load Balancing Linux scheduler attempts to evenly distribute load across CPUs Load of CPU (run queue): sum of task weights Load balancing is triggered by Timer interrupts
Computer-System Architecture
Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage Structure Storage Hierarchy Hardware Protection General System Architecture 2.1 Computer-System Architecture 2.2 Computer-System
A Survey of Parallel Processing in Linux
A Survey of Parallel Processing in Linux Kojiro Akasaka Computer Science Department San Jose State University San Jose, CA 95192 408 924 1000 [email protected] ABSTRACT Any kernel with parallel processing
Overlapping Data Transfer With Application Execution on Clusters
Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm [email protected] [email protected] Department of Computer Science Department of Electrical and Computer
Using Synology SSD Technology to Enhance System Performance. Based on DSM 5.2
Using Synology SSD Technology to Enhance System Performance Based on DSM 5.2 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD Cache as Solution...
Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance
Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance Hybrid Storage Performance Gains for IOPS and Bandwidth Utilizing Colfax Servers and Enmotus FuzeDrive Software NVMe Hybrid
1: B asic S imu lati on Modeling
Network Simulation Chapter 1: Basic Simulation Modeling Prof. Dr. Jürgen Jasperneite 1 Contents The Nature of Simulation Systems, Models and Simulation Discrete Event Simulation Simulation of a Single-Server
Client-aware Cloud Storage
Client-aware Cloud Storage Feng Chen Computer Science & Engineering Louisiana State University Michael Mesnier Circuits & Systems Research Intel Labs Scott Hahn Circuits & Systems Research Intel Labs Cloud
Linux Process Scheduling. sched.c. schedule() scheduler_tick() hooks. try_to_wake_up() ... CFS CPU 0 CPU 1 CPU 2 CPU 3
Linux Process Scheduling sched.c schedule() scheduler_tick() try_to_wake_up() hooks RT CPU 0 CPU 1 CFS CPU 2 CPU 3 Linux Process Scheduling 1. Task Classification 2. Scheduler Skeleton 3. Completely Fair
White paper. QNAP Turbo NAS with SSD Cache
White paper QNAP Turbo NAS with SSD Cache 1 Table of Contents Introduction... 3 Audience... 3 Terminology... 3 SSD cache technology... 4 Applications and benefits... 5 Limitations... 6 Performance Test...
OPTIMIZE DMA CONFIGURATION IN ENCRYPTION USE CASE. Guillène Ribière, CEO, System Architect
OPTIMIZE DMA CONFIGURATION IN ENCRYPTION USE CASE Guillène Ribière, CEO, System Architect Problem Statement Low Performances on Hardware Accelerated Encryption: Max Measured 10MBps Expectations: 90 MBps
Application-Focused Flash Acceleration
IBM System Storage Application-Focused Flash Acceleration XIV SDD Caching Overview FLASH MEMORY SUMMIT 2012 Anthony Vattathil [email protected] 1 Overview Flash technology is an excellent choice to service
Getting maximum mileage out of tickless
Getting maximum mileage out of tickless Suresh Siddha Venkatesh Pallipadi Arjan Van De Ven Intel Open Source Technology Center {suresh.b.siddha venkatesh.pallipadi arjan.van.de.ven}@intel.com Abstract
Deep Dive: Maximizing EC2 & EBS Performance
Deep Dive: Maximizing EC2 & EBS Performance Tom Maddox, Solutions Architect 2015, Amazon Web Services, Inc. or its affiliates. All rights reserved What we ll cover Amazon EBS overview Volumes Snapshots
SSD Performance Tips: Avoid The Write Cliff
ebook 100% KBs/sec 12% GBs Written SSD Performance Tips: Avoid The Write Cliff An Inexpensive and Highly Effective Method to Keep SSD Performance at 100% Through Content Locality Caching Share this ebook
BHyVe. BSD Hypervisor. Neel Natu Peter Grehan
BHyVe BSD Hypervisor Neel Natu Peter Grehan 1 Introduction BHyVe stands for BSD Hypervisor Pronounced like beehive Type 2 Hypervisor (aka hosted hypervisor) FreeBSD is the Host OS Availability NetApp is
SH-Sim: A Flexible Simulation Platform for Hybrid Storage Systems
, pp.61-70 http://dx.doi.org/10.14257/ijgdc.2014.7.3.07 SH-Sim: A Flexible Simulation Platform for Hybrid Storage Systems Puyuan Yang 1, Peiquan Jin 1,2 and Lihua Yue 1,2 1 School of Computer Science and
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at
Lecture 3: Scaling by Load Balancing 1. Comments on reviews i. 2. Topic 1: Scalability a. QUESTION: What are problems? i. These papers look at distributing load b. QUESTION: What is the context? i. How
Freezing Exabytes of Data at Facebook s Cold Storage. Kestutis Patiejunas ([email protected])
Freezing Exabytes of Data at Facebook s Cold Storage Kestutis Patiejunas ([email protected]) 1990 vs. 2014 Seagate 94171-327 (300MB) iphone 5 16 GB Specs Value Form 3.5" Platters 5 Heads 9 Capacity 300MB
Mobile Devices - An Introduction to the Android Operating Environment. Design, Architecture, and Performance Implications
Mobile Devices - An Introduction to the Android Operating Environment Design, Architecture, and Performance Implications Dominique A. Heger DHTechnologies (DHT) [email protected] 1.0 Introduction With
Using Synology SSD Technology to Enhance System Performance Synology Inc.
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...
Analysis of VDI Storage Performance During Bootstorm
Analysis of VDI Storage Performance During Bootstorm Introduction Virtual desktops are gaining popularity as a more cost effective and more easily serviceable solution. The most resource-dependent process
Operating Systems Concepts: Chapter 7: Scheduling Strategies
Operating Systems Concepts: Chapter 7: Scheduling Strategies Olav Beckmann Huxley 449 http://www.doc.ic.ac.uk/~ob3 Acknowledgements: There are lots. See end of Chapter 1. Home Page for the course: http://www.doc.ic.ac.uk/~ob3/teaching/operatingsystemsconcepts/
Wireless Microcontrollers for Environment Management, Asset Tracking and Consumer. October 2009
Wireless Microcontrollers for Environment Management, Asset Tracking and Consumer October 2009 Jennic highlights Jennic is a fabless semiconductor company providing Wireless Microcontrollers to high-growth
A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems
A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems Anton Beloglazov, Rajkumar Buyya, Young Choon Lee, and Albert Zomaya Present by Leping Wang 1/25/2012 Outline Background
1. Comments on reviews a. Need to avoid just summarizing web page asks you for:
1. Comments on reviews a. Need to avoid just summarizing web page asks you for: i. A one or two sentence summary of the paper ii. A description of the problem they were trying to solve iii. A summary of
OpenMosix Presented by Dr. Moshe Bar and MAASK [01]
OpenMosix Presented by Dr. Moshe Bar and MAASK [01] openmosix is a kernel extension for single-system image clustering. openmosix [24] is a tool for a Unix-like kernel, such as Linux, consisting of adaptive
Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012
Unstructured Data Accelerator (UDA) Author: Motti Beck, Mellanox Technologies Date: March 27, 2012 1 Market Trends Big Data Growing technology deployments are creating an exponential increase in the volume
Memory Channel Storage ( M C S ) Demystified. Jerome McFarland
ory nel Storage ( M C S ) Demystified Jerome McFarland Principal Product Marketer AGENDA + INTRO AND ARCHITECTURE + PRODUCT DETAILS + APPLICATIONS THE COMPUTE-STORAGE DISCONNECT + Compute And Data Have
Multilevel Load Balancing in NUMA Computers
FACULDADE DE INFORMÁTICA PUCRS - Brazil http://www.pucrs.br/inf/pos/ Multilevel Load Balancing in NUMA Computers M. Corrêa, R. Chanin, A. Sales, R. Scheer, A. Zorzo Technical Report Series Number 049 July,
Virtualization in the ARMv7 Architecture Lecture for the Embedded Systems Course CSD, University of Crete (May 20, 2014)
Virtualization in the ARMv7 Architecture Lecture for the Embedded Systems Course CSD, University of Crete (May 20, 2014) ManolisMarazakis ([email protected]) Institute of Computer Science (ICS) Foundation
CPS104 Computer Organization and Programming Lecture 18: Input-Output. Robert Wagner
CPS104 Computer Organization and Programming Lecture 18: Input-Output Robert Wagner cps 104 I/O.1 RW Fall 2000 Outline of Today s Lecture The I/O system Magnetic Disk Tape Buses DMA cps 104 I/O.2 RW Fall
Delivering Quality in Software Performance and Scalability Testing
Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,
Methods to achieve low latency and consistent performance
Methods to achieve low latency and consistent performance Alan Wu, Architect, Memblaze [email protected] 2015/8/13 1 Software Defined Flash Storage System Memblaze provides software defined flash
Deciding which process to run. (Deciding which thread to run) Deciding how long the chosen process can run
SFWR ENG 3BB4 Software Design 3 Concurrent System Design 2 SFWR ENG 3BB4 Software Design 3 Concurrent System Design 11.8 10 CPU Scheduling Chapter 11 CPU Scheduling Policies Deciding which process to run
Confinement Problem. The confinement problem Isolating entities. Example Problem. Server balances bank accounts for clients Server security issues:
Confinement Problem The confinement problem Isolating entities Virtual machines Sandboxes Covert channels Mitigation 1 Example Problem Server balances bank accounts for clients Server security issues:
SLIDE 1 www.bitmicro.com. Previous Next Exit
SLIDE 1 MAXio All Flash Storage Array Popular Applications MAXio N1A6 SLIDE 2 MAXio All Flash Storage Array Use Cases High speed centralized storage for IO intensive applications email, OLTP, databases
Chapter 19: Real-Time Systems. Overview of Real-Time Systems. Objectives. System Characteristics. Features of Real-Time Systems
Chapter 19: Real-Time Systems System Characteristics Features of Real-Time Systems Chapter 19: Real-Time Systems Implementing Real-Time Operating Systems Real-Time CPU Scheduling VxWorks 5.x 19.2 Silberschatz,
Comparison of Hybrid Flash Storage System Performance
Test Validation Comparison of Hybrid Flash Storage System Performance Author: Russ Fellows March 23, 2015 Enabling you to make the best technology decisions 2015 Evaluator Group, Inc. All rights reserved.
BridgeWays Management Pack for VMware ESX
Bridgeways White Paper: Management Pack for VMware ESX BridgeWays Management Pack for VMware ESX Ensuring smooth virtual operations while maximizing your ROI. Published: July 2009 For the latest information,
Section 29. Real-Time Clock and Calendar (RTCC)
Section 29. Real-Time Clock and Calendar (RTCC) HIGHLIGHTS This section of the manual contains the following topics: 29.1 Introduction... 29-2 29.2 Status and Control Registers... 29-3 29.3 Modes of Operation...
Accelerating Database Applications on Linux Servers
White Paper Accelerating Database Applications on Linux Servers Introducing OCZ s LXL Software - Delivering a Data-Path Optimized Solution for Flash Acceleration Allon Cohen, PhD Yaron Klein Eli Ben Namer
Best Practices for Increasing Ceph Performance with SSD
Best Practices for Increasing Ceph Performance with SSD Jian Zhang [email protected] Jiangang Duan [email protected] Agenda Introduction Filestore performance on All Flash Array KeyValueStore
HP Z Turbo Drive PCIe SSD
Performance Evaluation of HP Z Turbo Drive PCIe SSD Powered by Samsung XP941 technology Evaluation Conducted Independently by: Hamid Taghavi Senior Technical Consultant June 2014 Sponsored by: P a g e
Chapter 5 Linux Load Balancing Mechanisms
Chapter 5 Linux Load Balancing Mechanisms Load balancing mechanisms in multiprocessor systems have two compatible objectives. One is to prevent processors from being idle while others processors still
System Software Integration: An Expansive View. Overview
Software Integration: An Expansive View Steven P. Smith Design of Embedded s EE382V Fall, 2009 EE382 SoC Design Software Integration SPS-1 University of Texas at Austin Overview Some Definitions Introduction:
Linux Scheduler. Linux Scheduler
or or Affinity Basic Interactive es 1 / 40 Reality... or or Affinity Basic Interactive es The Linux scheduler tries to be very efficient To do that, it uses some complex data structures Some of what it
File System Management
Lecture 7: Storage Management File System Management Contents Non volatile memory Tape, HDD, SSD Files & File System Interface Directories & their Organization File System Implementation Disk Space Allocation
Solid State Storage in Massive Data Environments Erik Eyberg
Solid State Storage in Massive Data Environments Erik Eyberg Senior Analyst Texas Memory Systems, Inc. Agenda Taxonomy Performance Considerations Reliability Considerations Q&A Solid State Storage Taxonomy
Energy aware RAID Configuration for Large Storage Systems
Energy aware RAID Configuration for Large Storage Systems Norifumi Nishikawa [email protected] Miyuki Nakano [email protected] Masaru Kitsuregawa [email protected] Abstract
OpenFlow Based Load Balancing
OpenFlow Based Load Balancing Hardeep Uppal and Dane Brandon University of Washington CSE561: Networking Project Report Abstract: In today s high-traffic internet, it is often desirable to have multiple
OCZ s NVMe SSDs provide Lower Latency and Faster, more Consistent Performance
OCZ s NVMe SSDs provide Lower Latency and Faster, more Consistent Performance by George Crump, Lead Analyst! When non-volatile flash memory-based solid-state drives (SSDs) were introduced, the protocol
Thread-based analysis of embedded applications with real-time and non real-time processing on a single-processor platform
Thread-based analysis of embedded applications with real-time and non real-time processing on a single-processor platform Dietmar Prisching AVL List GmbH Graz AUSTRIA [email protected] Bernhard
Priority Based Implementation in Pintos
Priority Based Implementation in Pintos Deepa D 1, Nivas K S 2, Preethi V 3 1, 2, 3 Students M-Tech, Dept. of Information Technology, SNS College of Engg., Coimbatore. Abstract Pintos is a simple operating
Efficient Load Balancing using VM Migration by QEMU-KVM
International Journal of Computer Science and Telecommunications [Volume 5, Issue 8, August 2014] 49 ISSN 2047-3338 Efficient Load Balancing using VM Migration by QEMU-KVM Sharang Telkikar 1, Shreyas Talele
EMC SCALEIO OPERATION OVERVIEW
EMC SCALEIO OPERATION OVERVIEW Ensuring Non-disruptive Operation and Upgrade ABSTRACT This white paper reviews the challenges organizations face as they deal with the growing need for always-on levels
Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays
Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays Database Solutions Engineering By Murali Krishnan.K Dell Product Group October 2009
Why Relative Share Does Not Work
Why Relative Share Does Not Work Introduction Velocity Software, Inc March 2010 Rob van der Heij rvdheij @ velocitysoftware.com Installations that run their production and development Linux servers on
Online Failure Prediction in Cloud Datacenters
Online Failure Prediction in Cloud Datacenters Yukihiro Watanabe Yasuhide Matsumoto Once failures occur in a cloud datacenter accommodating a large number of virtual resources, they tend to spread rapidly
Lesson Objectives. To provide a grand tour of the major operating systems components To provide coverage of basic computer system organization
Lesson Objectives To provide a grand tour of the major operating systems components To provide coverage of basic computer system organization AE3B33OSD Lesson 1 / Page 2 What is an Operating System? A
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What
EZ GPO. Power Management Tool for Network Administrators.
EZ GPO Power Management Tool for Network Administrators. Table of Contents Introduction...3 Installation...4 Configuration...6 Base Options...7 Options Properties Dialogue...8 Suggested Configuration Settings...9
Kerrighed: use cases. Cyril Brulebois. Kerrighed. Kerlabs
Kerrighed: use cases Cyril Brulebois [email protected] Kerrighed http://www.kerrighed.org/ Kerlabs http://www.kerlabs.com/ 1 / 23 Introducing Kerrighed What s Kerrighed? Single-System Image (SSI)
Operating System Multilevel Load Balancing
Operating System Multilevel Load Balancing M. Corrêa, A. Zorzo Faculty of Informatics - PUCRS Porto Alegre, Brazil {mcorrea, zorzo}@inf.pucrs.br R. Scheer HP Brazil R&D Porto Alegre, Brazil [email protected]
Using the TASKING Software Platform for AURIX
Using the TASKING Software Platform for AURIX MA160-869 (v1.0rb3) June 19, 2015 Copyright 2015 Altium BV. All rights reserved. You are permitted to print this document provided that (1) the use of such
